Spaces:
Paused
Paused
Commit
Β·
526db9a
1
Parent(s):
8b378ca
feat: add dockerfile
Browse files
README.md
CHANGED
@@ -1,11 +1,31 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
pinned: false
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
title: "WhisperFusion"
|
2 |
+
emoji: "π"
|
3 |
+
colorFrom: "pink"
|
4 |
+
colorTo: "green"
|
5 |
+
sdk: "docker"
|
6 |
+
python_version: "3.10"
|
7 |
+
sdk_version: "latest"
|
8 |
+
suggested_hardware: "t4-small"
|
9 |
+
suggested_storage: "medium"
|
10 |
+
app_file: "examples/chatbot/html/main.py"
|
11 |
+
app_port: 7860
|
12 |
+
base_path: "/"
|
13 |
+
fullWidth: false
|
14 |
+
models: ["teknium/OpenHermes-2.5-Mistral-7B"]
|
15 |
+
datasets: []
|
16 |
+
tags: ["AI", "chatbot", "speech-to-text", "real-time", "TensorRT", "LLM"]
|
17 |
pinned: false
|
18 |
+
hf_oauth: false
|
19 |
+
hf_oauth_scopes: []
|
20 |
+
disable_embedding: false
|
21 |
+
startup_duration_timeout: "30m"
|
22 |
+
custom_headers:
|
23 |
+
cross-origin-embedder-policy: "require-corp"
|
24 |
+
cross-origin-opener-policy: "same-origin"
|
25 |
+
cross-origin-resource-policy: "cross-origin"
|
26 |
+
preload_from_hub:
|
27 |
+
- "NVIDIA/TensorRT-LLM examples/whisper/whisper_small_en,examples/phi/phi_engine,examples/phi/phi-2"
|
28 |
+
description: "WhisperFusion is an AI chatbot that provides ultra-low latency conversations. It integrates Mistral, a Large Language Model (LLM), on top of the real-time speech-to-text pipeline. It utilizes OpenAI WhisperLive to convert spoken language into text in real-time and is optimized to run as TensorRT engines, ensuring high-performance and low-latency processing."
|
29 |
+
installation: "Install [TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM/blob/main/docs/source/installation.md) to build Whisper and Mistral TensorRT engines. Refer to the README and the [Dockerfile.multi](https://github.com/NVIDIA/TensorRT-LLM/blob/main/docker/Dockerfile.multi) to install the required packages in the base pytorch docker image."
|
30 |
+
usage: "Run the main.py script with the appropriate arguments to start the chatbot."
|
31 |
+
source: "This information is provided by Hugging Face."
|