Spaces:
Running
on
Zero
Running
on
Zero
Minor fixes
Browse files- Dockerfile +1 -1
- README.md +2 -1
- docker-compose.yml +1 -3
- requirements.txt +9 -9
- xtts.py +7 -0
Dockerfile
CHANGED
@@ -7,7 +7,7 @@ RUN apt-get update && \
|
|
7 |
|
8 |
WORKDIR /app
|
9 |
COPY requirements.txt .
|
10 |
-
RUN python -m pip install
|
11 |
&& python -m pip cache purge
|
12 |
|
13 |
RUN python -m unidic download
|
|
|
7 |
|
8 |
WORKDIR /app
|
9 |
COPY requirements.txt .
|
10 |
+
RUN python -m pip install -r requirements.txt \
|
11 |
&& python -m pip cache purge
|
12 |
|
13 |
RUN python -m unidic download
|
README.md
CHANGED
@@ -4,9 +4,10 @@ emoji: ⚡
|
|
4 |
colorFrom: red
|
5 |
colorTo: green
|
6 |
sdk: gradio
|
7 |
-
sdk_version: 4.
|
8 |
app_file: app.py
|
9 |
pinned: false
|
|
|
10 |
---
|
11 |
|
12 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
4 |
colorFrom: red
|
5 |
colorTo: green
|
6 |
sdk: gradio
|
7 |
+
sdk_version: 4.44.0
|
8 |
app_file: app.py
|
9 |
pinned: false
|
10 |
+
startup_duration_timeout: 2h
|
11 |
---
|
12 |
|
13 |
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
docker-compose.yml
CHANGED
@@ -1,5 +1,3 @@
|
|
1 |
-
name: webui-docker
|
2 |
-
|
3 |
volumes:
|
4 |
servel-model-root:
|
5 |
|
@@ -13,7 +11,7 @@ services:
|
|
13 |
COQUI_TOS_AGREED: 1
|
14 |
CUSTOM_MODEL_PATH: /root/.local/share/tts/tts_models--multilingual--multi-dataset--xtts_v2
|
15 |
ports:
|
16 |
-
-
|
17 |
volumes:
|
18 |
- type: volume
|
19 |
source: servel-model-root
|
|
|
|
|
|
|
1 |
volumes:
|
2 |
servel-model-root:
|
3 |
|
|
|
11 |
COQUI_TOS_AGREED: 1
|
12 |
CUSTOM_MODEL_PATH: /root/.local/share/tts/tts_models--multilingual--multi-dataset--xtts_v2
|
13 |
ports:
|
14 |
+
- 8080:7860
|
15 |
volumes:
|
16 |
- type: volume
|
17 |
source: servel-model-root
|
requirements.txt
CHANGED
@@ -1,15 +1,15 @@
|
|
1 |
-
torch
|
2 |
-
torchvision
|
3 |
-
torchaudio
|
4 |
-
gradio
|
5 |
-
numpy
|
6 |
TTS @ git+https://github.com/coqui-ai/TTS@fa28f99f1508b5b5366539b2149963edcb80ba62
|
7 |
uvicorn[standard]==0.23.2
|
8 |
-
deepspeed
|
9 |
-
pydantic
|
10 |
-
python-multipart==0.0.
|
11 |
typing-extensions>=4.8.0
|
12 |
-
cutlet
|
13 |
mecab-python3==1.0.6
|
14 |
unidic-lite==1.0.8
|
15 |
unidic==1.1.0
|
|
|
1 |
+
torch==2.1.0
|
2 |
+
torchvision==0.16.0
|
3 |
+
torchaudio==2.1.0
|
4 |
+
gradio==4.44
|
5 |
+
numpy==1.22.0
|
6 |
TTS @ git+https://github.com/coqui-ai/TTS@fa28f99f1508b5b5366539b2149963edcb80ba62
|
7 |
uvicorn[standard]==0.23.2
|
8 |
+
deepspeed==0.15.1
|
9 |
+
pydantic==2.9.0
|
10 |
+
python-multipart==0.0.9
|
11 |
typing-extensions>=4.8.0
|
12 |
+
cutlet==0.4.0
|
13 |
mecab-python3==1.0.6
|
14 |
unidic-lite==1.0.8
|
15 |
unidic==1.1.0
|
xtts.py
CHANGED
@@ -150,6 +150,13 @@ class TTSInputs(BaseModel):
|
|
150 |
def predict_speech(parsed_input: TTSInputs):
|
151 |
speaker_embedding = torch.tensor(parsed_input.speaker_embedding).unsqueeze(0).unsqueeze(-1)
|
152 |
gpt_cond_latent = torch.tensor(parsed_input.gpt_cond_latent).reshape((-1, 1024)).unsqueeze(0)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
153 |
text = parsed_input.text
|
154 |
language = parsed_input.language
|
155 |
temperature = parsed_input.temperature
|
|
|
150 |
def predict_speech(parsed_input: TTSInputs):
|
151 |
speaker_embedding = torch.tensor(parsed_input.speaker_embedding).unsqueeze(0).unsqueeze(-1)
|
152 |
gpt_cond_latent = torch.tensor(parsed_input.gpt_cond_latent).reshape((-1, 1024)).unsqueeze(0)
|
153 |
+
|
154 |
+
print("speaker embedding")
|
155 |
+
print(speaker_embedding)
|
156 |
+
|
157 |
+
print("latent")
|
158 |
+
print(gpt_cond_latent)
|
159 |
+
|
160 |
text = parsed_input.text
|
161 |
language = parsed_input.language
|
162 |
temperature = parsed_input.temperature
|