rrg92 commited on
Commit
3e62df4
1 Parent(s): 140b723

Minor fixes

Browse files
Files changed (5) hide show
  1. Dockerfile +1 -1
  2. README.md +2 -1
  3. docker-compose.yml +1 -3
  4. requirements.txt +9 -9
  5. xtts.py +7 -0
Dockerfile CHANGED
@@ -7,7 +7,7 @@ RUN apt-get update && \
7
 
8
  WORKDIR /app
9
  COPY requirements.txt .
10
- RUN python -m pip install --use-deprecated=legacy-resolver -r requirements.txt \
11
  && python -m pip cache purge
12
 
13
  RUN python -m unidic download
 
7
 
8
  WORKDIR /app
9
  COPY requirements.txt .
10
+ RUN python -m pip install -r requirements.txt \
11
  && python -m pip cache purge
12
 
13
  RUN python -m unidic download
README.md CHANGED
@@ -4,9 +4,10 @@ emoji: ⚡
4
  colorFrom: red
5
  colorTo: green
6
  sdk: gradio
7
- sdk_version: 4.43.0
8
  app_file: app.py
9
  pinned: false
 
10
  ---
11
 
12
  Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
4
  colorFrom: red
5
  colorTo: green
6
  sdk: gradio
7
+ sdk_version: 4.44.0
8
  app_file: app.py
9
  pinned: false
10
+ startup_duration_timeout: 2h
11
  ---
12
 
13
  Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
docker-compose.yml CHANGED
@@ -1,5 +1,3 @@
1
- name: webui-docker
2
-
3
  volumes:
4
  servel-model-root:
5
 
@@ -13,7 +11,7 @@ services:
13
  COQUI_TOS_AGREED: 1
14
  CUSTOM_MODEL_PATH: /root/.local/share/tts/tts_models--multilingual--multi-dataset--xtts_v2
15
  ports:
16
- - 3000:7860
17
  volumes:
18
  - type: volume
19
  source: servel-model-root
 
 
 
1
  volumes:
2
  servel-model-root:
3
 
 
11
  COQUI_TOS_AGREED: 1
12
  CUSTOM_MODEL_PATH: /root/.local/share/tts/tts_models--multilingual--multi-dataset--xtts_v2
13
  ports:
14
+ - 8080:7860
15
  volumes:
16
  - type: volume
17
  source: servel-model-root
requirements.txt CHANGED
@@ -1,15 +1,15 @@
1
- torch
2
- torchvision
3
- torchaudio
4
- gradio
5
- numpy
6
  TTS @ git+https://github.com/coqui-ai/TTS@fa28f99f1508b5b5366539b2149963edcb80ba62
7
  uvicorn[standard]==0.23.2
8
- deepspeed
9
- pydantic
10
- python-multipart==0.0.6
11
  typing-extensions>=4.8.0
12
- cutlet
13
  mecab-python3==1.0.6
14
  unidic-lite==1.0.8
15
  unidic==1.1.0
 
1
+ torch==2.1.0
2
+ torchvision==0.16.0
3
+ torchaudio==2.1.0
4
+ gradio==4.44
5
+ numpy==1.22.0
6
  TTS @ git+https://github.com/coqui-ai/TTS@fa28f99f1508b5b5366539b2149963edcb80ba62
7
  uvicorn[standard]==0.23.2
8
+ deepspeed==0.15.1
9
+ pydantic==2.9.0
10
+ python-multipart==0.0.9
11
  typing-extensions>=4.8.0
12
+ cutlet==0.4.0
13
  mecab-python3==1.0.6
14
  unidic-lite==1.0.8
15
  unidic==1.1.0
xtts.py CHANGED
@@ -150,6 +150,13 @@ class TTSInputs(BaseModel):
150
  def predict_speech(parsed_input: TTSInputs):
151
  speaker_embedding = torch.tensor(parsed_input.speaker_embedding).unsqueeze(0).unsqueeze(-1)
152
  gpt_cond_latent = torch.tensor(parsed_input.gpt_cond_latent).reshape((-1, 1024)).unsqueeze(0)
 
 
 
 
 
 
 
153
  text = parsed_input.text
154
  language = parsed_input.language
155
  temperature = parsed_input.temperature
 
150
  def predict_speech(parsed_input: TTSInputs):
151
  speaker_embedding = torch.tensor(parsed_input.speaker_embedding).unsqueeze(0).unsqueeze(-1)
152
  gpt_cond_latent = torch.tensor(parsed_input.gpt_cond_latent).reshape((-1, 1024)).unsqueeze(0)
153
+
154
+ print("speaker embedding")
155
+ print(speaker_embedding)
156
+
157
+ print("latent")
158
+ print(gpt_cond_latent)
159
+
160
  text = parsed_input.text
161
  language = parsed_input.language
162
  temperature = parsed_input.temperature