Spaces:
Sleeping
Sleeping
mrfakename
commited on
Commit
•
0559e57
1
Parent(s):
5975221
Sync from GitHub repo
Browse filesThis Space is synced from the GitHub repo: https://github.com/SWivid/F5-TTS. Please submit contributions to the Space there
app.py
CHANGED
@@ -454,13 +454,31 @@ Have a conversation with an AI using your reference voice!
|
|
454 |
"""
|
455 |
)
|
456 |
|
457 |
-
|
|
|
458 |
|
459 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
460 |
|
461 |
-
@gpu_decorator
|
462 |
-
def load_chat_model():
|
463 |
-
global chat_model_state, chat_tokenizer_state
|
464 |
if chat_model_state is None:
|
465 |
show_info = gr.Info
|
466 |
show_info("Loading chat model...")
|
@@ -469,10 +487,6 @@ Have a conversation with an AI using your reference voice!
|
|
469 |
chat_tokenizer_state = AutoTokenizer.from_pretrained(model_name)
|
470 |
show_info("Chat model loaded.")
|
471 |
|
472 |
-
return gr.update(visible=False), gr.update(visible=True)
|
473 |
-
|
474 |
-
load_chat_model_btn.click(load_chat_model, outputs=[load_chat_model_btn, chat_interface_container])
|
475 |
-
|
476 |
with chat_interface_container:
|
477 |
with gr.Row():
|
478 |
with gr.Column():
|
|
|
454 |
"""
|
455 |
)
|
456 |
|
457 |
+
if not USING_SPACES:
|
458 |
+
load_chat_model_btn = gr.Button("Load Chat Model", variant="primary")
|
459 |
|
460 |
+
chat_interface_container = gr.Column(visible=False)
|
461 |
+
|
462 |
+
@gpu_decorator
|
463 |
+
def load_chat_model():
|
464 |
+
global chat_model_state, chat_tokenizer_state
|
465 |
+
if chat_model_state is None:
|
466 |
+
show_info = gr.Info
|
467 |
+
show_info("Loading chat model...")
|
468 |
+
model_name = "Qwen/Qwen2.5-3B-Instruct"
|
469 |
+
chat_model_state = AutoModelForCausalLM.from_pretrained(
|
470 |
+
model_name, torch_dtype="auto", device_map="auto"
|
471 |
+
)
|
472 |
+
chat_tokenizer_state = AutoTokenizer.from_pretrained(model_name)
|
473 |
+
show_info("Chat model loaded.")
|
474 |
+
|
475 |
+
return gr.update(visible=False), gr.update(visible=True)
|
476 |
+
|
477 |
+
load_chat_model_btn.click(load_chat_model, outputs=[load_chat_model_btn, chat_interface_container])
|
478 |
+
|
479 |
+
else:
|
480 |
+
chat_interface_container = gr.Column()
|
481 |
|
|
|
|
|
|
|
482 |
if chat_model_state is None:
|
483 |
show_info = gr.Info
|
484 |
show_info("Loading chat model...")
|
|
|
487 |
chat_tokenizer_state = AutoTokenizer.from_pretrained(model_name)
|
488 |
show_info("Chat model loaded.")
|
489 |
|
|
|
|
|
|
|
|
|
490 |
with chat_interface_container:
|
491 |
with gr.Row():
|
492 |
with gr.Column():
|