Text2Video-Zero / app_text_to_video.py
lev1's picture
T2V Tab improvements
2d7762b
raw
history blame
2.35 kB
import gradio as gr
from model import Model
from functools import partial
examples = [
["an astronaut waving the arm on the moon"],
["a sloth surfing on a wakeboard"],
["an astronaut walking on a street"],
["a cute cat walking on grass"],
["a horse is galloping on a street"],
["an astronaut is skiing down the hill"],
["a gorilla walking alone down the street"],
["a gorilla dancing on times square"],
["A panda dancing dancing like crazy on Times Square"],
]
def create_demo(model: Model):
with gr.Blocks() as demo:
with gr.Row():
gr.Markdown('## Text2Video-Zero: Video Generation')
with gr.Row():
with gr.Column():
prompt = gr.Textbox(label='Prompt')
run_button = gr.Button(label='Run')
with gr.Accordion('Advanced options', open=False):
motion_field_strength_x = gr.Slider(label='Global Translation $\delta_{x}$',
minimum=-20,
maximum=20,
value=12,
step=1)
motion_field_strength_y = gr.Slider(label='Global Translation $\delta_{y}$',
minimum=-20,
maximum=20,
value=12,
step=1)
# a_prompt = gr.Textbox(label="Added Prompt", value='best quality, extremely detailed')
n_prompt = gr.Textbox(label="Optional Negative Prompt",
value='')
with gr.Column():
result = gr.Video(label="Generated Video")
inputs = [
prompt,
motion_field_strength_x,
motion_field_strength_y,
n_prompt
]
gr.Examples(examples=examples,
inputs=inputs,
outputs=result,
# cache_examples=False,
cache_examples=os.getenv('SYSTEM') == 'spaces',
run_on_click=False,
)
run_button.click(fn=model.process_text2video,
inputs=inputs,
outputs=result,)
return demo