Spaces:
Sleeping
Sleeping
added markdown title per gradio row
Browse files
app.py
CHANGED
@@ -67,6 +67,7 @@ def scrape_config_from_hub(repo_id):
|
|
67 |
|
68 |
def build_interface(estimate_vram_fn):
|
69 |
with gr.Blocks() as app:
|
|
|
70 |
option = gr.Radio(["Repo ID", "Model Parameters"], label="Select Input Type")
|
71 |
repo_id = gr.Textbox(label="Repo ID", visible=False, placeholder="mistralai/Mistral-7B-v0.1")
|
72 |
|
@@ -91,7 +92,7 @@ def build_interface(estimate_vram_fn):
|
|
91 |
outputs=[repo_id, model_params_row]
|
92 |
)
|
93 |
|
94 |
-
|
95 |
with gr.Row(equal_height=True):
|
96 |
training_params = [gr.Dropdown(label="Micro batch size", choices=BATCH_SIZES, value=4, info="Micro batch size (batch size per device/GPU)"),
|
97 |
gr.Dropdown(label="ZeRO stage", choices=ZERO_STAGES, value=0, info="ZeRO optimization stage"),
|
@@ -100,7 +101,7 @@ def build_interface(estimate_vram_fn):
|
|
100 |
gr.Dropdown(label="Optimizer", choices=OPTIMIZERS, value="adamw", info="Type of optimizer"),
|
101 |
gr.Dropdown(label="QLoRA", choices=[False, True], value=False, info="Finetune with QLoRA enabled"),
|
102 |
gr.Slider(label="Num GPUs", minimum=1, maximum=64, step=1, value=4, info="Number of GPUs. Necessary for estimating ZeRO stages"),
|
103 |
-
gr.Textbox(label="Cache dir", value=None, placeholder=".huggingface_configs", info="HuggingFace cache directory to download config
|
104 |
]
|
105 |
|
106 |
submit_btn = gr.Button("Estimate!")
|
|
|
67 |
|
68 |
def build_interface(estimate_vram_fn):
|
69 |
with gr.Blocks() as app:
|
70 |
+
gr.Markdown("## Select either an existing HF model from a repository or choose your own model parameters")
|
71 |
option = gr.Radio(["Repo ID", "Model Parameters"], label="Select Input Type")
|
72 |
repo_id = gr.Textbox(label="Repo ID", visible=False, placeholder="mistralai/Mistral-7B-v0.1")
|
73 |
|
|
|
92 |
outputs=[repo_id, model_params_row]
|
93 |
)
|
94 |
|
95 |
+
gr.Markdown("## Select training parameters")
|
96 |
with gr.Row(equal_height=True):
|
97 |
training_params = [gr.Dropdown(label="Micro batch size", choices=BATCH_SIZES, value=4, info="Micro batch size (batch size per device/GPU)"),
|
98 |
gr.Dropdown(label="ZeRO stage", choices=ZERO_STAGES, value=0, info="ZeRO optimization stage"),
|
|
|
101 |
gr.Dropdown(label="Optimizer", choices=OPTIMIZERS, value="adamw", info="Type of optimizer"),
|
102 |
gr.Dropdown(label="QLoRA", choices=[False, True], value=False, info="Finetune with QLoRA enabled"),
|
103 |
gr.Slider(label="Num GPUs", minimum=1, maximum=64, step=1, value=4, info="Number of GPUs. Necessary for estimating ZeRO stages"),
|
104 |
+
gr.Textbox(label="Cache dir", value=None, placeholder=".huggingface_configs", info="HuggingFace cache directory to download config to")
|
105 |
]
|
106 |
|
107 |
submit_btn = gr.Button("Estimate!")
|