root commited on
Commit
0ea1b17
1 Parent(s): aadbd6e
Files changed (4) hide show
  1. app.py +30 -9
  2. model.README.md +24 -0
  3. token_identifier.txt +1 -0
  4. train_dreambooth.py +2 -0
app.py CHANGED
@@ -34,13 +34,12 @@ if(is_gpu_associated):
34
  # model_v1 = snapshot_download(repo_id="multimodalart/sd-fine-tunable")
35
  # model_v2 = snapshot_download(repo_id="stabilityai/stable-diffusion-2")
36
  # model_v2_512 = snapshot_download(repo_id="stabilityai/stable-diffusion-2-base")
37
- model_alt = snapshot_download(repo_id="BAAI/AltDiffusion")
38
- model_alt_m9 = snapshot_download(repo_id="BAAI/AltDiffusion-m9")
39
- safety_checker = snapshot_download(repo_id="multimodalart/sd-sc")
40
- model_to_load = model_alt_m9
41
-
42
- with zipfile.ZipFile("mix.zip", 'r') as zip_ref:
43
- zip_ref.extractall(".")
44
 
45
  def swap_text(option, base):
46
  resize_width = 768 if base == "v2-768" else 512
@@ -450,6 +449,27 @@ def checkbox_swap(checkbox):
450
  return [gr.update(visible=checkbox), gr.update(visible=checkbox), gr.update(visible=checkbox), gr.update(visible=checkbox)]
451
 
452
  with gr.Blocks(css=css) as demo:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
453
  with gr.Box():
454
  if is_shared_ui:
455
  top_description = gr.HTML(f'''
@@ -482,12 +502,13 @@ with gr.Blocks(css=css) as demo:
482
  <p>Do a <code>pip install requirements-local.txt</code></p>
483
  </div>
484
  ''')
 
485
  gr.Markdown("# Dreambooth Training UI 💭")
486
- gr.Markdown("Customize Stable Diffusion v1 or v2 (ⁿᵉʷ!) by giving it a few examples of a concept. Based on the [🧨 diffusers](https://github.com/huggingface/diffusers) implementation, additional techniques from [TheLastBen](https://github.com/TheLastBen/diffusers) and [ShivamShrirao](https://github.com/ShivamShrirao/diffusers)")
487
 
488
  with gr.Row() as what_are_you_training:
489
  type_of_thing = gr.Dropdown(label="What would you like to train?", choices=["object", "person", "style"], value="object", interactive=True)
490
- base_model_to_use = gr.Dropdown(label="Which base model would you like to use?", choices=["v1-5", "v2-512", "v2-768", "alt", "alt_m9"], value="alt_m9", interactive=True)
491
 
492
  #Very hacky approach to emulate dynamically created Gradio components
493
  with gr.Row() as upload_your_concept:
 
34
  # model_v1 = snapshot_download(repo_id="multimodalart/sd-fine-tunable")
35
  # model_v2 = snapshot_download(repo_id="stabilityai/stable-diffusion-2")
36
  # model_v2_512 = snapshot_download(repo_id="stabilityai/stable-diffusion-2-base")
37
+ # model_alt = snapshot_download(repo_id="BAAI/AltDiffusion")
38
+ # model_alt_m9 = snapshot_download(repo_id="BAAI/AltDiffusion-m9")
39
+ # safety_checker = snapshot_download(repo_id="multimodalart/sd-sc")
40
+ model_to_load = None # model_alt_m9
41
+ # with zipfile.ZipFile("mix.zip", 'r') as zip_ref:
42
+ # zip_ref.extractall(".")
 
43
 
44
  def swap_text(option, base):
45
  resize_width = 768 if base == "v2-768" else 512
 
449
  return [gr.update(visible=checkbox), gr.update(visible=checkbox), gr.update(visible=checkbox), gr.update(visible=checkbox)]
450
 
451
  with gr.Blocks(css=css) as demo:
452
+ gr.HTML(f'''
453
+ <div style="text-align: center; max-width: 650px; margin: 0 auto;">
454
+ <div
455
+ style="
456
+ display: inline-flex;
457
+ gap: 1.2rem;
458
+ font-size: 1.75rem;
459
+ margin-bottom: 10px;
460
+ width: 700px;
461
+ height: 100px;
462
+ margin: 0 auto;
463
+ /* border: 1px solid red; */
464
+ justify-content: center;
465
+ ">
466
+ <a href="https://github.com/FlagAI-Open/FlagAI"><img src="https://raw.githubusercontent.com/920232796/test/master/WechatIMG6906.png" alt="FlagAI" width="80%" height="80%" style="margin: 0 auto;"></a>
467
+ </div>
468
+ <p style="margin-bottom: 10px; font-size: 94%">
469
+ This is a dreambooth Training UI for <a href="https://huggingface.co/BAAI/AltDiffusion-m9" style="text-decoration: underline;">AltDiffusion-m9 model</a>,which is a multilingual image-to-text model supported 9 languages.
470
+ </p>
471
+ </div>
472
+ ''')
473
  with gr.Box():
474
  if is_shared_ui:
475
  top_description = gr.HTML(f'''
 
502
  <p>Do a <code>pip install requirements-local.txt</code></p>
503
  </div>
504
  ''')
505
+
506
  gr.Markdown("# Dreambooth Training UI 💭")
507
+ gr.Markdown("Customize AltDiffusion and AltDiffusion-m9(ⁿᵉʷ!) by giving it a few examples of a concept. Based on the [🧨 diffusers](https://github.com/huggingface/diffusers) implementation, additional techniques from [TheLastBen](https://github.com/TheLastBen/diffusers) and [ShivamShrirao](https://github.com/ShivamShrirao/diffusers)")
508
 
509
  with gr.Row() as what_are_you_training:
510
  type_of_thing = gr.Dropdown(label="What would you like to train?", choices=["object", "person", "style"], value="object", interactive=True)
511
+ base_model_to_use = gr.Dropdown(label="Which base model would you like to use?", choices=["alt", "alt_m9"], value="alt_m9", interactive=True)
512
 
513
  #Very hacky approach to emulate dynamically created Gradio components
514
  with gr.Row() as upload_your_concept:
model.README.md ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ tags:
4
+ - text-to-image
5
+ widget:
6
+ - text: sks
7
+ ---
8
+ ### test-m9 Dreambooth model trained by Alon77777 with [Hugging Face Dreambooth Training Space](https://huggingface.co/spaces/multimodalart/dreambooth-training) with the alt_m9 base model
9
+
10
+ You run your new concept via `diffusers` [Colab Notebook for Inference](https://colab.research.google.com/github/huggingface/notebooks/blob/main/diffusers/sd_dreambooth_inference.ipynb). Don't forget to use the concept prompts!
11
+
12
+ Sample pictures of:
13
+
14
+
15
+
16
+
17
+
18
+
19
+
20
+
21
+
22
+
23
+ sks (use that on your prompt)
24
+ ![sks 0](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%2810%29.jpg)![sks 1](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%285%29.jpg)![sks 2](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%282%29.jpg)![sks 3](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%283%29.jpg)![sks 4](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%2811%29.jpg)![sks 5](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%288%29.jpg)![sks 6](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%281%29.jpg)![sks 7](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%289%29.jpg)![sks 8](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%286%29.jpg)![sks 9](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%284%29.jpg)![sks 10](https://huggingface.co/Alon77777/test-m9/resolve/main/concept_images/sks_%287%29.jpg)
token_identifier.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ sks
train_dreambooth.py CHANGED
@@ -728,6 +728,8 @@ def run_training(args_imported):
728
  if args.train_text_encoder:
729
  text_encoder.train()
730
  for step, batch in enumerate(train_dataloader):
 
 
731
  with accelerator.accumulate(unet):
732
  # Convert images to latent space
733
  with torch.no_grad():
 
728
  if args.train_text_encoder:
729
  text_encoder.train()
730
  for step, batch in enumerate(train_dataloader):
731
+ import pdb
732
+ pdb.set_trace()
733
  with accelerator.accumulate(unet):
734
  # Convert images to latent space
735
  with torch.no_grad():