NYUAD-ComNets commited on
Commit
8f6b310
·
1 Parent(s): 14975f0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +92 -14
README.md CHANGED
@@ -12,24 +12,102 @@ tags:
12
  inference: true
13
  ---
14
 
15
- # LoRA text2image fine-tuning - NYUAD-ComNets/Asian_Male_Profession_Model
 
16
 
17
- These are LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. The weights were fine-tuned on the NYUAD-ComNets/Asian_Male_Profession dataset. You can find some example images in the following.
 
18
 
19
- ![img_0](./image_0.png)
20
- ![img_1](./image_1.png)
21
- ![img_2](./image_2.png)
22
- ![img_3](./image_3.png)
23
- ![img_4](./image_4.png)
24
- ![img_5](./image_5.png)
25
- ![img_6](./image_6.png)
26
- ![img_7](./image_7.png)
27
- ![img_8](./image_8.png)
28
- ![img_9](./image_9.png)
29
- ![img_10](./image_10.png)
30
- ![img_11](./image_11.png)
31
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
  LoRA for the text encoder was enabled: False.
34
 
35
  Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  inference: true
13
  ---
14
 
15
+ # Model description
16
+ LoRA text2image fine-tuning - NYUAD-ComNets/Asian_Male_Profession_Model
17
 
18
+ These are LoRA adaption weights for stabilityai/stable-diffusion-xl-base-1.0. The weights were fine-tuned on the NYUAD-ComNets/Asian_Male_Profession dataset.
19
+ You can find some example images.
20
 
21
+ prompt: a photo of a {profession}, looking at the camera, closeup headshot facing forward, ultra quality, sharp focus
 
 
 
 
 
 
 
 
 
 
 
22
 
23
+ # How to use this model:
24
+
25
+ ``` python
26
+
27
+
28
+ import torch
29
+ from compel import Compel, ReturnedEmbeddingsType
30
+ from diffusers import DiffusionPipeline
31
+
32
+ import random
33
+
34
+
35
+ negative_prompt = "cartoon, anime, 3d, painting, b&w, low quality"
36
+
37
+
38
+ models=["NYUAD-ComNets/Asian_Female_Profession_Model","NYUAD-ComNets/Black_Female_Profession_Model","NYUAD-ComNets/White_Female_Profession_Model",
39
+ "NYUAD-ComNets/Indian_Female_Profession_Model","NYUAD-ComNets/Latino_Hispanic_Female_Profession_Model","NYUAD-ComNets/Middle_Eastern_Female_Profession_Model",
40
+ "NYUAD-ComNets/Asian_Male_Profession_Model","NYUAD-ComNets/Black_Male_Profession_Model","NYUAD-ComNets/White_Male_Profession_Model",
41
+ "NYUAD-ComNets/Indian_Male_Profession_Model","NYUAD-ComNets/Latino_Hispanic_Male_Profession_Model","NYUAD-ComNets/Middle_Eastern_Male_Profession_Model"]
42
+
43
+ adapters=["asian_female","black_female","white_female","indian_female","latino_female","middle_east_female",
44
+ "asian_male","black_male","white_male","indian_male","latino_male","middle_east_male"]
45
+
46
+ pipeline = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-base-1.0", variant="fp16", use_safetensors=True, torch_dtype=torch.float16).to("cuda")
47
+
48
+
49
+ for i,j in zip(models,adapters):
50
+ pipeline.load_lora_weights(i, weight_name="pytorch_lora_weights.safetensors",adapter_name=j)
51
+
52
+
53
+ prof='doctor'
54
+
55
+
56
+ pipeline.set_adapters(random.choice(adapters))
57
+
58
+
59
+ compel = Compel(tokenizer=[pipeline.tokenizer, pipeline.tokenizer_2] ,
60
+ text_encoder=[pipeline.text_encoder, pipeline.text_encoder_2],
61
+ returned_embeddings_type=ReturnedEmbeddingsType.PENULTIMATE_HIDDEN_STATES_NON_NORMALIZED,
62
+ requires_pooled=[False, True],truncate_long_prompts=False)
63
+
64
+
65
+ conditioning, pooled = compel("a photo of a {}, looking at the camera, closeup headshot facing forward, ultra quality, sharp focus".format(prof))
66
+
67
+ negative_conditioning, negative_pooled = compel(negative_prompt)
68
+ [conditioning, negative_conditioning] = compel.pad_conditioning_tensors_to_same_length([conditioning, negative_conditioning])
69
+
70
+ image = pipeline(prompt_embeds=conditioning, negative_prompt_embeds=negative_conditioning,
71
+ pooled_prompt_embeds=pooled, negative_pooled_prompt_embeds=negative_pooled,
72
+ num_inference_steps=40).images[0]
73
+
74
+ image.save('/../../x.jpg')
75
+
76
+ ```
77
+
78
+
79
+ # Examples
80
+
81
+ | | | |
82
+ |:-------------------------:|:-------------------------:|:-------------------------:|
83
+ |<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_0.png"> | <img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_1.png">|<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_2.png">|
84
+ |<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_3.png"> | <img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_4.png">|<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_5.png">|
85
+ |<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_6.png"> | <img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_7.png">|<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_8.png">|
86
+ |<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_9.png"> | <img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_10.png">|<img width="500" alt="screen shot 2017-08-07 at 12 18 15 pm" src="./image_11.png">|
87
+
88
+
89
+
90
+
91
+ # Training data
92
+
93
+ NYUAD-ComNets/Asian_Male_Profession dataset was used to fine-tune stabilityai/stable-diffusion-xl-base-1.0
94
+
95
+
96
+
97
+ # Configurations
98
 
99
  LoRA for the text encoder was enabled: False.
100
 
101
  Special VAE used for training: madebyollin/sdxl-vae-fp16-fix.
102
+
103
+
104
+
105
+ # BibTeX entry and citation info
106
+
107
+ ```
108
+ @misc{ComNets,
109
+ url={[https://huggingface.co/NYUAD-ComNets/Asian_Male_Profession_Model](https://huggingface.co/NYUAD-ComNets/Asian_Male_Profession_Model)},
110
+ title={Asian_Male_Profession_Model},
111
+ author={Nouar AlDahoul, Talal Rahwan, Yasir Zaki}
112
+ }
113
+ ```