EpsilonGreedy commited on
Commit
3df3ede
1 Parent(s): 901cc50

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ images_0.png filter=lfs diff=lfs merge=lfs -text
37
+ images_1.png filter=lfs diff=lfs merge=lfs -text
38
+ images_2.png filter=lfs diff=lfs merge=lfs -text
39
+ images_3.png filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ library_name: diffusers
4
+ tags:
5
+ - stable-diffusion
6
+ - stable-diffusion-diffusers
7
+ - text-to-image
8
+ - diffusers
9
+ - controlnet
10
+ - diffusers-training
11
+ base_model: runwayml/stable-diffusion-v1-5
12
+ inference: true
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the training script had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+
19
+ # controlnet-EpsilonGreedy/Clothes2PersonCL1e-050.5
20
+
21
+ These are controlnet weights trained on runwayml/stable-diffusion-v1-5 with new type of conditioning.
22
+ You can find some example images below.
23
+
24
+ prompt: C2PDress, a woman, refracted line and sparkles, bold graphics, suki, barcodes, gliter, round-cropped, flat texture, galactic, palladium, lower part of face, stands in front of a white background
25
+ ![images_0)](./images_0.png)
26
+ prompt: C2PUpperBody, a woman wearing black leather skirt, anthro, minimalist photorealist, t-top, pelisse, a new, features between french, inspired by James Baynes, flat icon, oppai, cut out, very ahestetic, stands in front of a white background
27
+ ![images_1)](./images_1.png)
28
+ prompt: C2PLowerBody, a woman in a black top with a black jacket, elegant cape, wearing a cropped top, 4 arms, posh, cloth sim, inspired by Rose ONeill, listing image, proportions off, bubblegum, lower part of face, stands in front of a white background
29
+ ![images_2)](./images_2.png)
30
+ prompt: C2PUpperBody, a man in tan shorts standing in front of a white wall, discord moderator, round-cropped, summer vibrance, simplified, dividing it into nine quarters, sculls, scandinavian style, nbc, lower part of face, stands in front of a white background
31
+ ![images_3)](./images_3.png)
32
+
33
+
34
+
35
+ ## Intended uses & limitations
36
+
37
+ #### How to use
38
+
39
+ ```python
40
+ # TODO: add an example code snippet for running this diffusion pipeline
41
+ ```
42
+
43
+ #### Limitations and bias
44
+
45
+ [TODO: provide examples of latent issues and potential remediations]
46
+
47
+ ## Training details
48
+
49
+ [TODO: describe the data used to train the model]
checkpoint-2500/controlnet/config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "ControlNetModel",
3
+ "_diffusers_version": "0.30.0.dev0",
4
+ "act_fn": "silu",
5
+ "addition_embed_type": null,
6
+ "addition_embed_type_num_heads": 64,
7
+ "addition_time_embed_dim": null,
8
+ "attention_head_dim": 8,
9
+ "block_out_channels": [
10
+ 320,
11
+ 640,
12
+ 1280,
13
+ 1280
14
+ ],
15
+ "class_embed_type": null,
16
+ "conditioning_channels": 3,
17
+ "conditioning_embedding_out_channels": [
18
+ 16,
19
+ 32,
20
+ 96,
21
+ 256
22
+ ],
23
+ "controlnet_conditioning_channel_order": "rgb",
24
+ "cross_attention_dim": 768,
25
+ "down_block_types": [
26
+ "CrossAttnDownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "CrossAttnDownBlock2D",
29
+ "DownBlock2D"
30
+ ],
31
+ "downsample_padding": 1,
32
+ "encoder_hid_dim": null,
33
+ "encoder_hid_dim_type": null,
34
+ "flip_sin_to_cos": true,
35
+ "freq_shift": 0,
36
+ "global_pool_conditions": false,
37
+ "in_channels": 4,
38
+ "layers_per_block": 2,
39
+ "mid_block_scale_factor": 1,
40
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
41
+ "norm_eps": 1e-05,
42
+ "norm_num_groups": 32,
43
+ "num_attention_heads": null,
44
+ "num_class_embeds": null,
45
+ "only_cross_attention": false,
46
+ "projection_class_embeddings_input_dim": null,
47
+ "resnet_time_scale_shift": "default",
48
+ "transformer_layers_per_block": 1,
49
+ "upcast_attention": false,
50
+ "use_linear_projection": false
51
+ }
checkpoint-2500/controlnet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af7d333deef3f95935ec9ef7ed3c0a3301089e9dd6fe102b5e9d5cd184c88417
3
+ size 1445157120
checkpoint-2500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cabf6315ba84eafa56ac551680fea1ec45ad3b6026f9e25bb7d786198de15924
3
+ size 725253818
checkpoint-2500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceace5b2a1414cc4b43410b35669b56c2db90e234f7b082d5a6aa6e57eef93e6
3
+ size 14604
checkpoint-2500/scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:920333155b08e76154b89a85170c320833074db7b84649a9cfe79fd10b6bf2d9
3
+ size 1000
config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "ControlNetModel",
3
+ "_diffusers_version": "0.30.0.dev0",
4
+ "act_fn": "silu",
5
+ "addition_embed_type": null,
6
+ "addition_embed_type_num_heads": 64,
7
+ "addition_time_embed_dim": null,
8
+ "attention_head_dim": 8,
9
+ "block_out_channels": [
10
+ 320,
11
+ 640,
12
+ 1280,
13
+ 1280
14
+ ],
15
+ "class_embed_type": null,
16
+ "conditioning_channels": 3,
17
+ "conditioning_embedding_out_channels": [
18
+ 16,
19
+ 32,
20
+ 96,
21
+ 256
22
+ ],
23
+ "controlnet_conditioning_channel_order": "rgb",
24
+ "cross_attention_dim": 768,
25
+ "down_block_types": [
26
+ "CrossAttnDownBlock2D",
27
+ "CrossAttnDownBlock2D",
28
+ "CrossAttnDownBlock2D",
29
+ "DownBlock2D"
30
+ ],
31
+ "downsample_padding": 1,
32
+ "encoder_hid_dim": null,
33
+ "encoder_hid_dim_type": null,
34
+ "flip_sin_to_cos": true,
35
+ "freq_shift": 0,
36
+ "global_pool_conditions": false,
37
+ "in_channels": 4,
38
+ "layers_per_block": 2,
39
+ "mid_block_scale_factor": 1,
40
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
41
+ "norm_eps": 1e-05,
42
+ "norm_num_groups": 32,
43
+ "num_attention_heads": null,
44
+ "num_class_embeds": null,
45
+ "only_cross_attention": false,
46
+ "projection_class_embeddings_input_dim": null,
47
+ "resnet_time_scale_shift": "default",
48
+ "transformer_layers_per_block": 1,
49
+ "upcast_attention": false,
50
+ "use_linear_projection": false
51
+ }
diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afe7d92c06e4e649ac313153ef68a8c9c79471b0b0cbf830894d46f8fb9a0ad3
3
+ size 1445157120
image_control.png ADDED
images_0.png ADDED

Git LFS Details

  • SHA256: a5f7f674f64135560babebc070102f09f94e9a927da488ac4eedd45f5b1dfb23
  • Pointer size: 132 Bytes
  • Size of remote file: 3.19 MB
images_1.png ADDED

Git LFS Details

  • SHA256: 4a6a9555738c411d00b34a5b63754497b8d53c4471561de9f3cd80001c07918e
  • Pointer size: 132 Bytes
  • Size of remote file: 2.9 MB
images_2.png ADDED

Git LFS Details

  • SHA256: f30f493d0b8544145fccbcdaf2696e91a97f8ffa2ba75bfb7f5914a743235f97
  • Pointer size: 132 Bytes
  • Size of remote file: 2.31 MB
images_3.png ADDED

Git LFS Details

  • SHA256: 38bfcd55e90d39e872f735b8ec8455377977f12cc53eed1b778859422baf9c24
  • Pointer size: 132 Bytes
  • Size of remote file: 2.67 MB
logs/train_controlnet/1720256327.7865648/events.out.tfevents.1720256327.4e51e2d133f2.111.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ca7301f789ad7135a385d696244fa3c0ffc2996fd75fc714397e0784be4482b
3
+ size 2470
logs/train_controlnet/1720256327.7889934/hparams.yml ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ adam_beta1: 0.9
2
+ adam_beta2: 0.999
3
+ adam_epsilon: 1.0e-08
4
+ adam_weight_decay: 0.01
5
+ allow_tf32: false
6
+ cache_dir: null
7
+ caption_column: text
8
+ checkpointing_steps: 2500
9
+ checkpoints_total_limit: null
10
+ conditioning_image_column: conditioning_image
11
+ controlnet_model_name_or_path: null
12
+ dataloader_num_workers: 4
13
+ dataset_config_name: null
14
+ dataset_name: EpsilonGreedy/Clothes2Clothes
15
+ enable_xformers_memory_efficient_attention: false
16
+ gradient_accumulation_steps: 4
17
+ gradient_checkpointing: true
18
+ hub_model_id: EpsilonGreedy/Clothes2PersonCL1e-050.5
19
+ hub_token: null
20
+ image_column: image
21
+ learning_rate: 1.0e-05
22
+ logging_dir: logs
23
+ lr_num_cycles: 1
24
+ lr_power: 1.0
25
+ lr_scheduler: constant_with_warmup
26
+ lr_warmup_steps: 500
27
+ max_grad_norm: 1.0
28
+ max_train_samples: null
29
+ max_train_steps: 2812
30
+ mixed_precision: null
31
+ num_train_epochs: 1
32
+ num_validation_images: 4
33
+ output_dir: /kaggle/working/output/Clothes2PersonCL
34
+ pretrained_model_name_or_path: runwayml/stable-diffusion-v1-5
35
+ proportion_empty_prompts: 0.5
36
+ push_to_hub: true
37
+ report_to: tensorboard
38
+ resolution: 512
39
+ resume_from_checkpoint: latest
40
+ revision: null
41
+ scale_lr: false
42
+ seed: null
43
+ set_grads_to_none: false
44
+ tokenizer_name: null
45
+ tracker_project_name: train_controlnet
46
+ train_batch_size: 3
47
+ train_data_dir: null
48
+ use_8bit_adam: true
49
+ validation_steps: 2500
50
+ variant: null
logs/train_controlnet/events.out.tfevents.1720256327.4e51e2d133f2.111.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:556d4a7f47ae611d79ec50583ad6405446715b4985a52cc27ccda32a6a3139b7
3
+ size 22861679