snow-leopard commited on
Commit
5f47b7b
·
1 Parent(s): 5f768b8
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. checkpoint-1000/optimizer.bin +3 -0
  2. checkpoint-1000/random_states_0.pkl +3 -0
  3. checkpoint-1000/unet/config.json +21 -0
  4. checkpoint-1000/unet/diffusion_pytorch_model.bin +3 -0
  5. checkpoint-1000/unet_ema/config.json +30 -0
  6. checkpoint-1000/unet_ema/diffusion_pytorch_model.bin +3 -0
  7. checkpoint-10000/optimizer.bin +3 -0
  8. checkpoint-10000/random_states_0.pkl +3 -0
  9. checkpoint-10000/unet/config.json +21 -0
  10. checkpoint-10000/unet/diffusion_pytorch_model.bin +3 -0
  11. checkpoint-10000/unet_ema/config.json +30 -0
  12. checkpoint-10000/unet_ema/diffusion_pytorch_model.bin +3 -0
  13. checkpoint-10500/optimizer.bin +3 -0
  14. checkpoint-10500/random_states_0.pkl +3 -0
  15. checkpoint-10500/unet/config.json +21 -0
  16. checkpoint-10500/unet/diffusion_pytorch_model.bin +3 -0
  17. checkpoint-10500/unet_ema/config.json +30 -0
  18. checkpoint-10500/unet_ema/diffusion_pytorch_model.bin +3 -0
  19. checkpoint-11000/optimizer.bin +3 -0
  20. checkpoint-11000/random_states_0.pkl +3 -0
  21. checkpoint-11000/unet/config.json +21 -0
  22. checkpoint-11000/unet/diffusion_pytorch_model.bin +3 -0
  23. checkpoint-11000/unet_ema/config.json +30 -0
  24. checkpoint-11000/unet_ema/diffusion_pytorch_model.bin +3 -0
  25. checkpoint-11500/optimizer.bin +3 -0
  26. checkpoint-11500/random_states_0.pkl +3 -0
  27. checkpoint-11500/unet/config.json +21 -0
  28. checkpoint-11500/unet/diffusion_pytorch_model.bin +3 -0
  29. checkpoint-11500/unet_ema/config.json +30 -0
  30. checkpoint-11500/unet_ema/diffusion_pytorch_model.bin +3 -0
  31. checkpoint-12000/optimizer.bin +3 -0
  32. checkpoint-12000/random_states_0.pkl +3 -0
  33. checkpoint-12000/unet/config.json +21 -0
  34. checkpoint-12000/unet/diffusion_pytorch_model.bin +3 -0
  35. checkpoint-12000/unet_ema/config.json +30 -0
  36. checkpoint-12000/unet_ema/diffusion_pytorch_model.bin +3 -0
  37. checkpoint-12500/optimizer.bin +3 -0
  38. checkpoint-12500/random_states_0.pkl +3 -0
  39. checkpoint-12500/unet/config.json +21 -0
  40. checkpoint-12500/unet/diffusion_pytorch_model.bin +3 -0
  41. checkpoint-12500/unet_ema/config.json +30 -0
  42. checkpoint-12500/unet_ema/diffusion_pytorch_model.bin +3 -0
  43. checkpoint-13000/optimizer.bin +3 -0
  44. checkpoint-13000/random_states_0.pkl +3 -0
  45. checkpoint-13000/unet/config.json +21 -0
  46. checkpoint-13000/unet/diffusion_pytorch_model.bin +3 -0
  47. checkpoint-13000/unet_ema/config.json +30 -0
  48. checkpoint-13000/unet_ema/diffusion_pytorch_model.bin +3 -0
  49. checkpoint-13500/optimizer.bin +3 -0
  50. checkpoint-13500/random_states_0.pkl +3 -0
checkpoint-1000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f0322b5888139ba00a4a41e45908f48b88e2bc4f848e0921ad6e94889ee66c0
3
+ size 2411350917
checkpoint-1000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c1105e2f308222cc8d3100b64feeb8ef8dd52c41c9d3ff101689419fc1d3aec
3
+ size 14663
checkpoint-1000/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-1000/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-1000/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 1000,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-1000/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6c171711b28723be190778ea5147b3509fa144074ac38fbb63c161a487fb4eb
3
+ size 1205673685
checkpoint-10000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86f9bbfcb0b5c6cf926251bc8f01c28ebdcc7bedfbeb8cb957f1d9f60133c134
3
+ size 2411350917
checkpoint-10000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e220b668168dcdf8d63f73bebd1d2727b7216494653990ee03818930f0deaf70
3
+ size 14663
checkpoint-10000/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-10000/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-10000/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 10000,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-10000/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1727c0c596b9772c999fb8f63f0372cbf11e041ffb1361490ae4595878f73fcc
3
+ size 1205673685
checkpoint-10500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d7bcf85b4ed43a187a644369fba99d63ac40ac8738a1b3cd17cf4ea9aa3a5fe
3
+ size 2411350917
checkpoint-10500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:682adc56d83498cbfb5178a4727df2281a426fa5037cc9dff82dea283b0ba19d
3
+ size 14663
checkpoint-10500/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-10500/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-10500/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 10500,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-10500/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a2851412c563e03c1000b237729e3a5464b653a0fc8de509590f50e0d71d53a
3
+ size 1205673685
checkpoint-11000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8901f284acf7e776684a68a0ae5faa7a9c6188af9fa1de8274487fe52697417
3
+ size 2411350917
checkpoint-11000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4d302b004222cacbbefb8820cc2443e33da75615652cdb67ef109705c6059a7
3
+ size 14663
checkpoint-11000/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-11000/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-11000/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 11000,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-11000/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:753c1c7f192d084f47a4dca3c521e262793b01bddeb7ac4331c3b612e7ee9c49
3
+ size 1205673685
checkpoint-11500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68772260b3eeadf3470108661d13b25403fed3698d146515f287f0fab7142ff9
3
+ size 2411350917
checkpoint-11500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76ce6ddb7918a19ac947c02441ae4bb08eaf82006be20699e49e0f9bfbb455a6
3
+ size 14663
checkpoint-11500/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-11500/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-11500/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 11500,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-11500/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa3e70a4b7d08e4241cb078ff07768bbdfe1df26f92a8ff093383a0d8d779eb1
3
+ size 1205673685
checkpoint-12000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6248432f7535de46bb3e5eaf1b88ecdde3a562e2305ca6ff5d342d52d8366299
3
+ size 2411350917
checkpoint-12000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7a147f4402f1155ffff434f21bfe91d79e1b80b2116f6a338cdb17d5f5dc637
3
+ size 14663
checkpoint-12000/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-12000/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-12000/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 12000,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-12000/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:487761adf46217a21e6b51fe1db18b7acad256e3885195624589b4d70fd0a0e6
3
+ size 1205673685
checkpoint-12500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:500e14b20b399b8580ff0caab6078940ab41a15e402d845ea3534a2c2a79aca4
3
+ size 2411350917
checkpoint-12500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69c439b2cc2c167b1be19b7ec911955a492aedf9a1a360c7f058c6d1e4d537f1
3
+ size 14663
checkpoint-12500/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-12500/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-12500/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 12500,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-12500/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00fe1d7950e889a3b18dc37b46e8f40d0ecb034e66106009d2b3373e0fce4fcc
3
+ size 1205673685
checkpoint-13000/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e4cd4a6d12f557fcde39293089e1b30acc7c5d94b46b800e5b4e8db84b6d95c
3
+ size 2411350917
checkpoint-13000/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b57a912f09f0a184e1f525e09059a3c46a63c47cd2878d2b122fcb6c08c08ea1
3
+ size 14663
checkpoint-13000/unet/config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "act_fn": "silu",
5
+ "block_out_channels": [
6
+ 64
7
+ ],
8
+ "down_block_types": [
9
+ "DownEncoderBlock2D"
10
+ ],
11
+ "in_channels": 3,
12
+ "latent_channels": 3,
13
+ "layers_per_block": 1,
14
+ "norm_num_groups": 32,
15
+ "out_channels": 3,
16
+ "sample_size": 32,
17
+ "scaling_factor": 0.18215,
18
+ "up_block_types": [
19
+ "UpDecoderBlock2D"
20
+ ]
21
+ }
checkpoint-13000/unet/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b44e2fb6d05620dd6b3c1de9d1bdb31bfef166c6a6d4f3ad47d3a7d64cebd80d
3
+ size 2281457
checkpoint-13000/unet_ema/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "Transformer2DModel",
3
+ "_diffusers_version": "0.13.0.dev0",
4
+ "activation_fn": "geglu",
5
+ "attention_bias": false,
6
+ "attention_head_dim": 88,
7
+ "cross_attention_dim": null,
8
+ "decay": 0.9999,
9
+ "dropout": 0.0,
10
+ "in_channels": 3,
11
+ "inv_gamma": 1.0,
12
+ "min_decay": 0.0,
13
+ "norm_elementwise_affine": true,
14
+ "norm_num_groups": 3,
15
+ "norm_type": "ada_norm",
16
+ "num_attention_heads": 16,
17
+ "num_embeds_ada_norm": 1000,
18
+ "num_layers": 8,
19
+ "num_vector_embeds": null,
20
+ "only_cross_attention": false,
21
+ "optimization_step": 13000,
22
+ "out_channels": 3,
23
+ "patch_size": 8,
24
+ "power": 0.75,
25
+ "sample_size": 64,
26
+ "upcast_attention": false,
27
+ "update_after_step": 0,
28
+ "use_ema_warmup": true,
29
+ "use_linear_projection": false
30
+ }
checkpoint-13000/unet_ema/diffusion_pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b256e206b1d4d8bfeb893ad615df55b6bbdc33bcbda10b455af1e257b71f258b
3
+ size 1205673685
checkpoint-13500/optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86eb266f10ccfae17171c9f64bcb0716627a0d0ea2b5fd6abeec9e23d6fb9448
3
+ size 2411350917
checkpoint-13500/random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87b8437bc2aa9f60f41fcc1f43e3779b9003a3ada1265433254221945fe96693
3
+ size 14663