ManglerFTW commited on
Commit
e6a19c9
1 Parent(s): 23faa14

Upload 13 files

Browse files
args.json ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "shuffle_per_epoch": false,
3
+ "attention": "xformers",
4
+ "model_variant": "base",
5
+ "aspect_mode": "dynamic",
6
+ "aspect_mode_action_preference": "truncate",
7
+ "use_ema": true,
8
+ "clip_penultimate": false,
9
+ "conditional_dropout": null,
10
+ "disable_cudnn_benchmark": true,
11
+ "use_text_files_as_captions": true,
12
+ "sample_from_batch": 0,
13
+ "flatten_sample_folder": false,
14
+ "stop_text_encoder_training": 30,
15
+ "use_bucketing": true,
16
+ "regenerate_latent_cache": true,
17
+ "sample_on_training_start": false,
18
+ "add_class_images_to_dataset": false,
19
+ "auto_balance_concept_datasets": false,
20
+ "sample_aspect_ratios": true,
21
+ "dataset_repeats": 1,
22
+ "save_every_n_epoch": 5,
23
+ "pretrained_model_name_or_path": "input_model/V5E20",
24
+ "pretrained_vae_name_or_path": "",
25
+ "tokenizer_name": null,
26
+ "instance_data_dir": null,
27
+ "class_data_dir": null,
28
+ "instance_prompt": null,
29
+ "class_prompt": null,
30
+ "save_sample_prompt": null,
31
+ "n_save_sample": 1,
32
+ "sample_height": 768,
33
+ "sample_width": 768,
34
+ "save_guidance_scale": 7.5,
35
+ "save_infer_steps": 30,
36
+ "with_prior_preservation": false,
37
+ "prior_loss_weight": 1.0,
38
+ "num_class_images": 0,
39
+ "output_dir": "output/",
40
+ "seed": 987654321,
41
+ "resolution": 768,
42
+ "center_crop": false,
43
+ "train_text_encoder": true,
44
+ "train_batch_size": 10,
45
+ "sample_batch_size": 4,
46
+ "num_train_epochs": 35,
47
+ "max_train_steps": 17150,
48
+ "gradient_accumulation_steps": 1,
49
+ "gradient_checkpointing": false,
50
+ "learning_rate": 3e-06,
51
+ "scale_lr": false,
52
+ "lr_scheduler": "cosine",
53
+ "lr_warmup_steps": 0,
54
+ "use_8bit_adam": true,
55
+ "adam_beta1": 0.9,
56
+ "adam_beta2": 0.999,
57
+ "adam_weight_decay": 0.01,
58
+ "adam_epsilon": 1e-08,
59
+ "max_grad_norm": 1.0,
60
+ "push_to_hub": false,
61
+ "hub_token": null,
62
+ "hub_model_id": null,
63
+ "logging_dir": "logs",
64
+ "log_interval": 10,
65
+ "sample_step_interval": 500,
66
+ "mixed_precision": "fp16",
67
+ "local_rank": -1,
68
+ "concepts_list": [
69
+ {
70
+ "instance_prompt": "Photography",
71
+ "instance_data_dir": "datasets/Photographs",
72
+ "class_prompt": "",
73
+ "class_data_dir": "",
74
+ "do_not_balance": 0,
75
+ "use_sub_dirs": 0
76
+ },
77
+ {
78
+ "instance_prompt": "Drawn",
79
+ "instance_data_dir": "datasets/Drawn",
80
+ "class_prompt": "",
81
+ "class_data_dir": "",
82
+ "do_not_balance": 0,
83
+ "use_sub_dirs": 0
84
+ }
85
+ ],
86
+ "save_sample_controlled_seed": [
87
+ 3434554,
88
+ 1234567890
89
+ ],
90
+ "detect_full_drive": true,
91
+ "send_telegram_updates": false,
92
+ "telegram_chat_id": "0",
93
+ "telegram_token": "0",
94
+ "use_deepspeed_adam": false,
95
+ "append_sample_controlled_seed_action": null,
96
+ "add_sample_prompt": [
97
+ "CHV3SDark, a dark forest filled with green trees and bushes, eerie jungle, ominous and eerie forest, dark forest and trees, gloomy forest, in the middle of dark forest, dark forest in the background, spooky forest, dark forest in background",
98
+ "CHV3SEldritch, an image of a creature that is in the water, serpentine water monster, lovecraftian creature, eldritch being, eldritch god, looming creature with a long, lovecraftian monster, eldritch abomination, creature concept art",
99
+ "CHV3CUrban, a young man sitting on the ground with an old fashioned boombox in his lap, holding it up to his face and looking at the camera while he's leaning against a metal fence behind him is wearing blue denim jeans "
100
+ ],
101
+ "use_image_names_as_captions": false,
102
+ "masked_training": false,
103
+ "normalize_masked_area_loss": false,
104
+ "unmasked_probability": 0.0,
105
+ "max_denoising_strength": 1.0,
106
+ "mask_prompts": null,
107
+ "batch_tokens": null
108
+ }
model_index.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.10.2",
4
+ "feature_extractor": [
5
+ null,
6
+ null
7
+ ],
8
+ "requires_safety_checker": null,
9
+ "safety_checker": [
10
+ null,
11
+ null
12
+ ],
13
+ "scheduler": [
14
+ "diffusers",
15
+ "DDIMScheduler"
16
+ ],
17
+ "text_encoder": [
18
+ "transformers",
19
+ "CLIPTextModel"
20
+ ],
21
+ "tokenizer": [
22
+ "transformers",
23
+ "CLIPTokenizer"
24
+ ],
25
+ "unet": [
26
+ "diffusers",
27
+ "UNet2DConditionModel"
28
+ ],
29
+ "vae": [
30
+ "diffusers",
31
+ "AutoencoderKL"
32
+ ]
33
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "DPMSolverMultistepScheduler",
3
+ "_diffusers_version": "0.10.2",
4
+ "algorithm_type": "dpmsolver++",
5
+ "beta_end": 0.012,
6
+ "beta_schedule": "scaled_linear",
7
+ "beta_start": 0.00085,
8
+ "clip_sample": false,
9
+ "dynamic_thresholding_ratio": 0.995,
10
+ "lower_order_final": true,
11
+ "num_train_timesteps": 1000,
12
+ "prediction_type": "v_prediction",
13
+ "sample_max_value": 1.0,
14
+ "set_alpha_to_one": false,
15
+ "skip_prk_steps": true,
16
+ "solver_order": 2,
17
+ "solver_type": "midpoint",
18
+ "steps_offset": 1,
19
+ "thresholding": false,
20
+ "trained_betas": null
21
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "input_model/V5E20",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_size": 1024,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 23,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 512,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.26.1",
24
+ "vocab_size": 49408
25
+ }
text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b85d9357cb60bb18ae1971733b69d17b88ae85b85e6fcdfc3a07216017d2ab2b
3
+ size 1361597018
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "!",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": {
4
+ "__type": "AddedToken",
5
+ "content": "<|startoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "do_lower_case": true,
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
+ "model_max_length": 77,
22
+ "name_or_path": "input_model/V5E20/tokenizer",
23
+ "pad_token": "<|endoftext|>",
24
+ "special_tokens_map_file": "./special_tokens_map.json",
25
+ "tokenizer_class": "CLIPTokenizer",
26
+ "unk_token": {
27
+ "__type": "AddedToken",
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.10.2",
4
+ "_name_or_path": "input_model/V5E20",
5
+ "act_fn": "silu",
6
+ "attention_head_dim": [
7
+ 5,
8
+ 10,
9
+ 20,
10
+ 20
11
+ ],
12
+ "block_out_channels": [
13
+ 320,
14
+ 640,
15
+ 1280,
16
+ 1280
17
+ ],
18
+ "center_input_sample": false,
19
+ "class_embed_type": null,
20
+ "cross_attention_dim": 1024,
21
+ "down_block_types": [
22
+ "CrossAttnDownBlock2D",
23
+ "CrossAttnDownBlock2D",
24
+ "CrossAttnDownBlock2D",
25
+ "DownBlock2D"
26
+ ],
27
+ "downsample_padding": 1,
28
+ "dual_cross_attention": false,
29
+ "flip_sin_to_cos": true,
30
+ "freq_shift": 0,
31
+ "in_channels": 4,
32
+ "layers_per_block": 2,
33
+ "mid_block_scale_factor": 1,
34
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
35
+ "norm_eps": 1e-05,
36
+ "norm_num_groups": 32,
37
+ "num_class_embeds": null,
38
+ "only_cross_attention": false,
39
+ "out_channels": 4,
40
+ "resnet_time_scale_shift": "default",
41
+ "sample_size": 32,
42
+ "up_block_types": [
43
+ "UpBlock2D",
44
+ "CrossAttnUpBlock2D",
45
+ "CrossAttnUpBlock2D",
46
+ "CrossAttnUpBlock2D"
47
+ ],
48
+ "upcast_attention": false,
49
+ "use_linear_projection": false
50
+ }
unet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e3b947cd9207670849dddf50569a6131be8e7c2bbc59886df7d0206e02b97f2
3
+ size 3463726595
vae/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.10.2",
4
+ "_name_or_path": "input_model/V5E20",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "in_channels": 3,
19
+ "latent_channels": 4,
20
+ "layers_per_block": 2,
21
+ "norm_num_groups": 32,
22
+ "out_channels": 3,
23
+ "sample_size": 256,
24
+ "up_block_types": [
25
+ "UpDecoderBlock2D",
26
+ "UpDecoderBlock2D",
27
+ "UpDecoderBlock2D",
28
+ "UpDecoderBlock2D"
29
+ ]
30
+ }
vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d0af75d1920dd51dfb48afd760d93f8771214de769a0aaf6db88c7961ccccaa
3
+ size 334643245