JunhaoZhuang
commited on
Commit
•
a890698
1
Parent(s):
babc1f5
Upload folder using huggingface_hub
Browse files- GraySD/MultiResNetModel.bin +3 -0
- GraySD/config.json +57 -0
- GraySD/diffusion_pytorch_model.bin +3 -0
- GraySD/unet_lora.bin +3 -0
- LE/erika.pth +3 -0
- ScreenStyle/ScreenVAE/latest_net_dec.pth +3 -0
- ScreenStyle/ScreenVAE/latest_net_enc.pth +3 -0
- ScreenStyle/color2manga/color2manga_cycle_ganstft/latest_net_E.pth +3 -0
- ScreenStyle/color2manga/color2manga_cycle_ganstft/latest_net_G_INTSCR2RGB.pth +3 -0
- ScreenStyle/color2manga/color2manga_cycle_ganstft/latest_net_G_RGB2INTSCR.pth +3 -0
- ScreenStyle/color2manga/color2manga_cycle_ganstft/opt.txt +55 -0
- image_encoder/config.json +23 -0
- image_encoder/model.safetensors +3 -0
- image_encoder/pytorch_model.bin +3 -0
- sketch/MultiResNetModel.bin +3 -0
- sketch/config.json +22 -0
- sketch/diffusion_pytorch_model.bin +3 -0
- sketch/transformer_lora.bin +3 -0
GraySD/MultiResNetModel.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7f536ea6bbeb0e68d290da57231681e18e6424ed9aca86b00c041b86371beba0
|
3 |
+
size 340148769
|
GraySD/config.json
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_class_name": "ColorGuiderSDModel",
|
3 |
+
"_diffusers_version": "0.27.0.dev0",
|
4 |
+
"act_fn": "silu",
|
5 |
+
"addition_embed_type": null,
|
6 |
+
"addition_embed_type_num_heads": 64,
|
7 |
+
"addition_time_embed_dim": null,
|
8 |
+
"attention_head_dim": 8,
|
9 |
+
"block_out_channels": [
|
10 |
+
320,
|
11 |
+
640,
|
12 |
+
1280,
|
13 |
+
1280
|
14 |
+
],
|
15 |
+
"brushnet_conditioning_channel_order": "rgb",
|
16 |
+
"class_embed_type": null,
|
17 |
+
"conditioning_channels": 9,
|
18 |
+
"conditioning_embedding_out_channels": [
|
19 |
+
16,
|
20 |
+
32,
|
21 |
+
96,
|
22 |
+
256
|
23 |
+
],
|
24 |
+
"cross_attention_dim": 768,
|
25 |
+
"down_block_types": [
|
26 |
+
"SelfAttnDownBlock2D",
|
27 |
+
"SelfAttnDownBlock2D",
|
28 |
+
"SelfAttnDownBlock2D",
|
29 |
+
"DownBlock2D"
|
30 |
+
],
|
31 |
+
"downsample_padding": 1,
|
32 |
+
"encoder_hid_dim": null,
|
33 |
+
"encoder_hid_dim_type": null,
|
34 |
+
"flip_sin_to_cos": true,
|
35 |
+
"freq_shift": 0,
|
36 |
+
"global_pool_conditions": false,
|
37 |
+
"in_channels": 4,
|
38 |
+
"layers_per_block": 2,
|
39 |
+
"mid_block_scale_factor": 1,
|
40 |
+
"mid_block_type": "UNetMidBlock2DSelfAttn",
|
41 |
+
"norm_eps": 1e-05,
|
42 |
+
"norm_num_groups": 32,
|
43 |
+
"num_attention_heads": null,
|
44 |
+
"num_class_embeds": null,
|
45 |
+
"only_cross_attention": false,
|
46 |
+
"projection_class_embeddings_input_dim": null,
|
47 |
+
"resnet_time_scale_shift": "default",
|
48 |
+
"transformer_layers_per_block": 1,
|
49 |
+
"up_block_types": [
|
50 |
+
"UpBlock2D",
|
51 |
+
"SelfAttnUpBlock2D",
|
52 |
+
"SelfAttnUpBlock2D",
|
53 |
+
"SelfAttnUpBlock2D"
|
54 |
+
],
|
55 |
+
"upcast_attention": false,
|
56 |
+
"use_linear_projection": false
|
57 |
+
}
|
GraySD/diffusion_pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed8b4e17102edc10c3e55bb92ba4e575eec71107b98358dcfd6124f97bd00ee1
|
3 |
+
size 3368615172
|
GraySD/unet_lora.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4e86ccf22f1f0f89245afe0473d6a1481b8f0182fe6706e4a13826630d5c0b98
|
3 |
+
size 95863546
|
LE/erika.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:badbd6baf013cefbd98993307b02cc14a26c770d067416e4fdecc8720b88feeb
|
3 |
+
size 172789563
|
ScreenStyle/ScreenVAE/latest_net_dec.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88624a2df7318ca98f8d52dc85b33faad32776399bb0b8f34cf347ebdf0e7e0a
|
3 |
+
size 52770370
|
ScreenStyle/ScreenVAE/latest_net_enc.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d64be0ace4e2f4678a660e84986d2eb16718067858580419810cd5eb3fbadce9
|
3 |
+
size 17398368
|
ScreenStyle/color2manga/color2manga_cycle_ganstft/latest_net_E.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4b9b08fabb4af34cbe7a221a08c8682d8f15f31fd3c926d06e2bb151037d4fd
|
3 |
+
size 12037739
|
ScreenStyle/color2manga/color2manga_cycle_ganstft/latest_net_G_INTSCR2RGB.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6534e3a49a3964b1f1756e037830e95e54bae0580fc82439865b84f7f5711458
|
3 |
+
size 51937787
|
ScreenStyle/color2manga/color2manga_cycle_ganstft/latest_net_G_RGB2INTSCR.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7724992688eab80f2d6ddb94312aa30e9d1330fb2bfc7343837dbfb03aaafc97
|
3 |
+
size 50681103
|
ScreenStyle/color2manga/color2manga_cycle_ganstft/opt.txt
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
----------------- Options ---------------
|
2 |
+
aspect_ratio: 1.0
|
3 |
+
batch_size: 2
|
4 |
+
center_crop: False
|
5 |
+
checkpoints_dir: ./ckpt//ScreenStyle/color2manga/
|
6 |
+
color2screen: False
|
7 |
+
conditional_D: False
|
8 |
+
crop_size: 1024
|
9 |
+
dataroot: None
|
10 |
+
dataset_mode: aligned
|
11 |
+
direction: BtoA
|
12 |
+
display_winsize: 256
|
13 |
+
epoch: latest
|
14 |
+
eval: False
|
15 |
+
folder: intra
|
16 |
+
gpu_ids: 0
|
17 |
+
init_gain: 0.02
|
18 |
+
init_type: kaiming
|
19 |
+
input_nc: 1
|
20 |
+
isTrain: False [default: None]
|
21 |
+
level: 0
|
22 |
+
load_size: 512
|
23 |
+
local_rank: 0
|
24 |
+
max_dataset_size: inf
|
25 |
+
model: cycle_ganstft
|
26 |
+
n_samples: 1
|
27 |
+
name: color2manga_cycle_ganstft
|
28 |
+
ndf: 32
|
29 |
+
nef: 48
|
30 |
+
netC: unet_128
|
31 |
+
netD: basic_256_multi
|
32 |
+
netD2: basic_256_multi
|
33 |
+
netE: conv_256
|
34 |
+
netG: unet_256
|
35 |
+
ngf: 48
|
36 |
+
nl: prelu
|
37 |
+
no_encode: False
|
38 |
+
no_flip: True
|
39 |
+
norm: layer
|
40 |
+
num_Ds: 2
|
41 |
+
num_test: 30
|
42 |
+
num_threads: 4
|
43 |
+
nz: 64
|
44 |
+
output_nc: 3
|
45 |
+
phase: val
|
46 |
+
preprocess: none
|
47 |
+
results_dir: ../results/
|
48 |
+
serial_batches: False
|
49 |
+
suffix:
|
50 |
+
sync: False
|
51 |
+
upsample: bilinear
|
52 |
+
use_dropout: False
|
53 |
+
verbose: False
|
54 |
+
where_add: all
|
55 |
+
----------------- End -------------------
|
image_encoder/config.json
ADDED
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "./image_encoder",
|
3 |
+
"architectures": [
|
4 |
+
"CLIPVisionModelWithProjection"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"dropout": 0.0,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_size": 1280,
|
10 |
+
"image_size": 224,
|
11 |
+
"initializer_factor": 1.0,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 5120,
|
14 |
+
"layer_norm_eps": 1e-05,
|
15 |
+
"model_type": "clip_vision_model",
|
16 |
+
"num_attention_heads": 16,
|
17 |
+
"num_channels": 3,
|
18 |
+
"num_hidden_layers": 32,
|
19 |
+
"patch_size": 14,
|
20 |
+
"projection_dim": 1024,
|
21 |
+
"torch_dtype": "float16",
|
22 |
+
"transformers_version": "4.28.0.dev0"
|
23 |
+
}
|
image_encoder/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6ca9667da1ca9e0b0f75e46bb030f7e011f44f86cbfb8d5a36590fcd7507b030
|
3 |
+
size 2528373448
|
image_encoder/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3d3ec1e66737f77a4f3bc2df3c52eacefc69ce7825e2784183b1d4e9877d9193
|
3 |
+
size 2528481905
|
sketch/MultiResNetModel.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80b34ef6426842f04d7f93d76e6b9015eee297cf202c9fd4b807f5d80b79ca03
|
3 |
+
size 340148769
|
sketch/config.json
ADDED
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"num_attention_heads": 16,
|
3 |
+
"attention_head_dim": 72,
|
4 |
+
"in_channels": 4,
|
5 |
+
"cond_channels": 9,
|
6 |
+
"out_channels": 8,
|
7 |
+
"num_layers": 28,
|
8 |
+
"dropout": 0.0,
|
9 |
+
"norm_num_groups": 32,
|
10 |
+
"cross_attention_dim": 1152,
|
11 |
+
"attention_bias": true,
|
12 |
+
"sample_size": 128,
|
13 |
+
"patch_size": 2,
|
14 |
+
"activation_fn": "gelu-approximate",
|
15 |
+
"num_embeds_ada_norm": 1000,
|
16 |
+
"upcast_attention": false,
|
17 |
+
"norm_type": "ada_norm_single",
|
18 |
+
"norm_elementwise_affine": false,
|
19 |
+
"norm_eps": 1e-06,
|
20 |
+
"caption_channels": 4096,
|
21 |
+
"attention_type": "default"
|
22 |
+
}
|
sketch/diffusion_pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:316bc5f1285ac87b19850bbd1b0bf4ecc38f50da330f68f6bfa5c891e023c4cd
|
3 |
+
size 1007841568
|
sketch/transformer_lora.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bac9146550fb430829a5db3e24ac248c8b713931f9c105d66a1725a6d29798f0
|
3 |
+
size 442134080
|