Upload Kawai Diffusion V31
Browse files- .gitattributes +1 -0
- README.md +6 -21
- asset/preview.png +0 -0
- model_index.json +32 -32
- scheduler/scheduler_config.json +6 -3
- text_encoder/config.json +1 -1
- text_encoder/model.safetensors +2 -2
- tokenizer/tokenizer_config.json +0 -1
- unet/config.json +1 -1
- unet/diffusion_pytorch_model.safetensors +2 -2
- vae/config.json +1 -1
.gitattributes
CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
35 |
+
asset/preview.png filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,27 +1,11 @@
|
|
1 |
-
---
|
2 |
-
license: creativeml-openrail-m
|
3 |
-
language:
|
4 |
-
- en
|
5 |
-
library_name: diffusers
|
6 |
-
tags:
|
7 |
-
- pytorch
|
8 |
-
- art
|
9 |
-
- anime
|
10 |
-
- text-to-image
|
11 |
-
---
|
12 |
-
|
13 |
-
# Kawai Diffusion (anime-base) v3.0 LTS Big Update (≧∇≦)ノ
|
14 |
-
**Waifumake from now on will be Kawai Diffusion.**
|
15 |
|
|
|
16 |
See more in CivitAI : https://civitai.com/models/21138/kawai-diffusion-sd15
|
17 |
![](asset/preview.png)
|
18 |
-
## What's new in Kawai v3.
|
19 |
-
-
|
20 |
-
|
21 |
-
-
|
22 |
-
- True "kawaii"... Haizzzzzzzzz
|
23 |
-
- Two versions: the [ema-only](kawaiDiffusionSD15_v29betaGuardHotfix.ckpt) model (4.1GB), and [pruned model](kawai-anime-sd-pruned_guard.safetensors) (7.7GB). Come on, don't be surprised by it, even I was surprised.
|
24 |
-
- Can work on some VAE. But the pruned model does not require any VAE.
|
25 |
|
26 |
## Introduction:
|
27 |
- It's an AI art model for converting text to images, images to images, inpainting, and outpainting using Stable Diffusion.
|
@@ -90,6 +74,7 @@ This wouldn't have happened if they hadn't made a breakthrough.
|
|
90 |
- [d8ahazard](https://github.com/d8ahazard/.sd_dreambooth_extension) : Dreambooth.
|
91 |
- [Automatic1111](https://github.com/AUTOMATIC1111/) : Web UI.
|
92 |
- [Mikubill](https://github.com/Mikubill/): Where my ideas started.
|
|
|
93 |
- Chat-GPT: Help me do crazy things that I thought I would never do.
|
94 |
- Novel AI, Anything Model, Abyss Orange Model: Dataset images. An AI made me thousands of pictures without worrying about copyright or dispute.
|
95 |
- Danbooru: Help me write the correct tag.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
|
2 |
+
# Kawai Diffusion (anime-base) v3.1 (≧∇≦)ノ
|
3 |
See more in CivitAI : https://civitai.com/models/21138/kawai-diffusion-sd15
|
4 |
![](asset/preview.png)
|
5 |
+
## What's new in Kawai v3.1:
|
6 |
+
- Inherit the best part of the deleted old Kawai v3.0 LTS version, add a little "grapefruit", avoid "grapefruit" too small. I don't want it to be deleted again because of the "airport". :v
|
7 |
+
|
8 |
+
- Holy "kawaii desu~"
|
|
|
|
|
|
|
9 |
|
10 |
## Introduction:
|
11 |
- It's an AI art model for converting text to images, images to images, inpainting, and outpainting using Stable Diffusion.
|
|
|
74 |
- [d8ahazard](https://github.com/d8ahazard/.sd_dreambooth_extension) : Dreambooth.
|
75 |
- [Automatic1111](https://github.com/AUTOMATIC1111/) : Web UI.
|
76 |
- [Mikubill](https://github.com/Mikubill/): Where my ideas started.
|
77 |
+
- Guard: It... is... a... secret.... Don't worry about it, Guard is just something we made for fun, but it does exist in the model. If you know a bit about safetensor, you can actually read it.
|
78 |
- Chat-GPT: Help me do crazy things that I thought I would never do.
|
79 |
- Novel AI, Anything Model, Abyss Orange Model: Dataset images. An AI made me thousands of pictures without worrying about copyright or dispute.
|
80 |
- Danbooru: Help me write the correct tag.
|
asset/preview.png
CHANGED
Git LFS Details
|
model_index.json
CHANGED
@@ -1,33 +1,33 @@
|
|
1 |
{
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
}
|
|
|
1 |
{
|
2 |
+
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.14.0",
|
4 |
+
"feature_extractor": [
|
5 |
+
null,
|
6 |
+
null
|
7 |
+
],
|
8 |
+
"requires_safety_checker": null,
|
9 |
+
"safety_checker": [
|
10 |
+
null,
|
11 |
+
null
|
12 |
+
],
|
13 |
+
"scheduler": [
|
14 |
+
"diffusers",
|
15 |
+
"DEISMultistepScheduler"
|
16 |
+
],
|
17 |
+
"text_encoder": [
|
18 |
+
"transformers",
|
19 |
+
"CLIPTextModel"
|
20 |
+
],
|
21 |
+
"tokenizer": [
|
22 |
+
"transformers",
|
23 |
+
"CLIPTokenizer"
|
24 |
+
],
|
25 |
+
"unet": [
|
26 |
+
"diffusers",
|
27 |
+
"UNet2DConditionModel"
|
28 |
+
],
|
29 |
+
"vae": [
|
30 |
+
"diffusers",
|
31 |
+
"AutoencoderKL"
|
32 |
+
]
|
33 |
+
}
|
scheduler/scheduler_config.json
CHANGED
@@ -1,16 +1,19 @@
|
|
1 |
{
|
2 |
-
"_class_name": "
|
3 |
-
"_diffusers_version": "0.
|
|
|
4 |
"beta_end": 0.012,
|
5 |
"beta_schedule": "scaled_linear",
|
6 |
"beta_start": 0.00085,
|
7 |
"clip_sample": false,
|
8 |
-
"clip_sample_range": 1.0,
|
9 |
"dynamic_thresholding_ratio": 0.995,
|
|
|
10 |
"num_train_timesteps": 1000,
|
11 |
"prediction_type": "epsilon",
|
12 |
"sample_max_value": 1.0,
|
13 |
"set_alpha_to_one": false,
|
|
|
|
|
14 |
"steps_offset": 1,
|
15 |
"thresholding": false,
|
16 |
"trained_betas": null
|
|
|
1 |
{
|
2 |
+
"_class_name": "DEISMultistepScheduler",
|
3 |
+
"_diffusers_version": "0.14.0",
|
4 |
+
"algorithm_type": "deis",
|
5 |
"beta_end": 0.012,
|
6 |
"beta_schedule": "scaled_linear",
|
7 |
"beta_start": 0.00085,
|
8 |
"clip_sample": false,
|
|
|
9 |
"dynamic_thresholding_ratio": 0.995,
|
10 |
+
"lower_order_final": true,
|
11 |
"num_train_timesteps": 1000,
|
12 |
"prediction_type": "epsilon",
|
13 |
"sample_max_value": 1.0,
|
14 |
"set_alpha_to_one": false,
|
15 |
+
"solver_order": 2,
|
16 |
+
"solver_type": "logrho",
|
17 |
"steps_offset": 1,
|
18 |
"thresholding": false,
|
19 |
"trained_betas": null
|
text_encoder/config.json
CHANGED
@@ -20,6 +20,6 @@
|
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 768,
|
22 |
"torch_dtype": "float32",
|
23 |
-
"transformers_version": "4.
|
24 |
"vocab_size": 49408
|
25 |
}
|
|
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 768,
|
22 |
"torch_dtype": "float32",
|
23 |
+
"transformers_version": "4.27.3",
|
24 |
"vocab_size": 49408
|
25 |
}
|
text_encoder/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d934a4375a0b0b22facfc479e5a8b81be37db59e7eb333f100034740fbbab36a
|
3 |
+
size 492265879
|
tokenizer/tokenizer_config.json
CHANGED
@@ -19,7 +19,6 @@
|
|
19 |
},
|
20 |
"errors": "replace",
|
21 |
"model_max_length": 77,
|
22 |
-
"name_or_path": "openai/clip-vit-large-patch14",
|
23 |
"pad_token": "<|endoftext|>",
|
24 |
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
"tokenizer_class": "CLIPTokenizer",
|
|
|
19 |
},
|
20 |
"errors": "replace",
|
21 |
"model_max_length": 77,
|
|
|
22 |
"pad_token": "<|endoftext|>",
|
23 |
"special_tokens_map_file": "./special_tokens_map.json",
|
24 |
"tokenizer_class": "CLIPTokenizer",
|
unet/config.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
-
"_diffusers_version": "0.
|
4 |
"act_fn": "silu",
|
5 |
"attention_head_dim": 8,
|
6 |
"block_out_channels": [
|
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
+
"_diffusers_version": "0.14.0",
|
4 |
"act_fn": "silu",
|
5 |
"attention_head_dim": 8,
|
6 |
"block_out_channels": [
|
unet/diffusion_pytorch_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9bda43092a179100c2f1d18e2a17353952c60ae72826955b8615d901a236fc42
|
3 |
+
size 3438167534
|
vae/config.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
-
"_diffusers_version": "0.
|
4 |
"act_fn": "silu",
|
5 |
"block_out_channels": [
|
6 |
128,
|
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
+
"_diffusers_version": "0.14.0",
|
4 |
"act_fn": "silu",
|
5 |
"block_out_channels": [
|
6 |
128,
|