Ojimi commited on
Commit
2d840e3
1 Parent(s): 065718b

Upload Kawai Diffusion V31

Browse files
.gitattributes CHANGED
@@ -32,3 +32,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ asset/preview.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,27 +1,11 @@
1
- ---
2
- license: creativeml-openrail-m
3
- language:
4
- - en
5
- library_name: diffusers
6
- tags:
7
- - pytorch
8
- - art
9
- - anime
10
- - text-to-image
11
- ---
12
-
13
- # Kawai Diffusion (anime-base) v3.0 LTS Big Update (≧∇≦)ノ
14
- **Waifumake from now on will be Kawai Diffusion.**
15
 
 
16
  See more in CivitAI : https://civitai.com/models/21138/kawai-diffusion-sd15
17
  ![](asset/preview.png)
18
- ## What's new in Kawai v3.0 LTS:
19
- - Fix color loss.
20
- - Image quality is greatly enhanced. Thank you my friend.
21
- - Kawai Diffusion's most powerful ability is "enhance" (img2img). It will make a bad photo look better.
22
- - True "kawaii"... Haizzzzzzzzz
23
- - Two versions: the [ema-only](kawaiDiffusionSD15_v29betaGuardHotfix.ckpt) model (4.1GB), and [pruned model](kawai-anime-sd-pruned_guard.safetensors) (7.7GB). Come on, don't be surprised by it, even I was surprised.
24
- - Can work on some VAE. But the pruned model does not require any VAE.
25
 
26
  ## Introduction:
27
  - It's an AI art model for converting text to images, images to images, inpainting, and outpainting using Stable Diffusion.
@@ -90,6 +74,7 @@ This wouldn't have happened if they hadn't made a breakthrough.
90
  - [d8ahazard](https://github.com/d8ahazard/.sd_dreambooth_extension) : Dreambooth.
91
  - [Automatic1111](https://github.com/AUTOMATIC1111/) : Web UI.
92
  - [Mikubill](https://github.com/Mikubill/): Where my ideas started.
 
93
  - Chat-GPT: Help me do crazy things that I thought I would never do.
94
  - Novel AI, Anything Model, Abyss Orange Model: Dataset images. An AI made me thousands of pictures without worrying about copyright or dispute.
95
  - Danbooru: Help me write the correct tag.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
 
2
+ # Kawai Diffusion (anime-base) v3.1 (≧∇≦)ノ
3
  See more in CivitAI : https://civitai.com/models/21138/kawai-diffusion-sd15
4
  ![](asset/preview.png)
5
+ ## What's new in Kawai v3.1:
6
+ - Inherit the best part of the deleted old Kawai v3.0 LTS version, add a little "grapefruit", avoid "grapefruit" too small. I don't want it to be deleted again because of the "airport". :v
7
+
8
+ - Holy "kawaii desu~"
 
 
 
9
 
10
  ## Introduction:
11
  - It's an AI art model for converting text to images, images to images, inpainting, and outpainting using Stable Diffusion.
 
74
  - [d8ahazard](https://github.com/d8ahazard/.sd_dreambooth_extension) : Dreambooth.
75
  - [Automatic1111](https://github.com/AUTOMATIC1111/) : Web UI.
76
  - [Mikubill](https://github.com/Mikubill/): Where my ideas started.
77
+ - Guard: It... is... a... secret.... Don't worry about it, Guard is just something we made for fun, but it does exist in the model. If you know a bit about safetensor, you can actually read it.
78
  - Chat-GPT: Help me do crazy things that I thought I would never do.
79
  - Novel AI, Anything Model, Abyss Orange Model: Dataset images. An AI made me thousands of pictures without worrying about copyright or dispute.
80
  - Danbooru: Help me write the correct tag.
asset/preview.png CHANGED

Git LFS Details

  • SHA256: f60b2738a233f2506a7fe5e597adebc90a02c0b4fdb059672ddd47e57d6a121d
  • Pointer size: 132 Bytes
  • Size of remote file: 1.05 MB
model_index.json CHANGED
@@ -1,33 +1,33 @@
1
  {
2
- "_class_name": "StableDiffusionPipeline",
3
- "_diffusers_version": "0.15.0.dev0",
4
- "feature_extractor": [
5
- "transformers",
6
- "CLIPFeatureExtractor"
7
- ],
8
- "requires_safety_checker": true,
9
- "safety_checker": [
10
- "stable_diffusion",
11
- "StableDiffusionSafetyChecker"
12
- ],
13
- "scheduler": [
14
- "diffusers",
15
- "DDIMScheduler"
16
- ],
17
- "text_encoder": [
18
- "transformers",
19
- "CLIPTextModel"
20
- ],
21
- "tokenizer": [
22
- "transformers",
23
- "CLIPTokenizer"
24
- ],
25
- "unet": [
26
- "diffusers",
27
- "UNet2DConditionModel"
28
- ],
29
- "vae": [
30
- "diffusers",
31
- "AutoencoderKL"
32
- ]
33
- }
 
1
  {
2
+ "_class_name": "StableDiffusionPipeline",
3
+ "_diffusers_version": "0.14.0",
4
+ "feature_extractor": [
5
+ null,
6
+ null
7
+ ],
8
+ "requires_safety_checker": null,
9
+ "safety_checker": [
10
+ null,
11
+ null
12
+ ],
13
+ "scheduler": [
14
+ "diffusers",
15
+ "DEISMultistepScheduler"
16
+ ],
17
+ "text_encoder": [
18
+ "transformers",
19
+ "CLIPTextModel"
20
+ ],
21
+ "tokenizer": [
22
+ "transformers",
23
+ "CLIPTokenizer"
24
+ ],
25
+ "unet": [
26
+ "diffusers",
27
+ "UNet2DConditionModel"
28
+ ],
29
+ "vae": [
30
+ "diffusers",
31
+ "AutoencoderKL"
32
+ ]
33
+ }
scheduler/scheduler_config.json CHANGED
@@ -1,16 +1,19 @@
1
  {
2
- "_class_name": "DDIMScheduler",
3
- "_diffusers_version": "0.15.0.dev0",
 
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
7
  "clip_sample": false,
8
- "clip_sample_range": 1.0,
9
  "dynamic_thresholding_ratio": 0.995,
 
10
  "num_train_timesteps": 1000,
11
  "prediction_type": "epsilon",
12
  "sample_max_value": 1.0,
13
  "set_alpha_to_one": false,
 
 
14
  "steps_offset": 1,
15
  "thresholding": false,
16
  "trained_betas": null
 
1
  {
2
+ "_class_name": "DEISMultistepScheduler",
3
+ "_diffusers_version": "0.14.0",
4
+ "algorithm_type": "deis",
5
  "beta_end": 0.012,
6
  "beta_schedule": "scaled_linear",
7
  "beta_start": 0.00085,
8
  "clip_sample": false,
 
9
  "dynamic_thresholding_ratio": 0.995,
10
+ "lower_order_final": true,
11
  "num_train_timesteps": 1000,
12
  "prediction_type": "epsilon",
13
  "sample_max_value": 1.0,
14
  "set_alpha_to_one": false,
15
+ "solver_order": 2,
16
+ "solver_type": "logrho",
17
  "steps_offset": 1,
18
  "thresholding": false,
19
  "trained_betas": null
text_encoder/config.json CHANGED
@@ -20,6 +20,6 @@
20
  "pad_token_id": 1,
21
  "projection_dim": 768,
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.26.1",
24
  "vocab_size": 49408
25
  }
 
20
  "pad_token_id": 1,
21
  "projection_dim": 768,
22
  "torch_dtype": "float32",
23
+ "transformers_version": "4.27.3",
24
  "vocab_size": 49408
25
  }
text_encoder/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:180c897d4c12afb44895fcb22e6789d1164b1fd2cac65907ea5d34202d59f998
3
- size 492265874
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d934a4375a0b0b22facfc479e5a8b81be37db59e7eb333f100034740fbbab36a
3
+ size 492265879
tokenizer/tokenizer_config.json CHANGED
@@ -19,7 +19,6 @@
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
22
- "name_or_path": "openai/clip-vit-large-patch14",
23
  "pad_token": "<|endoftext|>",
24
  "special_tokens_map_file": "./special_tokens_map.json",
25
  "tokenizer_class": "CLIPTokenizer",
 
19
  },
20
  "errors": "replace",
21
  "model_max_length": 77,
 
22
  "pad_token": "<|endoftext|>",
23
  "special_tokens_map_file": "./special_tokens_map.json",
24
  "tokenizer_class": "CLIPTokenizer",
unet/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
- "_diffusers_version": "0.15.0.dev0",
4
  "act_fn": "silu",
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.14.0",
4
  "act_fn": "silu",
5
  "attention_head_dim": 8,
6
  "block_out_channels": [
unet/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d13530a9ff30fcf50fdddd32d7f66efe72058358117e13948f843017a907acee
3
- size 3438167540
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bda43092a179100c2f1d18e2a17353952c60ae72826955b8615d901a236fc42
3
+ size 3438167534
vae/config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.15.0.dev0",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.14.0",
4
  "act_fn": "silu",
5
  "block_out_channels": [
6
  128,