Core ML Converted Model:
- This model was converted to Core ML for use on Apple Silicon devices. Conversion instructions can be found here.
- Provide the model to an app such as Mochi Diffusion Github / Discord to generate images.
split_einsum
version is compatible with all compute unit options including Neural Engine.original
version is only compatible withCPU & GPU
option.- Custom resolution versions are tagged accordingly.
- The
vae-ft-mse-840000-ema-pruned.ckpt
VAE is embedded into the model. - This model was converted with a
vae-encoder
for use withimage2image
. - This model is
fp16
. - Descriptions are posted as-is from original model source.
- Not all features and/or results may be available in
CoreML
format. - This model does not have the unet split into chunks.
- This model does not include a
safety checker
(for NSFW content). - This model can be used with ControlNet.
inkpunkDiffusion-v2_cn:
Source(s): CivitAI / Hugging Face
Inkpunk Diffusion v2
Finetuned Stable Diffusion model trained on Dreambooth.
Vaguely inspired by Gorillaz, FLCL, and Yoji Shinkawa
Use nvinkpunk in your prompts.
About Version 2
Improvements:
Excorsized the woman that seemed to be haunting the model and appearing in basically every prompt. (She still pops in here and there, I honestly don't know where she comes from.)
Faces are better overall, and characters are more responsive to descriptive language in prompts. Try playing around with emotions/nationality/age/poses to get the best results.
Hair anarchy has been toned down slightly and added more diverse hairstyles to the dataset to avoid everyone having tumbleweed hair. Again, descriptive language in prompts helps.
Full body poses are also more diverse, though keep in mind that wide angle shots tend to lead to distorted faces that may need to be touched up with img2img/inpainting.