Upload README.md
Browse files
README.md
CHANGED
@@ -11,56 +11,46 @@ tags:
|
|
11 |
|
12 |
The SD models are all "original" (not split-einsum) and built for CPU and GPU. They are each for the output size noted. They are fp16, with the standard SD-1.5 VAE embedded.
|
13 |
|
14 |
-
The Stable Diffusion v1.5 model and the other SD 1.5 type models
|
15 |
|
16 |
-
They
|
17 |
|
18 |
-
All the ControlNet models are also "original" ones, built for CPU and GPU compute units (cpuAndGPU) and for SD-1.5 type models. The
|
19 |
|
20 |
-
All of the models in this repo will only work with Swift and the current ml-stable-diffusion pipeline (0.4.0). They were not built for a python diffusers pipeline. They need apple/ml-stable-diffusion (from GitHub) for command line use or a Swift app
|
21 |
|
22 |
-
The full SD models are in the "SD" folder here. They are individually zipped
|
23 |
|
24 |
-
The ControlNet model files are in the "CN" folder here. They are also zipped and need to be unzipped after downloading. Note that
|
25 |
|
26 |
-
There is also a MISC folder that has text files with my notes and a screencap of my directory structure.
|
27 |
|
28 |
-
For command line use, it all runs in a miniconda3 environment, covered in one of the notes. If you are using the command line, please read the notes concerning naming and placement of your ControlNet model folder. If you are using a GUI, it will guide you to the correct location/arrangement.
|
29 |
|
30 |
-
|
31 |
-
**You can just rename them yourself, or download them again as the file names have been corrected.**
|
32 |
|
33 |
-
**
|
34 |
-
|
35 |
-
**Sorry if my early transposing of sizes messed with your mind**
|
36 |
-
|
37 |
-
## Notes
|
38 |
-
|
39 |
-
- There ia also a branch to main here called "For-Mochi-Model-Env".
|
40 |
-
- It was going to be a shortcut version of the conversion and generation pipelines for people who already have a setup for converting models per the Wiki at Mochi Diffusion. Development of a new version of Mochi Diffusion, with ControlNet included, is moving along very quickly, so I don't plan to spend more time on the CLI instructions.
|
41 |
-
- If you downloaded Stable Diffusion v1.5 Orignal 768x768 For ControlNet before 4/27/23, or Stable Diffusion v1.5 Original 512x768 before 5/4/23, please re-download. Those models were not supporting all intended features.
|
42 |
-
- If you encounter any models that do not work fully with image2image and ControlNet using the current CLI pipeline or Mochi Diffusion 3.2, please leave a report in the Community area here.
|
43 |
|
44 |
## Model List
|
45 |
-
**Each zip
|
46 |
|
47 |
-
-
|
48 |
-
- MyMerge of 8 1.5-type NSFW models, original, for ControlNet & Standard
|
49 |
-
- MeinaMix9 1.5-type anime model, original, for ControlNet & Standard
|
50 |
- GhostMix v1.1, 1.5-type anime model, original, for ControlNet & Standard
|
|
|
|
|
51 |
- Realistic Vision v2.0, 1.5-type model, original, for ControlNet & Standard
|
52 |
-
-
|
53 |
|
54 |
## ControlNet List
|
55 |
-
**
|
56 |
|
57 |
- Canny -- Edge Detection, Outlines As Input
|
58 |
-
-
|
59 |
-
- InstrP2P -- Instruct Picture2Picture, Modified By Text ("change dog to cat")
|
60 |
-
- MLSD -- Find And Reuse Straight Lines And Edges
|
61 |
- InPaint -- Modify An Indicated Area Of An Image (not sure how this works)
|
|
|
62 |
- LineArt -- Find And Reuse Small Outlines
|
|
|
63 |
- OpenPose -- Copy Body Poses
|
|
|
64 |
- SoftEdge -- Find And Reuse Soft Edges
|
65 |
- Tile -- Subtle Variations In Batch Runs
|
66 |
-
- Depth -- Reproduces Depth Relationships From An Image
|
|
|
11 |
|
12 |
The SD models are all "original" (not split-einsum) and built for CPU and GPU. They are each for the output size noted. They are fp16, with the standard SD-1.5 VAE embedded.
|
13 |
|
14 |
+
The Stable Diffusion v1.5 model and the other SD 1.5 type models contain both the standard Unet and the ControlledUnet used for a ControlNet pipeline. The correct one will be used automatically based on whether a ControlNet is enabled or not.
|
15 |
|
16 |
+
They have VAEEncoder.mlmodelc bundles that allow Image2Image to operate correctly at the noted resolutions, when used with a current Swift CLI pipeline or a current GUI built with ml-stable-diffusion 0.4.0, such as Mochi Diffusion 3.2 or later.
|
17 |
|
18 |
+
All of the ControlNet models are also "original" ones, built for CPU and GPU compute units (cpuAndGPU) and for SD-1.5 type models. The zip files each contain have a set of models at 4 resolutions. They may also work with split-einsum models, using CPU and GPU (not CPU and NE), but they will not work with SD-2.1 type models at all.
|
19 |
|
20 |
+
All of the models in this repo will only work with Swift and the current ml-stable-diffusion pipeline (0.4.0). They were not built for a python diffusers pipeline. They need apple/ml-stable-diffusion (from GitHub) for command line use or a Swift app that supports ControlNet, such as the Mochi Diffusion test version currently in a closed beta test at https://github.com/godly-devotion/MochiDiffusion that supports ControlNet. Join the Mochi Difusion Discord server at https://discord.gg/x2kartzxGv to request access to the beta test version.
|
21 |
|
22 |
+
The full SD models are in the "SD" folder here. They are in subfolders by model name and individually zipped for a particular resolution. They need to be unzipped for use after downloading.
|
23 |
|
24 |
+
The ControlNet model files are in the "CN" folder here. They are also zipped and need to be unzipped after downloading. Note that they are zipped into sets of 4: 512x512, 512x768, 768x512, 768x768 for each ControlNet type.
|
25 |
|
26 |
+
There is also a MISC folder that has text files with my notes and a screencap of my directory structure. These are provided for folks who want to try converting models themselves and/or running the models with a SwiftCLI. The notes are not perfect, and may be out of date as the various python and CoreML packages are updated.
|
27 |
|
28 |
+
For command line use, it all runs in a miniconda3 environment, covered in one of the notes. If you are using the command line, please read the notes concerning naming and placement of your ControlNet model folder. If you are using a GUI, it will most likely guide you to the correct location/arrangement.
|
29 |
|
30 |
+
The sizes are always meant to be WIDTH x HEIGHT. A 512x768 is "portrait" orientation and a 768x512 is "landscape" orientation.**
|
|
|
31 |
|
32 |
+
**If you encounter any models that do not work fully with image2image and ControlNet, using the current apple/ml-stable-diffusion SwiftCLI pipeline or Mochi Diffusion 3.2 or the Mochi Diffusion CN test build, please leave a report in the Community area here. If you would like to add models that you have converted, leave a message as well, and I'll try to figure out out to grant you access to this repo.**
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
|
34 |
## Model List
|
35 |
+
**Models are organized into folders by model name. Each folder contains 4 zip files of single models for the output size indicated: 512x512, 512x768, 768x512 or 768x768.**
|
36 |
|
37 |
+
- DreamShaper v5.0, 1.5-type model, original, for ControlNet & Standard <<<=== NEW <<<=== NEW
|
|
|
|
|
38 |
- GhostMix v1.1, 1.5-type anime model, original, for ControlNet & Standard
|
39 |
+
- MeinaMix v9.0 1.5-type anime model, original, for ControlNet & Standard
|
40 |
+
- MyMerge v1.0 1.5-type NSFW model, original, for ControlNet & Standard
|
41 |
- Realistic Vision v2.0, 1.5-type model, original, for ControlNet & Standard
|
42 |
+
- Stable Diffusion v1.5, original, for ControlNet & Standard
|
43 |
|
44 |
## ControlNet List
|
45 |
+
**Each file is a set of 4 resolutions zipped together: 512x512, 512x768, 768x512, 768x768**
|
46 |
|
47 |
- Canny -- Edge Detection, Outlines As Input
|
48 |
+
- Depth -- Reproduces Depth Relationships From An Image
|
|
|
|
|
49 |
- InPaint -- Modify An Indicated Area Of An Image (not sure how this works)
|
50 |
+
- InstrP2P -- Instruct Picture2Picture, Modified By Text ("change dog to cat")
|
51 |
- LineArt -- Find And Reuse Small Outlines
|
52 |
+
- MLSD -- Find And Reuse Straight Lines And Edges
|
53 |
- OpenPose -- Copy Body Poses
|
54 |
+
- Scribble -- Freehand Sketch As Input
|
55 |
- SoftEdge -- Find And Reuse Soft Edges
|
56 |
- Tile -- Subtle Variations In Batch Runs
|
|