Update README.md
Browse files
README.md
CHANGED
@@ -50,7 +50,7 @@ From left to right: Input image, Masked image, SDXL inpainting, Ours.
|
|
50 |
|
51 |
Step1: Make sure you upgrade to the latest version of diffusers(>=0.29.2): pip install -U diffusers.
|
52 |
|
53 |
-
Step2: Download the two required Python files(pipeline_sd3_controlnet_inpainting.py and controlnet_sd3.py) from either the current repo or from (
|
54 |
(We will merge this Feature to official Diffusers.)
|
55 |
|
56 |
Step3: And then you can run demo.py or following:
|
@@ -126,4 +126,4 @@ The model was trained on 12M laion2B and internal source images for 20k steps at
|
|
126 |
Due to the fact that only 1024*1024 pixel resolution was used during the training phase, the inference performs best at this size, with other sizes yielding suboptimal results. We will initiate multi-resolution training in the future, and at that time, we will open-source the new weights.
|
127 |
|
128 |
## LICENSE
|
129 |
-
The model is based on SD3 finetuning; therefore, the license follows the original
|
|
|
50 |
|
51 |
Step1: Make sure you upgrade to the latest version of diffusers(>=0.29.2): pip install -U diffusers.
|
52 |
|
53 |
+
Step2: Download the two required Python files(pipeline_sd3_controlnet_inpainting.py and controlnet_sd3.py) from either the current repo or from [GitHub](https://github.com/JPlin/SD3-Controlnet-Inpainting).
|
54 |
(We will merge this Feature to official Diffusers.)
|
55 |
|
56 |
Step3: And then you can run demo.py or following:
|
|
|
126 |
Due to the fact that only 1024*1024 pixel resolution was used during the training phase, the inference performs best at this size, with other sizes yielding suboptimal results. We will initiate multi-resolution training in the future, and at that time, we will open-source the new weights.
|
127 |
|
128 |
## LICENSE
|
129 |
+
The model is based on SD3 finetuning; therefore, the license follows the original [SD3 license](https://huggingface.co/stabilityai/stable-diffusion-3-medium#license).
|