|
import gradio as gr |
|
|
|
gr.load("models/cutycat2000/InterDiffusion-2.5").launch() |
|
from diffusers import ( |
|
StableDiffusionXLPipeline, |
|
KDPM2AncestralDiscreteScheduler, |
|
AutoencoderKL |
|
) |
|
vae = AutoencoderKL.from_pretrained( |
|
"madebyollin/sdxl-vae-fp16-fix", |
|
torch_dtype=torch.float16 |
|
) |
|
pipe = StableDiffusionXLPipeline.from_pretrained( |
|
"cutycat2000/InterDiffusion-2.5", |
|
vae=vae, |
|
torch_dtype=torch.float16 |
|
) |
|
pipe.scheduler = KDPM2AncestralDiscreteScheduler.from_config(pipe.scheduler.config) |
|
pipe.to('cuda') |
|
prompt = "Hello World" |
|
negative_prompt = "text" |
|
|
|
image = pipe( |
|
prompt, |
|
negative_prompt=negative_prompt, |
|
width=1024, |
|
height=1024, |
|
guidance_scale=7, |
|
num_inference_steps=20 |
|
).images[0] |