Spaces:
Sleeping
Sleeping
import os | |
os.environ["PYTORCH_ENABLE_MPS_FALLBACK"] = "1" | |
from pathlib import Path | |
import pytest | |
import torch | |
from lama_cleaner.model_manager import ModelManager | |
from lama_cleaner.schema import HDStrategy, SDSampler | |
from lama_cleaner.tests.test_model import get_config, assert_equal | |
current_dir = Path(__file__).parent.absolute().resolve() | |
save_dir = current_dir / "result" | |
save_dir.mkdir(exist_ok=True, parents=True) | |
device = "cuda" if torch.cuda.is_available() else "cpu" | |
device = torch.device(device) | |
def test_runway_sd_1_5_ddim( | |
sd_device, strategy, sampler, cpu_textencoder, disable_nsfw | |
): | |
def callback(i, t, latents): | |
pass | |
if sd_device == "cuda" and not torch.cuda.is_available(): | |
return | |
sd_steps = 50 if sd_device == "cuda" else 1 | |
model = ModelManager( | |
name="sd1.5", | |
device=torch.device(sd_device), | |
hf_access_token="", | |
sd_run_local=True, | |
disable_nsfw=disable_nsfw, | |
sd_cpu_textencoder=cpu_textencoder, | |
callback=callback, | |
) | |
cfg = get_config(strategy, prompt="a fox sitting on a bench", sd_steps=sd_steps) | |
cfg.sd_sampler = sampler | |
name = f"device_{sd_device}_{sampler}_cpu_textencoder_{cpu_textencoder}_disnsfw_{disable_nsfw}" | |
assert_equal( | |
model, | |
cfg, | |
f"runway_sd_{strategy.capitalize()}_{name}.png", | |
img_p=current_dir / "overture-creations-5sI6fQgYIuo.png", | |
mask_p=current_dir / "overture-creations-5sI6fQgYIuo_mask.png", | |
fx=1.3, | |
) | |
def test_runway_sd_1_5(sd_device, strategy, sampler, cpu_textencoder, disable_nsfw): | |
def callback(i, t, latents): | |
print(f"sd_step_{i}") | |
if sd_device == "cuda" and not torch.cuda.is_available(): | |
return | |
sd_steps = 50 if sd_device == "cuda" else 1 | |
model = ModelManager( | |
name="sd1.5", | |
device=torch.device(sd_device), | |
hf_access_token="", | |
sd_run_local=True, | |
disable_nsfw=disable_nsfw, | |
sd_cpu_textencoder=cpu_textencoder, | |
callback=callback, | |
) | |
cfg = get_config(strategy, prompt="a fox sitting on a bench", sd_steps=sd_steps) | |
cfg.sd_sampler = sampler | |
name = f"device_{sd_device}_{sampler}_cpu_textencoder_{cpu_textencoder}_disnsfw_{disable_nsfw}" | |
assert_equal( | |
model, | |
cfg, | |
f"runway_sd_{strategy.capitalize()}_{name}.png", | |
img_p=current_dir / "overture-creations-5sI6fQgYIuo.png", | |
mask_p=current_dir / "overture-creations-5sI6fQgYIuo_mask.png", | |
fx=1.3, | |
) | |
def test_runway_sd_1_5_negative_prompt(sd_device, strategy, sampler): | |
def callback(i, t, latents): | |
pass | |
if sd_device == "cuda" and not torch.cuda.is_available(): | |
return | |
sd_steps = 50 if sd_device == "cuda" else 1 | |
model = ModelManager( | |
name="sd1.5", | |
device=torch.device(sd_device), | |
hf_access_token="", | |
sd_run_local=True, | |
disable_nsfw=False, | |
sd_cpu_textencoder=False, | |
callback=callback, | |
) | |
cfg = get_config( | |
strategy, | |
sd_steps=sd_steps, | |
prompt="Face of a fox, high resolution, sitting on a park bench", | |
negative_prompt="orange, yellow, small", | |
sd_sampler=sampler, | |
sd_match_histograms=True, | |
) | |
name = f"{sampler}_negative_prompt" | |
assert_equal( | |
model, | |
cfg, | |
f"runway_sd_{strategy.capitalize()}_{name}.png", | |
img_p=current_dir / "overture-creations-5sI6fQgYIuo.png", | |
mask_p=current_dir / "overture-creations-5sI6fQgYIuo_mask.png", | |
fx=1, | |
) | |
def test_runway_sd_1_5_sd_scale( | |
sd_device, strategy, sampler, cpu_textencoder, disable_nsfw | |
): | |
if sd_device == "cuda" and not torch.cuda.is_available(): | |
return | |
sd_steps = 50 if sd_device == "cuda" else 1 | |
model = ModelManager( | |
name="sd1.5", | |
device=torch.device(sd_device), | |
hf_access_token="", | |
sd_run_local=True, | |
disable_nsfw=disable_nsfw, | |
sd_cpu_textencoder=cpu_textencoder, | |
) | |
cfg = get_config( | |
strategy, prompt="a fox sitting on a bench", sd_steps=sd_steps, sd_scale=0.85 | |
) | |
cfg.sd_sampler = sampler | |
name = f"device_{sd_device}_{sampler}_cpu_textencoder_{cpu_textencoder}_disnsfw_{disable_nsfw}" | |
assert_equal( | |
model, | |
cfg, | |
f"runway_sd_{strategy.capitalize()}_{name}_sdscale.png", | |
img_p=current_dir / "overture-creations-5sI6fQgYIuo.png", | |
mask_p=current_dir / "overture-creations-5sI6fQgYIuo_mask.png", | |
fx=1.3, | |
) | |
def test_runway_sd_1_5_cpu_offload(sd_device, strategy, sampler): | |
if sd_device == "cuda" and not torch.cuda.is_available(): | |
return | |
sd_steps = 50 if sd_device == "cuda" else 1 | |
model = ModelManager( | |
name="sd1.5", | |
device=torch.device(sd_device), | |
hf_access_token="", | |
sd_run_local=True, | |
disable_nsfw=True, | |
sd_cpu_textencoder=False, | |
cpu_offload=True, | |
) | |
cfg = get_config( | |
strategy, prompt="a fox sitting on a bench", sd_steps=sd_steps, sd_scale=0.85 | |
) | |
cfg.sd_sampler = sampler | |
name = f"device_{sd_device}_{sampler}" | |
assert_equal( | |
model, | |
cfg, | |
f"runway_sd_{strategy.capitalize()}_{name}_cpu_offload.png", | |
img_p=current_dir / "overture-creations-5sI6fQgYIuo.png", | |
mask_p=current_dir / "overture-creations-5sI6fQgYIuo_mask.png", | |
) | |
def test_local_file_path(sd_device, sampler): | |
if sd_device == "cuda" and not torch.cuda.is_available(): | |
return | |
sd_steps = 1 if sd_device == "cpu" else 50 | |
model = ModelManager( | |
name="sd1.5", | |
device=torch.device(sd_device), | |
hf_access_token="", | |
sd_run_local=True, | |
disable_nsfw=True, | |
sd_cpu_textencoder=False, | |
cpu_offload=True, | |
sd_local_model_path="/Users/cwq/data/models/sd-v1-5-inpainting.ckpt", | |
) | |
cfg = get_config( | |
HDStrategy.ORIGINAL, | |
prompt="a fox sitting on a bench", | |
sd_steps=sd_steps, | |
) | |
cfg.sd_sampler = sampler | |
name = f"device_{sd_device}_{sampler}" | |
assert_equal( | |
model, | |
cfg, | |
f"sd_local_model_{name}.png", | |
img_p=current_dir / "overture-creations-5sI6fQgYIuo.png", | |
mask_p=current_dir / "overture-creations-5sI6fQgYIuo_mask.png", | |
) | |