instruct-pix2pix / metrics /compute_metrics.py
holynski's picture
Updates for v1 release
926ff6c
raw
history blame
8.19 kB
from __future__ import annotations
import math
import random
import sys
from argparse import ArgumentParser
import einops
import k_diffusion as K
import numpy as np
import torch
import torch.nn as nn
from tqdm.auto import tqdm
from einops import rearrange
from omegaconf import OmegaConf
from PIL import Image, ImageOps
from torch import autocast
import json
import matplotlib.pyplot as plt
import seaborn
from pathlib import Path
sys.path.append("./")
from clip_similarity import ClipSimilarity
from edit_dataset import EditDatasetEval
sys.path.append("./stable_diffusion")
from ldm.util import instantiate_from_config
class CFGDenoiser(nn.Module):
def __init__(self, model):
super().__init__()
self.inner_model = model
def forward(self, z, sigma, cond, uncond, text_cfg_scale, image_cfg_scale):
cfg_z = einops.repeat(z, "1 ... -> n ...", n=3)
cfg_sigma = einops.repeat(sigma, "1 ... -> n ...", n=3)
cfg_cond = {
"c_crossattn": [torch.cat([cond["c_crossattn"][0], uncond["c_crossattn"][0], uncond["c_crossattn"][0]])],
"c_concat": [torch.cat([cond["c_concat"][0], cond["c_concat"][0], uncond["c_concat"][0]])],
}
out_cond, out_img_cond, out_uncond = self.inner_model(cfg_z, cfg_sigma, cond=cfg_cond).chunk(3)
return out_uncond + text_cfg_scale * (out_cond - out_img_cond) + image_cfg_scale * (out_img_cond - out_uncond)
def load_model_from_config(config, ckpt, vae_ckpt=None, verbose=False):
print(f"Loading model from {ckpt}")
pl_sd = torch.load(ckpt, map_location="cpu")
if "global_step" in pl_sd:
print(f"Global Step: {pl_sd['global_step']}")
sd = pl_sd["state_dict"]
if vae_ckpt is not None:
print(f"Loading VAE from {vae_ckpt}")
vae_sd = torch.load(vae_ckpt, map_location="cpu")["state_dict"]
sd = {
k: vae_sd[k[len("first_stage_model.") :]] if k.startswith("first_stage_model.") else v
for k, v in sd.items()
}
model = instantiate_from_config(config.model)
m, u = model.load_state_dict(sd, strict=False)
if len(m) > 0 and verbose:
print("missing keys:")
print(m)
if len(u) > 0 and verbose:
print("unexpected keys:")
print(u)
return model
class ImageEditor(nn.Module):
def __init__(self, config, ckpt, vae_ckpt=None):
super().__init__()
config = OmegaConf.load(config)
self.model = load_model_from_config(config, ckpt, vae_ckpt)
self.model.eval().cuda()
self.model_wrap = K.external.CompVisDenoiser(self.model)
self.model_wrap_cfg = CFGDenoiser(self.model_wrap)
self.null_token = self.model.get_learned_conditioning([""])
def forward(
self,
image: torch.Tensor,
edit: str,
scale_txt: float = 7.5,
scale_img: float = 1.0,
steps: int = 100,
) -> torch.Tensor:
assert image.dim() == 3
assert image.size(1) % 64 == 0
assert image.size(2) % 64 == 0
with torch.no_grad(), autocast("cuda"), self.model.ema_scope():
cond = {
"c_crossattn": [self.model.get_learned_conditioning([edit])],
"c_concat": [self.model.encode_first_stage(image[None]).mode()],
}
uncond = {
"c_crossattn": [self.model.get_learned_conditioning([""])],
"c_concat": [torch.zeros_like(cond["c_concat"][0])],
}
extra_args = {
"uncond": uncond,
"cond": cond,
"image_cfg_scale": scale_img,
"text_cfg_scale": scale_txt,
}
sigmas = self.model_wrap.get_sigmas(steps)
x = torch.randn_like(cond["c_concat"][0]) * sigmas[0]
x = K.sampling.sample_euler_ancestral(self.model_wrap_cfg, x, sigmas, extra_args=extra_args)
x = self.model.decode_first_stage(x)[0]
return x
def compute_metrics(config,
model_path,
vae_ckpt,
data_path,
output_path,
scales_img,
scales_txt,
num_samples = 5000,
split = "test",
steps = 50,
res = 512,
seed = 0):
editor = ImageEditor(config, model_path, vae_ckpt).cuda()
clip_similarity = ClipSimilarity().cuda()
outpath = Path(output_path, f"n={num_samples}_p={split}_s={steps}_r={res}_e={seed}.jsonl")
Path(output_path).mkdir(parents=True, exist_ok=True)
for scale_txt in scales_txt:
for scale_img in scales_img:
dataset = EditDatasetEval(
path=data_path,
split=split,
res=res
)
assert num_samples <= len(dataset)
print(f'Processing t={scale_txt}, i={scale_img}')
torch.manual_seed(seed)
perm = torch.randperm(len(dataset))
count = 0
i = 0
sim_0_avg = 0
sim_1_avg = 0
sim_direction_avg = 0
sim_image_avg = 0
count = 0
pbar = tqdm(total=num_samples)
while count < num_samples:
idx = perm[i].item()
sample = dataset[idx]
i += 1
gen = editor(sample["image_0"].cuda(), sample["edit"], scale_txt=scale_txt, scale_img=scale_img, steps=steps)
sim_0, sim_1, sim_direction, sim_image = clip_similarity(
sample["image_0"][None].cuda(), gen[None].cuda(), [sample["input_prompt"]], [sample["output_prompt"]]
)
sim_0_avg += sim_0.item()
sim_1_avg += sim_1.item()
sim_direction_avg += sim_direction.item()
sim_image_avg += sim_image.item()
count += 1
pbar.update(count)
pbar.close()
sim_0_avg /= count
sim_1_avg /= count
sim_direction_avg /= count
sim_image_avg /= count
with open(outpath, "a") as f:
f.write(f"{json.dumps(dict(sim_0=sim_0_avg, sim_1=sim_1_avg, sim_direction=sim_direction_avg, sim_image=sim_image_avg, num_samples=num_samples, split=split, scale_txt=scale_txt, scale_img=scale_img, steps=steps, res=res, seed=seed))}\n")
return outpath
def plot_metrics(metrics_file, output_path):
with open(metrics_file, 'r') as f:
data = [json.loads(line) for line in f]
plt.rcParams.update({'font.size': 11.5})
seaborn.set_style("darkgrid")
plt.figure(figsize=(20.5* 0.7, 10.8* 0.7), dpi=200)
x = [d["sim_direction"] for d in data]
y = [d["sim_image"] for d in data]
plt.plot(x, y, marker='o', linewidth=2, markersize=4)
plt.xlabel("CLIP Text-Image Direction Similarity", labelpad=10)
plt.ylabel("CLIP Image Similarity", labelpad=10)
plt.savefig(Path(output_path) / Path("plot.pdf"), bbox_inches="tight")
def main():
parser = ArgumentParser()
parser.add_argument("--resolution", default=512, type=int)
parser.add_argument("--steps", default=100, type=int)
parser.add_argument("--config", default="configs/generate.yaml", type=str)
parser.add_argument("--output_path", default="analysis/", type=str)
parser.add_argument("--ckpt", default="checkpoints/instruct-pix2pix-00-22000.ckpt", type=str)
parser.add_argument("--dataset", default="data/clip-filtered-dataset/", type=str)
parser.add_argument("--vae-ckpt", default=None, type=str)
args = parser.parse_args()
scales_img = [1.0, 1.2, 1.4, 1.6, 1.8, 2.0, 2.2]
scales_txt = [7.5]
metrics_file = compute_metrics(
args.config,
args.ckpt,
args.vae_ckpt,
args.dataset,
args.output_path,
scales_img,
scales_txt
steps = args.steps
)
plot_metrics(metrics_file, args.output_path)
if __name__ == "__main__":
main()