import gradio as gr import cv2 import torch import numpy as np from torchvision import transforms import torch from pytorch_lightning import seed_everything from torchvision.utils import save_image from model_lib.modules import MoMA_main_modal from model_lib.utils import parse_args import os os.environ["CUDA_VISIBLE_DEVICES"]="0" title = "MoMA" description = "This model has to run on GPU" article = "

Blog | Github Repo

" def MoMA_demo(rgb, mask, subject, prompt): # move the input and model to GPU for speed if available with torch.no_grad(): generated_image = model.generate_images(rgb, mask, subject, prompt, strength=1.0, seed=2) return generated_image def inference(rgb, mask, subject, prompt): result = MoMA_demo(rgb, mask, subject, prompt) return result seed_everything(0) args = parse_args() #load MoMA from HuggingFace. Auto download model = MoMA_main_modal(args).to(args.device, dtype=torch.float16) ################ change texture ################## # prompt = "A wooden sculpture of a car on the table." # generated_image = model.generate_images(rgb_path, mask_path, subject, prompt, strength=0.4, seed=4, return_mask=True) # set strength to 0.4 for better prompt fidelity # save_image(generated_image,f"{args.output_path}/{subject}_{prompt}.jpg") gr.Interface( inference, [gr.Image(type="pil", label="Input RGB"), gr.Image(type="pil", label="Input Mask"), gr.Textbox(lines=1, label="subject"), gr.Textbox(lines=5, label="Prompt")], gr.Image(type="pil", label="Output"), title=title, description=description, article=article, examples=[["example_images/newImages/3.jpg",'example_images/newImages/3_mask.jpg','car','A car in autumn with falling leaves.']], # enable_queue=True ).launch(debug=False)