|
""" |
|
User script for AUTOMATIC111's SD WebUI that integrates MagicPrompt |
|
Copyright (C) 2022 Spaceginner |
|
|
|
This user script is free software: you can redistribute it and/or modify |
|
it under the terms of the GNU General Public License as published by |
|
the Free Software Foundation, either version 3 of the License, or |
|
(at your option) any later version. |
|
|
|
This user script is distributed in the hope that it will be useful, |
|
but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
|
GNU General Public License for more details. |
|
|
|
You should have received a copy of the GNU General Public License |
|
along with this user script. If not, see <https://www.gnu.org/licenses/>. |
|
|
|
Contact me via Discord (Spaceginner#7688), email (ivan.demian2009@gmail.com) |
|
or via "Issues" tab on Github page of this script |
|
(https://github.com/Spaceginner/MagicPrompt-awebui) |
|
|
|
Credits: |
|
Special thanks to Gustavosta for creating MagicPrompt AI model. |
|
Also credits go to u/Letharguss (for creating basic script) |
|
and SoCalGuitarist#2586 (for figuring out how to change prompt batch to batch) |
|
|
|
Version: 1.0.0 |
|
""" |
|
|
|
import os |
|
import sys |
|
import subprocess |
|
|
|
|
|
try: |
|
from aitextgen import aitextgen |
|
except: |
|
print("[MagicPrompt script] aitextgen module is not found, downloading...") |
|
if os.path.exists("./venv"): |
|
subprocess.call(["./venv/Scripts/python", "-m", "pip", "-q", "--disable-pip-version-check", "--no-input", "install", "aitextgen"]) |
|
else: |
|
subprocess.call(["python", "-m", "pip", "-q", "--disable-pip-version-check", "--no-input", "install", "aitextgen"]) |
|
print("[MagicPrompt script] aitextgen module is downloaded") |
|
|
|
import gradio as gr |
|
import torch |
|
|
|
import modules.scripts as scripts |
|
from modules.processing import Processed, process_images |
|
from modules.shared import state |
|
|
|
|
|
if (not os.path.isdir("./models/MagicPrompt/")): |
|
print("[MagicPrompt script] MagicPrompt model is not found, downloading MagicPrompt model...") |
|
os.mkdir("./models/MagicPrompt/") |
|
subprocess.call(["git", "clone", "--quiet", "https://huggingface.co/Gustavosta/MagicPrompt-Stable-Diffusion", "./models/MagicPrompt/."]) |
|
print("[MagicPrompt script] MagicPrompt model is downloaded") |
|
|
|
def getOrdinalNum(n): |
|
if str(n)[-1] == "1": |
|
return f"{n}st" |
|
elif str(n)[-1] == "2": |
|
return f"{n}nd" |
|
elif str(n)[-1] == "3": |
|
return f"{n}rd" |
|
else: |
|
return f"{n}th" |
|
|
|
class Script(scripts.Script): |
|
|
|
gpt = None |
|
|
|
def title(self): |
|
return "MagicPrompt" |
|
|
|
def show(self, isImg2img): |
|
|
|
return not isImg2img |
|
|
|
def ui(self, isImg2img): |
|
|
|
promptLength = gr.Slider(label="Prompt max. length", value=75, minimum=1, maximum=300, step=1) |
|
temp = gr.Slider(label="Temperature", value=0.7, minimum=0.1, maximum=2, step=0.1) |
|
useSameSeed = gr.Checkbox(label="Use same seed for each batch", value=False) |
|
useUniquePrompt = gr.Checkbox(label="Use unique prompt for each batch", value=True) |
|
isPrioritized = gr.Checkbox(label="Iniatial prompt will have more prority over generated one", value=False) |
|
doPregenerating = gr.Checkbox(label="Enable prompt pregenerating (Theoretical perfomance boost). If you dont know how many images do you want to generate, disable it", value=True) |
|
doUnloadModel = gr.Checkbox(label="Unload MagicPrompt model from VRAM/RAM after this run. (Decreased perfomance between runs, as it need to load again)", value=False) |
|
|
|
return [promptLength, temp, useSameSeed, useUniquePrompt, isPrioritized, doPregenerating, doUnloadModel] |
|
|
|
def run(self, p, promptLength, temp, useSameSeed, useUniquePrompt, isPrioritized, doPregenerating, doUnloadModel): |
|
print() |
|
|
|
|
|
if type(self.gpt) != aitextgen: |
|
self.gpt = aitextgen(model_folder="./models/MagicPrompt/", tokenizer_file="./models/MagicPrompt/tokenizer.json", to_gpu=torch.cuda.is_available()) |
|
|
|
p.do_not_save_grid = True |
|
|
|
|
|
p.prompt = p.prompt[0] if type(p.prompt) == list else p.prompt |
|
|
|
|
|
|
|
state.job_count = p.n_iter |
|
p.n_iter = 1 |
|
|
|
|
|
originalPrompt = p.prompt |
|
if (originalPrompt != "" and isPrioritized): |
|
originalPrompt = "(" + originalPrompt + ")" |
|
|
|
|
|
prompts = [] |
|
if (doPregenerating): |
|
print(f"[MagicPrompt script] Pregenerating prompt{'s' if state.job_count > 1 else ''}...") |
|
for i in range(state.job_count): |
|
if (i == 0 or useUniquePrompt): |
|
if state.interrupted: |
|
print(f"[MagicPrompt script] Pregeneration interrupted") |
|
break |
|
|
|
prompts.append(self.gpt.generate_one(prompt=originalPrompt, max_length=promptLength, temperature=temp)) |
|
if state.job_count > 1: |
|
print(f"[MagicPrompt script] Pregenerated {getOrdinalNum(i+1)} prompt...") |
|
else: |
|
break |
|
print("[MagicPrompt script] Pregenerating finished") |
|
|
|
images = [] |
|
for i in range(state.job_count): |
|
if state.skipped: |
|
print("Rendering of current batch skipped") |
|
continue |
|
|
|
if state.interrupted: |
|
print(f"Rendering interrupted") |
|
break |
|
|
|
state.job = f"{i+1} out of {state.job_count}" |
|
|
|
|
|
sys.stdout.write('\033[2K\033[1G') |
|
print("\n") |
|
|
|
|
|
if (i == 0 or useUniquePrompt): |
|
if doPregenerating: |
|
p.prompt = prompts[i] |
|
else: |
|
|
|
print(f"[MagicPrompt script] Generating prompt for {getOrdinalNum(i+1)} batch...") |
|
p.prompt = self.gpt.generate_one(prompt=originalPrompt, max_length=promptLength, temperature=temp) |
|
|
|
print(f"[MagicPrompt script] Generated prompt for {getOrdinalNum(i+1)} batch: {p.prompt}") |
|
|
|
|
|
images += process_images(p).images |
|
|
|
if not useSameSeed: |
|
if not p.seed == -1: |
|
p.seed += 1 |
|
|
|
|
|
if doUnloadModel: |
|
del self.gpt |
|
|
|
return Processed(p, images, p.seed, "") |
|
|