Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -67,15 +67,16 @@ def generate(
|
|
67 |
max_new_tokens=512, temperature=0.5, top_p=0.85, top_k=50, repetition_penalty=1.05)
|
68 |
llm_result = llm.generate(**generation_kwargs)
|
69 |
llm_result = BOT_PREFIX + tokenizer.decode(llm_result.cpu()[0], skip_special_tokens=True)
|
|
|
70 |
print(llm_result)
|
71 |
-
print(
|
72 |
|
73 |
seed = random.randint(0, 2147483647)
|
74 |
diffusion_pipe.to(device)
|
75 |
generator = torch.Generator().manual_seed(seed)
|
76 |
|
77 |
images = diffusion_pipe(
|
78 |
-
prompt=
|
79 |
negative_prompt=None,
|
80 |
width=1024,
|
81 |
height=1024,
|
|
|
67 |
max_new_tokens=512, temperature=0.5, top_p=0.85, top_k=50, repetition_penalty=1.05)
|
68 |
llm_result = llm.generate(**generation_kwargs)
|
69 |
llm_result = BOT_PREFIX + tokenizer.decode(llm_result.cpu()[0], skip_special_tokens=True)
|
70 |
+
expanded_description = json.loads(llm_result)["expanded_description"]
|
71 |
print(llm_result)
|
72 |
+
print(expanded_description)
|
73 |
|
74 |
seed = random.randint(0, 2147483647)
|
75 |
diffusion_pipe.to(device)
|
76 |
generator = torch.Generator().manual_seed(seed)
|
77 |
|
78 |
images = diffusion_pipe(
|
79 |
+
prompt=expanded_description,
|
80 |
negative_prompt=None,
|
81 |
width=1024,
|
82 |
height=1024,
|