Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -32,8 +32,8 @@ JS = """function () {
|
|
32 |
}
|
33 |
}"""
|
34 |
|
35 |
-
|
36 |
-
|
37 |
|
38 |
def enable_lora(lora_in, lora_add):
|
39 |
if not lora_in and not lora_add:
|
@@ -61,7 +61,7 @@ async def generate_image(
|
|
61 |
|
62 |
#generator = torch.Generator().manual_seed(seed)
|
63 |
|
64 |
-
image1 = await
|
65 |
prompt=text,
|
66 |
height=height,
|
67 |
width=width,
|
@@ -69,7 +69,7 @@ async def generate_image(
|
|
69 |
num_inference_steps=steps,
|
70 |
model=basemodel,
|
71 |
)
|
72 |
-
image2 = await
|
73 |
prompt=text,
|
74 |
height=height,
|
75 |
width=width,
|
|
|
32 |
}
|
33 |
}"""
|
34 |
|
35 |
+
client1 = AsyncInferenceClient()
|
36 |
+
client2 = AsyncInferenceClient()
|
37 |
|
38 |
def enable_lora(lora_in, lora_add):
|
39 |
if not lora_in and not lora_add:
|
|
|
61 |
|
62 |
#generator = torch.Generator().manual_seed(seed)
|
63 |
|
64 |
+
image1 = await client1.text_to_image(
|
65 |
prompt=text,
|
66 |
height=height,
|
67 |
width=width,
|
|
|
69 |
num_inference_steps=steps,
|
70 |
model=basemodel,
|
71 |
)
|
72 |
+
image2 = await client2.text_to_image(
|
73 |
prompt=text,
|
74 |
height=height,
|
75 |
width=width,
|