Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -16,12 +16,13 @@ def main():
|
|
16 |
global pipe
|
17 |
# device = "cuda"
|
18 |
if torch.cuda.is_available():
|
19 |
-
torch.cuda.max_memory_allocated(device=device)
|
20 |
# pipe = DiffusionPipeline.from_pretrained("stabilityai/sdxl-turbo", torch_dtype=torch.float16, variant="fp16", use_safetensors=True)
|
21 |
-
pipe.enable_xformers_memory_efficient_attention()
|
22 |
pipe = pipe.to(device)
|
23 |
else:
|
24 |
-
|
|
|
25 |
# pipe = pipe.to(device)
|
26 |
|
27 |
MAX_SEED = np.iinfo(np.int32).max
|
|
|
16 |
global pipe
|
17 |
# device = "cuda"
|
18 |
if torch.cuda.is_available():
|
19 |
+
#torch.cuda.max_memory_allocated(device=device)
|
20 |
# pipe = DiffusionPipeline.from_pretrained("stabilityai/sdxl-turbo", torch_dtype=torch.float16, variant="fp16", use_safetensors=True)
|
21 |
+
#pipe.enable_xformers_memory_efficient_attention()
|
22 |
pipe = pipe.to(device)
|
23 |
else:
|
24 |
+
pass
|
25 |
+
#pipe = DiffusionPipeline.from_pretrained("Lykon/AAM_XL_AnimeMix", use_safetensors=True)
|
26 |
# pipe = pipe.to(device)
|
27 |
|
28 |
MAX_SEED = np.iinfo(np.int32).max
|