Spaces:
Running
on
A10G
Running
on
A10G
Increase queue size, link to ViTL14 HF model
Browse files
app.py
CHANGED
@@ -29,7 +29,7 @@ blip_model.eval()
|
|
29 |
blip_model = blip_model.to(device)
|
30 |
|
31 |
print("Loading CLIP model...")
|
32 |
-
clip_model_name = 'ViT-L/14'
|
33 |
clip_model, clip_preprocess = clip.load(clip_model_name, device=device)
|
34 |
clip_model.to(device).eval()
|
35 |
|
@@ -241,5 +241,5 @@ io = gr.Interface(
|
|
241 |
article=article,
|
242 |
examples=[['example01.jpg'], ['example02.jpg']]
|
243 |
)
|
244 |
-
io.queue(max_size=
|
245 |
io.launch()
|
|
|
29 |
blip_model = blip_model.to(device)
|
30 |
|
31 |
print("Loading CLIP model...")
|
32 |
+
clip_model_name = 'ViT-L/14' # https://huggingface.co/openai/clip-vit-large-patch14
|
33 |
clip_model, clip_preprocess = clip.load(clip_model_name, device=device)
|
34 |
clip_model.to(device).eval()
|
35 |
|
|
|
241 |
article=article,
|
242 |
examples=[['example01.jpg'], ['example02.jpg']]
|
243 |
)
|
244 |
+
io.queue(max_size=32)
|
245 |
io.launch()
|