add examples
Browse files- app.py +2 -10
- bike.jpg +0 -0
- surfer.jpg +0 -0
app.py
CHANGED
@@ -1,11 +1,3 @@
|
|
1 |
-
# import gradio as gr
|
2 |
-
|
3 |
-
# def greet(image):
|
4 |
-
# return "Shape " + image.shape + "!!"
|
5 |
-
|
6 |
-
# iface = gr.Interface(fn=greet, inputs="image", outputs="text")
|
7 |
-
# iface.launch()
|
8 |
-
|
9 |
import gradio as gr
|
10 |
import sys
|
11 |
from BLIP.models.blip import blip_decoder
|
@@ -42,7 +34,7 @@ inputs = [gr.Image(type='pil', interactive=False),]
|
|
42 |
outputs = gr.outputs.Textbox(label="Caption")
|
43 |
|
44 |
description = "Gradio demo for FuseCap: Leveraging Large Language Models to Fuse Visual Data into Enriched Image Captions. This demo features a BLIP-based model, trained using FuseCap."
|
45 |
-
|
46 |
article = "<p style='text-align: center'><a href='google.com' target='_blank'>place holder</a>/p>"
|
47 |
|
48 |
|
@@ -52,6 +44,6 @@ iface = gr.Interface(fn=inference,
|
|
52 |
title="FuseCap",
|
53 |
description=description,
|
54 |
article=article,
|
55 |
-
examples=
|
56 |
enable_queue=True)
|
57 |
iface.launch()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
import gradio as gr
|
2 |
import sys
|
3 |
from BLIP.models.blip import blip_decoder
|
|
|
34 |
outputs = gr.outputs.Textbox(label="Caption")
|
35 |
|
36 |
description = "Gradio demo for FuseCap: Leveraging Large Language Models to Fuse Visual Data into Enriched Image Captions. This demo features a BLIP-based model, trained using FuseCap."
|
37 |
+
examples = [["birthday_dog.jpeg"], ["surfer.png"], ["bike.jpg"]]
|
38 |
article = "<p style='text-align: center'><a href='google.com' target='_blank'>place holder</a>/p>"
|
39 |
|
40 |
|
|
|
44 |
title="FuseCap",
|
45 |
description=description,
|
46 |
article=article,
|
47 |
+
examples=examples,
|
48 |
enable_queue=True)
|
49 |
iface.launch()
|
bike.jpg
ADDED
surfer.jpg
ADDED