iakarshu commited on
Commit
75d6c3b
1 Parent(s): 751479f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -132,7 +132,7 @@ def answer_question(image, question):
132
 
133
 
134
  # Taken from here: https://huggingface.co/spaces/nielsr/vilt-vqa/blob/main/app.py
135
- title = "Interactive demo: laTr (Layout Aware Transformer) for VQA"
136
  description = "Gradio Demo for LaTr (Layout Aware Transformer),trained on TextVQA Dataset. To use it, simply upload your image and type a question and click 'submit', or click one of the examples to load them. Read more at the links below."
137
  article = "<p style='text-align: center'><a href='https://arxiv.org/abs/2112.12494' target='_blank'>LaTr: Layout-aware transformer for scene-text VQA,a novel multimodal architecture for Scene Text Visual Question Answering (STVQA)</a> | <a href='https://github.com/uakarsh/latr' target='_blank'>Github Repo</a></p>"
138
  examples = [['remote.png', "Is remote present in the picture?"]]
 
132
 
133
 
134
  # Taken from here: https://huggingface.co/spaces/nielsr/vilt-vqa/blob/main/app.py
135
+ title = "Interactive demo: LaTr (Layout Aware Transformer) for VQA"
136
  description = "Gradio Demo for LaTr (Layout Aware Transformer),trained on TextVQA Dataset. To use it, simply upload your image and type a question and click 'submit', or click one of the examples to load them. Read more at the links below."
137
  article = "<p style='text-align: center'><a href='https://arxiv.org/abs/2112.12494' target='_blank'>LaTr: Layout-aware transformer for scene-text VQA,a novel multimodal architecture for Scene Text Visual Question Answering (STVQA)</a> | <a href='https://github.com/uakarsh/latr' target='_blank'>Github Repo</a></p>"
138
  examples = [['remote.png', "Is remote present in the picture?"]]