LinWeizheDragon
commited on
Commit
•
67d9380
1
Parent(s):
f4edb9c
Update README.md
Browse files
README.md
CHANGED
@@ -3,6 +3,12 @@ library_name: transformers
|
|
3 |
license: mit
|
4 |
language:
|
5 |
- en
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
---
|
7 |
|
8 |
# PreFLMR model card
|
@@ -37,11 +43,11 @@ This model can be used combined with language models to create a retrieval-augme
|
|
37 |
|
38 |
## How to Get Started with the Model
|
39 |
|
40 |
-
For details of training, indexing and performing retrieval, please refer to [here](https://github.com/LinWeizheDragon/FLMR).
|
41 |
|
42 |
## Training datasets
|
43 |
-
The model is
|
44 |
-
1. Image to Text retrieval: WIT, KVQA and CC3M
|
45 |
2. Question to Text retrieval: MSMARCO
|
46 |
3. Image & Question to Text retrieval: LLaVA, OVEN, OKVQA, Infoseek and E-VQA
|
47 |
|
|
|
3 |
license: mit
|
4 |
language:
|
5 |
- en
|
6 |
+
tags:
|
7 |
+
- retrieval
|
8 |
+
- multi-modal
|
9 |
+
- knowledge-based visual question answering
|
10 |
+
- FLMR
|
11 |
+
- PreFLMR
|
12 |
---
|
13 |
|
14 |
# PreFLMR model card
|
|
|
43 |
|
44 |
## How to Get Started with the Model
|
45 |
|
46 |
+
For details of training, indexing, and performing retrieval, please refer to [here](https://github.com/LinWeizheDragon/FLMR).
|
47 |
|
48 |
## Training datasets
|
49 |
+
The model is pre-trained on three types of tasks with a total of nine datasets:
|
50 |
+
1. Image to Text retrieval: WIT, KVQA, and CC3M
|
51 |
2. Question to Text retrieval: MSMARCO
|
52 |
3. Image & Question to Text retrieval: LLaVA, OVEN, OKVQA, Infoseek and E-VQA
|
53 |
|