ruohongz commited on
Commit
2e6751c
1 Parent(s): 993a959

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -0
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: image-text-to-text
3
+ inference: false
4
+ license: apache-2.0
5
+ ---
6
+
7
+ <br>
8
+ <br>
9
+
10
+ # LLaVA-Reasoner Model Card
11
+
12
+ ## Model details
13
+
14
+ **Model type:**
15
+ LLaVA-Reasoner is an open-source image vision language model, fine-tuned from GPT4-o distilled chain-of-thought (CoT) reasoning data.
16
+
17
+ This model is the **SFT** version, with additional math data compared to SFT-preview.
18
+
19
+ Base LLM: [Lin-Chen/open-llava-next-llama3-8b](https://huggingface.co/Lin-Chen/open-llava-next-llama3-8b)
20
+
21
+ **Model date:**
22
+ Trained on Sep, 2024.
23
+
24
+ **Paper or resources for more information:**
25
+
26
+ Paper: https://arxiv.org/abs/2410.16198
27
+
28
+ Code: https://github.com/RifleZhang/LLaVA-Reasoner-DPO/tree/main
29
+
30
+ ## License
31
+ [Lin-Chen/open-llava-next-llama3-8b](https://huggingface.co/Lin-Chen/open-llava-next-llama3-8b) license.
32
+
33
+ **Where to send questions or comments about the model:**
34
+ https://github.com/RifleZhang/LLaVA-Reasoner-DPO/issues
35
+
36
+ ## Intended use
37
+ **Primary intended uses:**
38
+ Image CoT reasoning
39
+
40
+ **Primary intended users:**
41
+ Researchers in artificial intelligence, large multimodal model, etc.
42
+
43
+ ## Training dataset
44
+ [ShareGPT4o-Reasoning](https://huggingface.co/datasets/Share4oReasoning/sft_data) dataset.
45
+
46
+ ## Evaluation
47
+ Follow https://github.com/RifleZhang/LLaVA-Reasoner-DPO/blob/main/README.md
48
+
49
+
50
+ ## citation
51
+ ```
52
+ @article{zhang2024improve,
53
+ title={Improve vision language model chain-of-thought reasoning},
54
+ author={Zhang, Ruohong and Zhang, Bowen and Li, Yanghao and Zhang, Haotian and Sun, Zhiqing and Gan, Zhe and Yang, Yinfei and Pang, Ruoming and Yang, Yiming},
55
+ journal={arXiv preprint arXiv:2410.16198},
56
+ year={2024}
57
+ }
58
+ ```