fede97 commited on
Commit
98f60da
1 Parent(s): 4da3a88

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -3
README.md CHANGED
@@ -1,3 +1,31 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ ---
5
+ license: apache-2.0
6
+ datasets:
7
+ - liuhaotian/LLaVA-CC3M-Pretrain-595K
8
+ library_name: transformers
9
+ pipeline_tag: image-text-to-text
10
+ ---
11
+
12
+ # Model Card: LLaVA_MORE-llama_3_1-8B-S2-pretrain
13
+
14
+ ```LLaVA-MORE``` enhances the well-known LLaVA architecture by integrating the use of LLaMA 3.1 as the language model. We are publicly releasing the checkpoints for stages one and two for the first model with 8B parameters.
15
+
16
+ In this model space, you will find the stage one (pretrain) weights of LLaVA-MORE LLaMA 3.1 8B.
17
+
18
+ For more information, visit our [LLaVA-MORE](https://github.com/aimagelab/LLaVA-MORE) repository.
19
+
20
+
21
+ ## Citation
22
+ If you make use of our work, please cite our repo:
23
+
24
+ ```bibtex
25
+ @misc{cocchi2024llavamore,
26
+ title={{LLaVA-MORE: Enhancing Visual Instruction Tuning with LLaMA 3.1}},
27
+ author={Cocchi, Federico and Moratelli, Nicholas and Caffagni, Davide and Sarto, Sara and Cornia, Marcella and Baraldi, Lorenzo and Cucchiara, Rita},
28
+ url={https://github.com/aimagelab/LLaVA-MORE},
29
+ year={2024}
30
+ }
31
+ ```