czczup commited on
Commit
a657451
1 Parent(s): d9a286a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -14
README.md CHANGED
@@ -2,18 +2,14 @@
2
  license: llama3
3
  pipeline_tag: image-text-to-text
4
  library_name: transformers
 
 
 
5
  language:
6
  - multilingual
7
  tags:
8
  - internvl
9
- - vision
10
- - ocr
11
- - multi-image
12
- - video
13
  - custom_code
14
- base_model: OpenGVLab/InternVL2-Llama3-76B
15
- base_model_relation: quantized
16
- new_version: OpenGVLab/InternVL2_5-78B-AWQ
17
  ---
18
 
19
  # InternVL2-Llama3-76B-AWQ
@@ -117,7 +113,7 @@ print(response)
117
 
118
  ## License
119
 
120
- This project is released under the MIT license, while Llama3 is licensed under the Llama 3 Community License.
121
 
122
  ## Citation
123
 
@@ -130,16 +126,16 @@ If you find this project useful in your research, please consider citing:
130
  journal={arXiv preprint arXiv:2410.16261},
131
  year={2024}
132
  }
133
- @article{chen2023internvl,
134
- title={InternVL: Scaling up Vision Foundation Models and Aligning for Generic Visual-Linguistic Tasks},
135
- author={Chen, Zhe and Wu, Jiannan and Wang, Wenhai and Su, Weijie and Chen, Guo and Xing, Sen and Zhong, Muyan and Zhang, Qinglong and Zhu, Xizhou and Lu, Lewei and Li, Bin and Luo, Ping and Lu, Tong and Qiao, Yu and Dai, Jifeng},
136
- journal={arXiv preprint arXiv:2312.14238},
137
- year={2023}
138
- }
139
  @article{chen2024far,
140
  title={How Far Are We to GPT-4V? Closing the Gap to Commercial Multimodal Models with Open-Source Suites},
141
  author={Chen, Zhe and Wang, Weiyun and Tian, Hao and Ye, Shenglong and Gao, Zhangwei and Cui, Erfei and Tong, Wenwen and Hu, Kongzhi and Luo, Jiapeng and Ma, Zheng and others},
142
  journal={arXiv preprint arXiv:2404.16821},
143
  year={2024}
144
  }
 
 
 
 
 
 
145
  ```
 
2
  license: llama3
3
  pipeline_tag: image-text-to-text
4
  library_name: transformers
5
+ base_model: OpenGVLab/InternVL2-Llama3-76B
6
+ new_version: OpenGVLab/InternVL2_5-78B-AWQ
7
+ base_model_relation: quantized
8
  language:
9
  - multilingual
10
  tags:
11
  - internvl
 
 
 
 
12
  - custom_code
 
 
 
13
  ---
14
 
15
  # InternVL2-Llama3-76B-AWQ
 
113
 
114
  ## License
115
 
116
+ This project is released under the MIT License. This project uses the pre-trained Hermes-2-Theta-Llama-3-70B as a component, which is licensed under the Llama 3 Community License.
117
 
118
  ## Citation
119
 
 
126
  journal={arXiv preprint arXiv:2410.16261},
127
  year={2024}
128
  }
 
 
 
 
 
 
129
  @article{chen2024far,
130
  title={How Far Are We to GPT-4V? Closing the Gap to Commercial Multimodal Models with Open-Source Suites},
131
  author={Chen, Zhe and Wang, Weiyun and Tian, Hao and Ye, Shenglong and Gao, Zhangwei and Cui, Erfei and Tong, Wenwen and Hu, Kongzhi and Luo, Jiapeng and Ma, Zheng and others},
132
  journal={arXiv preprint arXiv:2404.16821},
133
  year={2024}
134
  }
135
+ @article{chen2023internvl,
136
+ title={InternVL: Scaling up Vision Foundation Models and Aligning for Generic Visual-Linguistic Tasks},
137
+ author={Chen, Zhe and Wu, Jiannan and Wang, Wenhai and Su, Weijie and Chen, Guo and Xing, Sen and Zhong, Muyan and Zhang, Qinglong and Zhu, Xizhou and Lu, Lewei and Li, Bin and Luo, Ping and Lu, Tong and Qiao, Yu and Dai, Jifeng},
138
+ journal={arXiv preprint arXiv:2312.14238},
139
+ year={2023}
140
+ }
141
  ```