Update README.md
Browse files
README.md
CHANGED
@@ -32,7 +32,7 @@ It is _**the largest open-source vision/vision-language foundation model (14B)**
|
|
32 |
- **Training Strategy:**
|
33 |
- Pretraining Stage
|
34 |
- Learnable Component: InternViT-6B
|
35 |
-
- Data: 72M samples
|
36 |
- SFT Stage
|
37 |
- Learnable Component: MLP + LLM
|
38 |
- Data: A comprehensive collection of open-source SFT datasets, along with their Chinese translation versions, totaling approximately 10M.
|
|
|
32 |
- **Training Strategy:**
|
33 |
- Pretraining Stage
|
34 |
- Learnable Component: InternViT-6B
|
35 |
+
- Data: Trained on 72M samples, including COYO, LAION, CC12M, CC3M, SBU, Wukong, GRIT, Objects365, OpenImages, and OCR data.
|
36 |
- SFT Stage
|
37 |
- Learnable Component: MLP + LLM
|
38 |
- Data: A comprehensive collection of open-source SFT datasets, along with their Chinese translation versions, totaling approximately 10M.
|