Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,6 @@ llava-dinov2-internlm2-7b-v1 is a LLaVA model fine-tuned from [InternLM2-Chat-7B
|
|
11 |
I did not carefully tune the training hyperparameters but the model still show capability to solve some tasks. It shows that a visual encoder can be integrated with an LLM, even when the encoder is not aligned with natural language with contrastive learning like CLIP.
|
12 |
|
13 |
## Example
|
14 |
-
|
15 |

|
16 |
Explain the photo in English:
|
17 |

|
@@ -77,9 +76,13 @@ You just need
|
|
77 |
```
|
78 |
pip install protobuf
|
79 |
```
|
|
|
|
|
|
|
|
|
|
|
80 |
|
81 |
## Data prepration
|
82 |
-
|
83 |
1. File structure
|
84 |
|
85 |
```
|
|
|
11 |
I did not carefully tune the training hyperparameters but the model still show capability to solve some tasks. It shows that a visual encoder can be integrated with an LLM, even when the encoder is not aligned with natural language with contrastive learning like CLIP.
|
12 |
|
13 |
## Example
|
|
|
14 |

|
15 |
Explain the photo in English:
|
16 |

|
|
|
76 |
```
|
77 |
pip install protobuf
|
78 |
```
|
79 |
+
4.
|
80 |
+
To use tensorboard to visualize the training loss curve:
|
81 |
+
```
|
82 |
+
pip install future tensorboard
|
83 |
+
```
|
84 |
|
85 |
## Data prepration
|
|
|
86 |
1. File structure
|
87 |
|
88 |
```
|