Update README.md
Browse files
README.md
CHANGED
@@ -63,12 +63,13 @@ wget https://huggingface.co/xtuner/llava-phi-3-mini-gguf/resolve/main/llava-phi-
|
|
63 |
|
64 |
# int4 llm
|
65 |
wget https://huggingface.co/xtuner/llava-phi-3-mini-gguf/resolve/main/llava-phi-3-mini-int4.gguf
|
66 |
-
```
|
67 |
|
68 |
-
|
|
|
69 |
|
70 |
-
|
71 |
-
|
|
|
72 |
|
73 |
### Chat by `ollama`
|
74 |
|
@@ -86,6 +87,9 @@ ollama run llava-phi3-int4 "xx.png Describe this image"
|
|
86 |
|
87 |
### Chat by `./llava-cli`
|
88 |
|
|
|
|
|
|
|
89 |
Note: llava-phi-3-mini uses the `Phi-3-instruct` chat template.
|
90 |
|
91 |
```bash
|
|
|
63 |
|
64 |
# int4 llm
|
65 |
wget https://huggingface.co/xtuner/llava-phi-3-mini-gguf/resolve/main/llava-phi-3-mini-int4.gguf
|
|
|
66 |
|
67 |
+
# (optional) ollama fp16 modelfile
|
68 |
+
wget https://huggingface.co/xtuner/llava-phi-3-mini-gguf/resolve/main/OLLAMA_MODELFILE_F16
|
69 |
|
70 |
+
# (optional) ollama int4 modelfile
|
71 |
+
wget https://huggingface.co/xtuner/llava-phi-3-mini-gguf/resolve/main/OLLAMA_MODELFILE_INT4
|
72 |
+
```
|
73 |
|
74 |
### Chat by `ollama`
|
75 |
|
|
|
87 |
|
88 |
### Chat by `./llava-cli`
|
89 |
|
90 |
+
1. Build [llama.cpp](https://github.com/ggerganov/llama.cpp) ([docs](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage)) .
|
91 |
+
2. Build `./llava-cli` ([docs](https://github.com/ggerganov/llama.cpp/tree/master/examples/llava#usage)).
|
92 |
+
|
93 |
Note: llava-phi-3-mini uses the `Phi-3-instruct` chat template.
|
94 |
|
95 |
```bash
|