nisten's picture
Update README.md
d156677
|
raw
history blame
907 Bytes
metadata
license: mit

NousResearch - Obsidian-3B-V0.5 gguf q6 version by Nisten

To run the server inside /llama.cpp/ folder IN YOUR TERMINAL

./server -m obsidian-q6.gguf --mmproj mmproj-obsidian-f16.gguf -ngl 42

that's it, it's literally one command, you open your browser now at http://127.0.0.1:8080

How to Run the server and download the models all in one command, you need ~2.9gb of space

FIRST TIME TO RUN mac or linux, make a new folder, THIS WILL DOWNLOAD AND RUN EVERYTHIN whole in ONE SHOT

git clone -b stablelm-support https://github.com/Galunid/llama.cpp.git && \
cd llama.cpp && \
make && \
wget https://huggingface.co/nisten/obsidian-3b-multimodal-q6-gguf/resolve/main/obsidian-q6.gguf && \
wget https://huggingface.co/nisten/obsidian-3b-multimodal-q6-gguf/resolve/main/mmproj-obsidian-f16.gguf && \
./server -m obsidian-q6.gguf --mmproj mmproj-obsidian-f16.gguf -ngl 42