matichon commited on
Commit
84009f5
·
verified ·
1 Parent(s): 78fd707

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -3
README.md CHANGED
@@ -1,3 +1,42 @@
1
- ---
2
- license: llama3.1
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.1
3
+ pipeline_tag: text-generation
4
+ base_model: scb10x/llama3.1-typhoon2-8b-instruct
5
+ tags:
6
+ - llama-cpp
7
+ - gguf-my-repo
8
+ ---
9
+ # Float16-cloud/llama3.1-typhoon2-8b-instruct-gguf
10
+ This model was converted to GGUF format from [`scb10x/llama3.1-typhoon2-8b-instruct`](https://huggingface.co/scb10x/llama3.1-typhoon2-8b-instruct) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
11
+ Refer to the [original model card](https://huggingface.co/scb10x/llama3.1-typhoon2-8b-instruct) for more details on the model.
12
+ ## Use with llama.cpp
13
+ Install llama.cpp through brew (works on Mac and Linux)
14
+ ```bash
15
+ brew install llama.cpp
16
+ ```
17
+ Invoke the llama.cpp server or the CLI.
18
+ ### CLI:
19
+ ```bash
20
+ llama-cli --hf-repo Float16-cloud/llama3.1-typhoon2-8b-instruct-gguf --hf-file llama3.1-typhoon2-8b-instruct-iq4_nl.gguf -p "The meaning to life and the universe is"
21
+ ```
22
+ ### Server:
23
+ ```bash
24
+ llama-server --hf-repo Float16-cloud/llama3.1-typhoon2-8b-instruct-gguf --hf-file llama3.1-typhoon2-8b-instruct-iq4_nl.gguf -c 2048
25
+ ```
26
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
27
+ Step 1: Clone llama.cpp from GitHub.
28
+ ```
29
+ git clone https://github.com/ggerganov/llama.cpp
30
+ ```
31
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
32
+ ```
33
+ cd llama.cpp && LLAMA_CURL=1 make
34
+ ```
35
+ Step 3: Run inference through the main binary.
36
+ ```
37
+ ./llama-cli --hf-repo Float16-cloud/llama3.1-typhoon2-8b-instruct-gguf --hf-file llama3.1-typhoon2-8b-instruct-iq4_nl.gguf -p "The meaning to life and the universe is"
38
+ ```
39
+ or
40
+ ```
41
+ ./llama-server --hf-repo Float16-cloud/llama3.1-typhoon2-8b-instruct-gguf --hf-file llama3.1-typhoon2-8b-instruct-iq4_nl.gguf -c 2048
42
+ ```