Triangle104 commited on
Commit
d182fcc
1 Parent(s): 66330ed

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +204 -0
README.md ADDED
@@ -0,0 +1,204 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: llama3.1
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ - shining-valiant
10
+ - shining-valiant-2
11
+ - cobalt
12
+ - plum
13
+ - valiant
14
+ - valiant-labs
15
+ - llama
16
+ - llama-3.1
17
+ - llama-3.1-instruct
18
+ - llama-3.1-instruct-8b
19
+ - llama-3
20
+ - llama-3-instruct
21
+ - llama-3-instruct-8b
22
+ - 8b
23
+ - math
24
+ - math-instruct
25
+ - science
26
+ - physics
27
+ - biology
28
+ - chemistry
29
+ - compsci
30
+ - computer-science
31
+ - engineering
32
+ - technical
33
+ - conversational
34
+ - chat
35
+ - instruct
36
+ - llama-cpp
37
+ - gguf-my-repo
38
+ base_model: sequelbox/Llama3.1-8B-PlumMath
39
+ pipeline_tag: text-generation
40
+ model_type: llama
41
+ model-index:
42
+ - name: sequelbox/Llama3.1-8B-PlumMath
43
+ results:
44
+ - task:
45
+ type: text-generation
46
+ name: Text Generation
47
+ dataset:
48
+ name: Winogrande (5-Shot)
49
+ type: Winogrande
50
+ args:
51
+ num_few_shot: 5
52
+ metrics:
53
+ - type: acc
54
+ value: 72.38
55
+ name: acc
56
+ - task:
57
+ type: text-generation
58
+ name: Text Generation
59
+ dataset:
60
+ name: MathQA (5-Shot)
61
+ type: MathQA
62
+ args:
63
+ num_few_shot: 5
64
+ metrics:
65
+ - type: acc
66
+ value: 40.27
67
+ name: acc
68
+ - task:
69
+ type: text-generation
70
+ name: Text Generation
71
+ dataset:
72
+ name: IFEval (0-Shot)
73
+ type: HuggingFaceH4/ifeval
74
+ args:
75
+ num_few_shot: 0
76
+ metrics:
77
+ - type: inst_level_strict_acc and prompt_level_strict_acc
78
+ value: 22.42
79
+ name: strict accuracy
80
+ source:
81
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumMath
82
+ name: Open LLM Leaderboard
83
+ - task:
84
+ type: text-generation
85
+ name: Text Generation
86
+ dataset:
87
+ name: BBH (3-Shot)
88
+ type: BBH
89
+ args:
90
+ num_few_shot: 3
91
+ metrics:
92
+ - type: acc_norm
93
+ value: 16.45
94
+ name: normalized accuracy
95
+ source:
96
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumMath
97
+ name: Open LLM Leaderboard
98
+ - task:
99
+ type: text-generation
100
+ name: Text Generation
101
+ dataset:
102
+ name: MATH Lvl 5 (4-Shot)
103
+ type: hendrycks/competition_math
104
+ args:
105
+ num_few_shot: 4
106
+ metrics:
107
+ - type: exact_match
108
+ value: 3.93
109
+ name: exact match
110
+ source:
111
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumMath
112
+ name: Open LLM Leaderboard
113
+ - task:
114
+ type: text-generation
115
+ name: Text Generation
116
+ dataset:
117
+ name: GPQA (0-shot)
118
+ type: Idavidrein/gpqa
119
+ args:
120
+ num_few_shot: 0
121
+ metrics:
122
+ - type: acc_norm
123
+ value: 9.06
124
+ name: acc_norm
125
+ source:
126
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumMath
127
+ name: Open LLM Leaderboard
128
+ - task:
129
+ type: text-generation
130
+ name: Text Generation
131
+ dataset:
132
+ name: MuSR (0-shot)
133
+ type: TAUR-Lab/MuSR
134
+ args:
135
+ num_few_shot: 0
136
+ metrics:
137
+ - type: acc_norm
138
+ value: 8.98
139
+ name: acc_norm
140
+ source:
141
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumMath
142
+ name: Open LLM Leaderboard
143
+ - task:
144
+ type: text-generation
145
+ name: Text Generation
146
+ dataset:
147
+ name: MMLU-PRO (5-shot)
148
+ type: TIGER-Lab/MMLU-Pro
149
+ config: main
150
+ split: test
151
+ args:
152
+ num_few_shot: 5
153
+ metrics:
154
+ - type: acc
155
+ value: 21.95
156
+ name: accuracy
157
+ source:
158
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sequelbox/Llama3.1-8B-PlumMath
159
+ name: Open LLM Leaderboard
160
+ ---
161
+
162
+ # Triangle104/Llama3.1-8B-PlumMath-Q6_K-GGUF
163
+ This model was converted to GGUF format from [`sequelbox/Llama3.1-8B-PlumMath`](https://huggingface.co/sequelbox/Llama3.1-8B-PlumMath) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
164
+ Refer to the [original model card](https://huggingface.co/sequelbox/Llama3.1-8B-PlumMath) for more details on the model.
165
+
166
+ ## Use with llama.cpp
167
+ Install llama.cpp through brew (works on Mac and Linux)
168
+
169
+ ```bash
170
+ brew install llama.cpp
171
+
172
+ ```
173
+ Invoke the llama.cpp server or the CLI.
174
+
175
+ ### CLI:
176
+ ```bash
177
+ llama-cli --hf-repo Triangle104/Llama3.1-8B-PlumMath-Q6_K-GGUF --hf-file llama3.1-8b-plummath-q6_k.gguf -p "The meaning to life and the universe is"
178
+ ```
179
+
180
+ ### Server:
181
+ ```bash
182
+ llama-server --hf-repo Triangle104/Llama3.1-8B-PlumMath-Q6_K-GGUF --hf-file llama3.1-8b-plummath-q6_k.gguf -c 2048
183
+ ```
184
+
185
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
186
+
187
+ Step 1: Clone llama.cpp from GitHub.
188
+ ```
189
+ git clone https://github.com/ggerganov/llama.cpp
190
+ ```
191
+
192
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
193
+ ```
194
+ cd llama.cpp && LLAMA_CURL=1 make
195
+ ```
196
+
197
+ Step 3: Run inference through the main binary.
198
+ ```
199
+ ./llama-cli --hf-repo Triangle104/Llama3.1-8B-PlumMath-Q6_K-GGUF --hf-file llama3.1-8b-plummath-q6_k.gguf -p "The meaning to life and the universe is"
200
+ ```
201
+ or
202
+ ```
203
+ ./llama-server --hf-repo Triangle104/Llama3.1-8B-PlumMath-Q6_K-GGUF --hf-file llama3.1-8b-plummath-q6_k.gguf -c 2048
204
+ ```