win10 commited on
Commit
f0771c6
1 Parent(s): e7d6d1d

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +161 -0
README.md ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - zh
5
+ license: apache-2.0
6
+ library_name: transformers
7
+ tags:
8
+ - chat
9
+ - conversational
10
+ - custom-research
11
+ - rp
12
+ - roleplay
13
+ - tool-use
14
+ - llama-cpp
15
+ - gguf-my-repo
16
+ base_model: sthenno-com/miscii-14b-1028
17
+ pipeline_tag: text-generation
18
+ datasets:
19
+ - nvidia/HelpSteer2
20
+ - google/Synthetic-Persona-Chat
21
+ - mlabonne/orpo-dpo-mix-40k
22
+ model-index:
23
+ - name: miscii-14b-1028
24
+ results:
25
+ - task:
26
+ type: text-generation
27
+ name: Text Generation
28
+ dataset:
29
+ name: IFEval (0-Shot)
30
+ type: HuggingFaceH4/ifeval
31
+ args:
32
+ num_few_shot: 0
33
+ metrics:
34
+ - type: inst_level_strict_acc and prompt_level_strict_acc
35
+ value: 82.37
36
+ name: strict accuracy
37
+ source:
38
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno-com/miscii-14b-1028
39
+ name: Open LLM Leaderboard
40
+ - task:
41
+ type: text-generation
42
+ name: Text Generation
43
+ dataset:
44
+ name: BBH (3-Shot)
45
+ type: BBH
46
+ args:
47
+ num_few_shot: 3
48
+ metrics:
49
+ - type: acc_norm
50
+ value: 49.26
51
+ name: normalized accuracy
52
+ source:
53
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno-com/miscii-14b-1028
54
+ name: Open LLM Leaderboard
55
+ - task:
56
+ type: text-generation
57
+ name: Text Generation
58
+ dataset:
59
+ name: MATH Lvl 5 (4-Shot)
60
+ type: hendrycks/competition_math
61
+ args:
62
+ num_few_shot: 4
63
+ metrics:
64
+ - type: exact_match
65
+ value: 6.34
66
+ name: exact match
67
+ source:
68
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno-com/miscii-14b-1028
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: GPQA (0-shot)
75
+ type: Idavidrein/gpqa
76
+ args:
77
+ num_few_shot: 0
78
+ metrics:
79
+ - type: acc_norm
80
+ value: 14.21
81
+ name: acc_norm
82
+ source:
83
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno-com/miscii-14b-1028
84
+ name: Open LLM Leaderboard
85
+ - task:
86
+ type: text-generation
87
+ name: Text Generation
88
+ dataset:
89
+ name: MuSR (0-shot)
90
+ type: TAUR-Lab/MuSR
91
+ args:
92
+ num_few_shot: 0
93
+ metrics:
94
+ - type: acc_norm
95
+ value: 12
96
+ name: acc_norm
97
+ source:
98
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno-com/miscii-14b-1028
99
+ name: Open LLM Leaderboard
100
+ - task:
101
+ type: text-generation
102
+ name: Text Generation
103
+ dataset:
104
+ name: MMLU-PRO (5-shot)
105
+ type: TIGER-Lab/MMLU-Pro
106
+ config: main
107
+ split: test
108
+ args:
109
+ num_few_shot: 5
110
+ metrics:
111
+ - type: acc
112
+ value: 46.14
113
+ name: accuracy
114
+ source:
115
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=sthenno-com/miscii-14b-1028
116
+ name: Open LLM Leaderboard
117
+ ---
118
+
119
+ # win10/miscii-14b-1028-Q8_0-GGUF
120
+ This model was converted to GGUF format from [`sthenno-com/miscii-14b-1028`](https://huggingface.co/sthenno-com/miscii-14b-1028) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
121
+ Refer to the [original model card](https://huggingface.co/sthenno-com/miscii-14b-1028) for more details on the model.
122
+
123
+ ## Use with llama.cpp
124
+ Install llama.cpp through brew (works on Mac and Linux)
125
+
126
+ ```bash
127
+ brew install llama.cpp
128
+
129
+ ```
130
+ Invoke the llama.cpp server or the CLI.
131
+
132
+ ### CLI:
133
+ ```bash
134
+ llama-cli --hf-repo win10/miscii-14b-1028-Q8_0-GGUF --hf-file miscii-14b-1028-q8_0.gguf -p "The meaning to life and the universe is"
135
+ ```
136
+
137
+ ### Server:
138
+ ```bash
139
+ llama-server --hf-repo win10/miscii-14b-1028-Q8_0-GGUF --hf-file miscii-14b-1028-q8_0.gguf -c 2048
140
+ ```
141
+
142
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
143
+
144
+ Step 1: Clone llama.cpp from GitHub.
145
+ ```
146
+ git clone https://github.com/ggerganov/llama.cpp
147
+ ```
148
+
149
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
150
+ ```
151
+ cd llama.cpp && LLAMA_CURL=1 make
152
+ ```
153
+
154
+ Step 3: Run inference through the main binary.
155
+ ```
156
+ ./llama-cli --hf-repo win10/miscii-14b-1028-Q8_0-GGUF --hf-file miscii-14b-1028-q8_0.gguf -p "The meaning to life and the universe is"
157
+ ```
158
+ or
159
+ ```
160
+ ./llama-server --hf-repo win10/miscii-14b-1028-Q8_0-GGUF --hf-file miscii-14b-1028-q8_0.gguf -c 2048
161
+ ```