noxinc commited on
Commit
543a07c
·
verified ·
1 Parent(s): 4fa11b5

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +289 -0
README.md ADDED
@@ -0,0 +1,289 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - pt
4
+ license: apache-2.0
5
+ library_name: transformers
6
+ tags:
7
+ - text-generation-inference
8
+ - llama-cpp
9
+ - gguf-my-repo
10
+ datasets:
11
+ - TucanoBR/GigaVerbo
12
+ metrics:
13
+ - perplexity
14
+ pipeline_tag: text-generation
15
+ widget:
16
+ - text: A floresta da Amazônia é conhecida por sua
17
+ example_title: Exemplo
18
+ - text: Uma das coisas que Portugal, Angola, Brasil e Moçambique tem em comum é o
19
+ example_title: Exemplo
20
+ - text: O Carnaval do Rio de Janeiro é
21
+ example_title: Exemplo
22
+ inference:
23
+ parameters:
24
+ repetition_penalty: 1.2
25
+ temperature: 0.2
26
+ top_k: 20
27
+ top_p: 0.2
28
+ max_new_tokens: 150
29
+ co2_eq_emissions:
30
+ emissions: 4475000
31
+ source: CodeCarbon
32
+ training_type: pre-training
33
+ geographical_location: Germany
34
+ hardware_used: NVIDIA A100-SXM4-80GB
35
+ model-index:
36
+ - name: Tucano-2b4
37
+ results:
38
+ - task:
39
+ type: text-generation
40
+ name: Text Generation
41
+ dataset:
42
+ name: CALAME-PT
43
+ type: NOVA-vision-language/calame-pt
44
+ split: all
45
+ args:
46
+ num_few_shot: 0
47
+ metrics:
48
+ - type: acc
49
+ value: 59.06
50
+ name: accuracy
51
+ source:
52
+ url: https://huggingface.co/datasets/NOVA-vision-language/calame-pt
53
+ name: Context-Aware LAnguage Modeling Evaluation for Portuguese
54
+ - task:
55
+ type: text-generation
56
+ name: Text Generation
57
+ dataset:
58
+ name: LAMBADA-PT
59
+ type: TucanoBR/lambada-pt
60
+ split: train
61
+ args:
62
+ num_few_shot: 0
63
+ metrics:
64
+ - type: acc
65
+ value: 37.67
66
+ name: accuracy
67
+ source:
68
+ url: https://huggingface.co/datasets/TucanoBR/lambada-pt
69
+ name: LAMBADA-PT
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: ENEM Challenge (No Images)
75
+ type: eduagarcia/enem_challenge
76
+ split: train
77
+ args:
78
+ num_few_shot: 3
79
+ metrics:
80
+ - type: acc
81
+ value: 20.5
82
+ name: accuracy
83
+ source:
84
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
85
+ name: Open Portuguese LLM Leaderboard
86
+ - task:
87
+ type: text-generation
88
+ name: Text Generation
89
+ dataset:
90
+ name: BLUEX (No Images)
91
+ type: eduagarcia-temp/BLUEX_without_images
92
+ split: train
93
+ args:
94
+ num_few_shot: 3
95
+ metrics:
96
+ - type: acc
97
+ value: 23.23
98
+ name: accuracy
99
+ source:
100
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
101
+ name: Open Portuguese LLM Leaderboard
102
+ - task:
103
+ type: text-generation
104
+ name: Text Generation
105
+ dataset:
106
+ name: OAB Exams
107
+ type: eduagarcia/oab_exams
108
+ split: train
109
+ args:
110
+ num_few_shot: 3
111
+ metrics:
112
+ - type: acc
113
+ value: 25.47
114
+ name: accuracy
115
+ source:
116
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
117
+ name: Open Portuguese LLM Leaderboard
118
+ - task:
119
+ type: text-generation
120
+ name: Text Generation
121
+ dataset:
122
+ name: Assin2 RTE
123
+ type: assin2
124
+ split: test
125
+ args:
126
+ num_few_shot: 15
127
+ metrics:
128
+ - type: f1_macro
129
+ value: 56.27
130
+ name: f1-macro
131
+ source:
132
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
133
+ name: Open Portuguese LLM Leaderboard
134
+ - task:
135
+ type: text-generation
136
+ name: Text Generation
137
+ dataset:
138
+ name: Assin2 STS
139
+ type: eduagarcia/portuguese_benchmark
140
+ split: test
141
+ args:
142
+ num_few_shot: 10
143
+ metrics:
144
+ - type: pearson
145
+ value: 1.93
146
+ name: pearson
147
+ source:
148
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
149
+ name: Open Portuguese LLM Leaderboard
150
+ - task:
151
+ type: text-generation
152
+ name: Text Generation
153
+ dataset:
154
+ name: FaQuAD NLI
155
+ type: ruanchaves/faquad-nli
156
+ split: test
157
+ args:
158
+ num_few_shot: 15
159
+ metrics:
160
+ - type: f1_macro
161
+ value: 43.97
162
+ name: f1-macro
163
+ source:
164
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
165
+ name: Open Portuguese LLM Leaderboard
166
+ - task:
167
+ type: text-generation
168
+ name: Text Generation
169
+ dataset:
170
+ name: HateBR Binary
171
+ type: ruanchaves/hatebr
172
+ split: test
173
+ args:
174
+ num_few_shot: 25
175
+ metrics:
176
+ - type: f1_macro
177
+ value: 29.49
178
+ name: f1-macro
179
+ source:
180
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
181
+ name: Open Portuguese LLM Leaderboard
182
+ - task:
183
+ type: text-generation
184
+ name: Text Generation
185
+ dataset:
186
+ name: PT Hate Speech Binary
187
+ type: hate_speech_portuguese
188
+ split: test
189
+ args:
190
+ num_few_shot: 25
191
+ metrics:
192
+ - type: f1_macro
193
+ value: 41.98
194
+ name: f1-macro
195
+ source:
196
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
197
+ name: Open Portuguese LLM Leaderboard
198
+ - task:
199
+ type: text-generation
200
+ name: Text Generation
201
+ dataset:
202
+ name: tweetSentBR
203
+ type: eduagarcia-temp/tweetsentbr
204
+ split: test
205
+ args:
206
+ num_few_shot: 25
207
+ metrics:
208
+ - type: f1_macro
209
+ value: 58.0
210
+ name: f1-macro
211
+ source:
212
+ url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard
213
+ name: Open Portuguese LLM Leaderboard
214
+ - task:
215
+ type: text-generation
216
+ name: Text Generation
217
+ dataset:
218
+ name: ARC-Challenge (PT)
219
+ type: arc_pt
220
+ args:
221
+ num_few_shot: 25
222
+ metrics:
223
+ - type: acc_norm
224
+ value: 30.43
225
+ name: normalized accuracy
226
+ source:
227
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
228
+ name: Evaluation Framework for Multilingual Large Language Models
229
+ - task:
230
+ type: text-generation
231
+ name: Text Generation
232
+ dataset:
233
+ name: HellaSwag (PT)
234
+ type: hellaswag_pt
235
+ args:
236
+ num_few_shot: 10
237
+ metrics:
238
+ - type: acc_norm
239
+ value: 47.17
240
+ name: normalized accuracy
241
+ source:
242
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
243
+ name: Evaluation Framework for Multilingual Large Language Models
244
+ - task:
245
+ type: text-generation
246
+ name: Text Generation
247
+ dataset:
248
+ name: TruthfulQA
249
+ type: truthfulqa_pt
250
+ args:
251
+ num_few_shot: 0
252
+ metrics:
253
+ - type: mc2
254
+ value: 39.3
255
+ name: bleurt
256
+ source:
257
+ url: https://github.com/nlp-uoregon/mlmm-evaluation
258
+ name: Evaluation Framework for Multilingual Large Language Models
259
+ ---
260
+
261
+ # noxinc/Tucano-2b4-Q4_K_M-GGUF
262
+ This model was converted to GGUF format from [`TucanoBR/Tucano-2b4`](https://huggingface.co/TucanoBR/Tucano-2b4) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
263
+ Refer to the [original model card](https://huggingface.co/TucanoBR/Tucano-2b4) for more details on the model.
264
+ ## Use with llama.cpp
265
+
266
+ Install llama.cpp through brew.
267
+
268
+ ```bash
269
+ brew install ggerganov/ggerganov/llama.cpp
270
+ ```
271
+ Invoke the llama.cpp server or the CLI.
272
+
273
+ CLI:
274
+
275
+ ```bash
276
+ llama-cli --hf-repo noxinc/Tucano-2b4-Q4_K_M-GGUF --model tucano-2b4.Q4_K_M.gguf -p "The meaning to life and the universe is"
277
+ ```
278
+
279
+ Server:
280
+
281
+ ```bash
282
+ llama-server --hf-repo noxinc/Tucano-2b4-Q4_K_M-GGUF --model tucano-2b4.Q4_K_M.gguf -c 2048
283
+ ```
284
+
285
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
286
+
287
+ ```
288
+ git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m tucano-2b4.Q4_K_M.gguf -n 128
289
+ ```