flozi00 commited on
Commit
343473b
1 Parent(s): d80e782

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -363
README.md CHANGED
@@ -1,372 +1,20 @@
1
  ---
2
- library_name: peft
 
 
 
3
  ---
4
- ## Training procedure
5
 
 
6
 
7
- The following `bitsandbytes` quantization config was used during training:
8
- - quant_method: bitsandbytes
9
- - load_in_8bit: False
10
- - load_in_4bit: True
11
- - llm_int8_threshold: 6.0
12
- - llm_int8_skip_modules: None
13
- - llm_int8_enable_fp32_cpu_offload: False
14
- - llm_int8_has_fp16_weight: False
15
- - bnb_4bit_quant_type: fp4
16
- - bnb_4bit_use_double_quant: False
17
- - bnb_4bit_compute_dtype: float16
18
 
19
- The following `bitsandbytes` quantization config was used during training:
20
- - quant_method: bitsandbytes
21
- - load_in_8bit: False
22
- - load_in_4bit: True
23
- - llm_int8_threshold: 6.0
24
- - llm_int8_skip_modules: None
25
- - llm_int8_enable_fp32_cpu_offload: False
26
- - llm_int8_has_fp16_weight: False
27
- - bnb_4bit_quant_type: fp4
28
- - bnb_4bit_use_double_quant: False
29
- - bnb_4bit_compute_dtype: float16
30
 
31
- The following `bitsandbytes` quantization config was used during training:
32
- - quant_method: bitsandbytes
33
- - load_in_8bit: False
34
- - load_in_4bit: True
35
- - llm_int8_threshold: 6.0
36
- - llm_int8_skip_modules: None
37
- - llm_int8_enable_fp32_cpu_offload: False
38
- - llm_int8_has_fp16_weight: False
39
- - bnb_4bit_quant_type: fp4
40
- - bnb_4bit_use_double_quant: False
41
- - bnb_4bit_compute_dtype: float16
42
 
43
- The following `bitsandbytes` quantization config was used during training:
44
- - quant_method: bitsandbytes
45
- - load_in_8bit: False
46
- - load_in_4bit: True
47
- - llm_int8_threshold: 6.0
48
- - llm_int8_skip_modules: None
49
- - llm_int8_enable_fp32_cpu_offload: False
50
- - llm_int8_has_fp16_weight: False
51
- - bnb_4bit_quant_type: fp4
52
- - bnb_4bit_use_double_quant: False
53
- - bnb_4bit_compute_dtype: float16
54
 
55
- The following `bitsandbytes` quantization config was used during training:
56
- - quant_method: bitsandbytes
57
- - load_in_8bit: False
58
- - load_in_4bit: True
59
- - llm_int8_threshold: 6.0
60
- - llm_int8_skip_modules: None
61
- - llm_int8_enable_fp32_cpu_offload: False
62
- - llm_int8_has_fp16_weight: False
63
- - bnb_4bit_quant_type: fp4
64
- - bnb_4bit_use_double_quant: False
65
- - bnb_4bit_compute_dtype: float16
66
 
67
- The following `bitsandbytes` quantization config was used during training:
68
- - quant_method: bitsandbytes
69
- - load_in_8bit: False
70
- - load_in_4bit: True
71
- - llm_int8_threshold: 6.0
72
- - llm_int8_skip_modules: None
73
- - llm_int8_enable_fp32_cpu_offload: False
74
- - llm_int8_has_fp16_weight: False
75
- - bnb_4bit_quant_type: fp4
76
- - bnb_4bit_use_double_quant: False
77
- - bnb_4bit_compute_dtype: float16
78
-
79
- The following `bitsandbytes` quantization config was used during training:
80
- - quant_method: bitsandbytes
81
- - load_in_8bit: False
82
- - load_in_4bit: True
83
- - llm_int8_threshold: 6.0
84
- - llm_int8_skip_modules: None
85
- - llm_int8_enable_fp32_cpu_offload: False
86
- - llm_int8_has_fp16_weight: False
87
- - bnb_4bit_quant_type: fp4
88
- - bnb_4bit_use_double_quant: False
89
- - bnb_4bit_compute_dtype: float16
90
-
91
- The following `bitsandbytes` quantization config was used during training:
92
- - quant_method: bitsandbytes
93
- - load_in_8bit: False
94
- - load_in_4bit: True
95
- - llm_int8_threshold: 6.0
96
- - llm_int8_skip_modules: None
97
- - llm_int8_enable_fp32_cpu_offload: False
98
- - llm_int8_has_fp16_weight: False
99
- - bnb_4bit_quant_type: fp4
100
- - bnb_4bit_use_double_quant: False
101
- - bnb_4bit_compute_dtype: float16
102
-
103
- The following `bitsandbytes` quantization config was used during training:
104
- - quant_method: bitsandbytes
105
- - load_in_8bit: False
106
- - load_in_4bit: True
107
- - llm_int8_threshold: 6.0
108
- - llm_int8_skip_modules: None
109
- - llm_int8_enable_fp32_cpu_offload: False
110
- - llm_int8_has_fp16_weight: False
111
- - bnb_4bit_quant_type: fp4
112
- - bnb_4bit_use_double_quant: False
113
- - bnb_4bit_compute_dtype: float16
114
-
115
- The following `bitsandbytes` quantization config was used during training:
116
- - quant_method: bitsandbytes
117
- - load_in_8bit: False
118
- - load_in_4bit: True
119
- - llm_int8_threshold: 6.0
120
- - llm_int8_skip_modules: None
121
- - llm_int8_enable_fp32_cpu_offload: False
122
- - llm_int8_has_fp16_weight: False
123
- - bnb_4bit_quant_type: fp4
124
- - bnb_4bit_use_double_quant: False
125
- - bnb_4bit_compute_dtype: float16
126
-
127
- The following `bitsandbytes` quantization config was used during training:
128
- - quant_method: bitsandbytes
129
- - load_in_8bit: False
130
- - load_in_4bit: True
131
- - llm_int8_threshold: 6.0
132
- - llm_int8_skip_modules: None
133
- - llm_int8_enable_fp32_cpu_offload: False
134
- - llm_int8_has_fp16_weight: False
135
- - bnb_4bit_quant_type: fp4
136
- - bnb_4bit_use_double_quant: False
137
- - bnb_4bit_compute_dtype: float16
138
-
139
- The following `bitsandbytes` quantization config was used during training:
140
- - quant_method: bitsandbytes
141
- - load_in_8bit: False
142
- - load_in_4bit: True
143
- - llm_int8_threshold: 6.0
144
- - llm_int8_skip_modules: None
145
- - llm_int8_enable_fp32_cpu_offload: False
146
- - llm_int8_has_fp16_weight: False
147
- - bnb_4bit_quant_type: fp4
148
- - bnb_4bit_use_double_quant: False
149
- - bnb_4bit_compute_dtype: float16
150
-
151
- The following `bitsandbytes` quantization config was used during training:
152
- - quant_method: bitsandbytes
153
- - load_in_8bit: False
154
- - load_in_4bit: True
155
- - llm_int8_threshold: 6.0
156
- - llm_int8_skip_modules: None
157
- - llm_int8_enable_fp32_cpu_offload: False
158
- - llm_int8_has_fp16_weight: False
159
- - bnb_4bit_quant_type: fp4
160
- - bnb_4bit_use_double_quant: False
161
- - bnb_4bit_compute_dtype: float16
162
-
163
- The following `bitsandbytes` quantization config was used during training:
164
- - quant_method: bitsandbytes
165
- - load_in_8bit: False
166
- - load_in_4bit: True
167
- - llm_int8_threshold: 6.0
168
- - llm_int8_skip_modules: None
169
- - llm_int8_enable_fp32_cpu_offload: False
170
- - llm_int8_has_fp16_weight: False
171
- - bnb_4bit_quant_type: fp4
172
- - bnb_4bit_use_double_quant: False
173
- - bnb_4bit_compute_dtype: float16
174
-
175
- The following `bitsandbytes` quantization config was used during training:
176
- - quant_method: bitsandbytes
177
- - load_in_8bit: False
178
- - load_in_4bit: True
179
- - llm_int8_threshold: 6.0
180
- - llm_int8_skip_modules: None
181
- - llm_int8_enable_fp32_cpu_offload: False
182
- - llm_int8_has_fp16_weight: False
183
- - bnb_4bit_quant_type: fp4
184
- - bnb_4bit_use_double_quant: False
185
- - bnb_4bit_compute_dtype: float16
186
-
187
- The following `bitsandbytes` quantization config was used during training:
188
- - quant_method: bitsandbytes
189
- - load_in_8bit: False
190
- - load_in_4bit: True
191
- - llm_int8_threshold: 6.0
192
- - llm_int8_skip_modules: None
193
- - llm_int8_enable_fp32_cpu_offload: False
194
- - llm_int8_has_fp16_weight: False
195
- - bnb_4bit_quant_type: fp4
196
- - bnb_4bit_use_double_quant: False
197
- - bnb_4bit_compute_dtype: float16
198
-
199
- The following `bitsandbytes` quantization config was used during training:
200
- - quant_method: bitsandbytes
201
- - load_in_8bit: False
202
- - load_in_4bit: True
203
- - llm_int8_threshold: 6.0
204
- - llm_int8_skip_modules: None
205
- - llm_int8_enable_fp32_cpu_offload: False
206
- - llm_int8_has_fp16_weight: False
207
- - bnb_4bit_quant_type: fp4
208
- - bnb_4bit_use_double_quant: False
209
- - bnb_4bit_compute_dtype: float16
210
-
211
- The following `bitsandbytes` quantization config was used during training:
212
- - quant_method: bitsandbytes
213
- - load_in_8bit: False
214
- - load_in_4bit: True
215
- - llm_int8_threshold: 6.0
216
- - llm_int8_skip_modules: None
217
- - llm_int8_enable_fp32_cpu_offload: False
218
- - llm_int8_has_fp16_weight: False
219
- - bnb_4bit_quant_type: fp4
220
- - bnb_4bit_use_double_quant: False
221
- - bnb_4bit_compute_dtype: float16
222
-
223
- The following `bitsandbytes` quantization config was used during training:
224
- - quant_method: bitsandbytes
225
- - load_in_8bit: False
226
- - load_in_4bit: True
227
- - llm_int8_threshold: 6.0
228
- - llm_int8_skip_modules: None
229
- - llm_int8_enable_fp32_cpu_offload: False
230
- - llm_int8_has_fp16_weight: False
231
- - bnb_4bit_quant_type: fp4
232
- - bnb_4bit_use_double_quant: False
233
- - bnb_4bit_compute_dtype: float16
234
-
235
- The following `bitsandbytes` quantization config was used during training:
236
- - quant_method: bitsandbytes
237
- - load_in_8bit: False
238
- - load_in_4bit: True
239
- - llm_int8_threshold: 6.0
240
- - llm_int8_skip_modules: None
241
- - llm_int8_enable_fp32_cpu_offload: False
242
- - llm_int8_has_fp16_weight: False
243
- - bnb_4bit_quant_type: fp4
244
- - bnb_4bit_use_double_quant: False
245
- - bnb_4bit_compute_dtype: float16
246
-
247
- The following `bitsandbytes` quantization config was used during training:
248
- - quant_method: bitsandbytes
249
- - load_in_8bit: False
250
- - load_in_4bit: True
251
- - llm_int8_threshold: 6.0
252
- - llm_int8_skip_modules: None
253
- - llm_int8_enable_fp32_cpu_offload: False
254
- - llm_int8_has_fp16_weight: False
255
- - bnb_4bit_quant_type: fp4
256
- - bnb_4bit_use_double_quant: False
257
- - bnb_4bit_compute_dtype: float16
258
-
259
- The following `bitsandbytes` quantization config was used during training:
260
- - quant_method: bitsandbytes
261
- - load_in_8bit: False
262
- - load_in_4bit: True
263
- - llm_int8_threshold: 6.0
264
- - llm_int8_skip_modules: None
265
- - llm_int8_enable_fp32_cpu_offload: False
266
- - llm_int8_has_fp16_weight: False
267
- - bnb_4bit_quant_type: fp4
268
- - bnb_4bit_use_double_quant: False
269
- - bnb_4bit_compute_dtype: float16
270
-
271
- The following `bitsandbytes` quantization config was used during training:
272
- - quant_method: bitsandbytes
273
- - load_in_8bit: False
274
- - load_in_4bit: True
275
- - llm_int8_threshold: 6.0
276
- - llm_int8_skip_modules: None
277
- - llm_int8_enable_fp32_cpu_offload: False
278
- - llm_int8_has_fp16_weight: False
279
- - bnb_4bit_quant_type: fp4
280
- - bnb_4bit_use_double_quant: False
281
- - bnb_4bit_compute_dtype: float16
282
-
283
- The following `bitsandbytes` quantization config was used during training:
284
- - quant_method: bitsandbytes
285
- - load_in_8bit: False
286
- - load_in_4bit: True
287
- - llm_int8_threshold: 6.0
288
- - llm_int8_skip_modules: None
289
- - llm_int8_enable_fp32_cpu_offload: False
290
- - llm_int8_has_fp16_weight: False
291
- - bnb_4bit_quant_type: fp4
292
- - bnb_4bit_use_double_quant: False
293
- - bnb_4bit_compute_dtype: float16
294
-
295
- The following `bitsandbytes` quantization config was used during training:
296
- - quant_method: bitsandbytes
297
- - load_in_8bit: False
298
- - load_in_4bit: True
299
- - llm_int8_threshold: 6.0
300
- - llm_int8_skip_modules: None
301
- - llm_int8_enable_fp32_cpu_offload: False
302
- - llm_int8_has_fp16_weight: False
303
- - bnb_4bit_quant_type: fp4
304
- - bnb_4bit_use_double_quant: False
305
- - bnb_4bit_compute_dtype: float16
306
-
307
- The following `bitsandbytes` quantization config was used during training:
308
- - quant_method: bitsandbytes
309
- - load_in_8bit: False
310
- - load_in_4bit: True
311
- - llm_int8_threshold: 6.0
312
- - llm_int8_skip_modules: None
313
- - llm_int8_enable_fp32_cpu_offload: False
314
- - llm_int8_has_fp16_weight: False
315
- - bnb_4bit_quant_type: fp4
316
- - bnb_4bit_use_double_quant: False
317
- - bnb_4bit_compute_dtype: float16
318
-
319
- The following `bitsandbytes` quantization config was used during training:
320
- - quant_method: bitsandbytes
321
- - load_in_8bit: False
322
- - load_in_4bit: True
323
- - llm_int8_threshold: 6.0
324
- - llm_int8_skip_modules: None
325
- - llm_int8_enable_fp32_cpu_offload: False
326
- - llm_int8_has_fp16_weight: False
327
- - bnb_4bit_quant_type: fp4
328
- - bnb_4bit_use_double_quant: False
329
- - bnb_4bit_compute_dtype: float16
330
-
331
- The following `bitsandbytes` quantization config was used during training:
332
- - quant_method: bitsandbytes
333
- - load_in_8bit: False
334
- - load_in_4bit: True
335
- - llm_int8_threshold: 6.0
336
- - llm_int8_skip_modules: None
337
- - llm_int8_enable_fp32_cpu_offload: False
338
- - llm_int8_has_fp16_weight: False
339
- - bnb_4bit_quant_type: fp4
340
- - bnb_4bit_use_double_quant: False
341
- - bnb_4bit_compute_dtype: float16
342
- ### Framework versions
343
-
344
- - PEFT 0.6.0.dev0
345
- - PEFT 0.6.0.dev0
346
- - PEFT 0.6.0.dev0
347
- - PEFT 0.6.0.dev0
348
- - PEFT 0.6.0.dev0
349
- - PEFT 0.6.0.dev0
350
- - PEFT 0.6.0.dev0
351
- - PEFT 0.6.0.dev0
352
- - PEFT 0.6.0.dev0
353
- - PEFT 0.6.0.dev0
354
- - PEFT 0.6.0.dev0
355
- - PEFT 0.6.0.dev0
356
- - PEFT 0.6.0.dev0
357
- - PEFT 0.6.0.dev0
358
- - PEFT 0.6.0.dev0
359
- - PEFT 0.6.0.dev0
360
- - PEFT 0.6.0.dev0
361
- - PEFT 0.6.0.dev0
362
- - PEFT 0.6.0.dev0
363
- - PEFT 0.6.0.dev0
364
- - PEFT 0.6.0.dev0
365
- - PEFT 0.6.0.dev0
366
- - PEFT 0.6.0.dev0
367
- - PEFT 0.6.0.dev0
368
- - PEFT 0.6.0.dev0
369
- - PEFT 0.6.0.dev0
370
- - PEFT 0.6.0.dev0
371
-
372
- - PEFT 0.6.0.dev0
 
1
  ---
2
+ datasets:
3
+ - flozi00/conversations
4
+ language:
5
+ - de
6
  ---
 
7
 
8
+ ## This project is sponsored by [ ![PrimeLine](https://www.primeline-solutions.com/skin/frontend/default/theme566/images/primeline-solutions-logo.png) ](https://www.primeline-solutions.com/de/server/nach-einsatzzweck/gpu-rendering-hpc/)
9
 
10
+ # Model Card
 
 
 
 
 
 
 
 
 
 
11
 
12
+ This model is an finetuned version for german instructions and conversations in style of Alpaca. "### Assistant:" "### User:"
13
+ The dataset used is deduplicated and cleaned, with no codes inside. The focus is on instruction following and conversational tasks.
 
 
 
 
 
 
 
 
 
14
 
15
+ The model archictecture is based on Llama version 2 with 13B parameters, trained on 100% renewable energy powered hardware.
 
 
 
 
 
 
 
 
 
 
16
 
17
+ This work is contributed by private research of [flozi00](https://huggingface.co/flozi00)
 
 
 
 
 
 
 
 
 
 
18
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
+ Join discussions about german llm research, and plan larger training runs together: https://join.slack.com/t/slack-dtc7771/shared_invite/zt-219keplqu-hLwjm0xcFAOX7enERfBz0Q