flozi00 commited on
Commit
5a5308c
1 Parent(s): 721d621

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -320
README.md DELETED
@@ -1,320 +0,0 @@
1
- ---
2
- library_name: peft
3
- ---
4
- ## Training procedure
5
-
6
-
7
- The following `bitsandbytes` quantization config was used during training:
8
- - load_in_8bit: False
9
- - load_in_4bit: True
10
- - llm_int8_threshold: 6.0
11
- - llm_int8_skip_modules: None
12
- - llm_int8_enable_fp32_cpu_offload: False
13
- - llm_int8_has_fp16_weight: False
14
- - bnb_4bit_quant_type: fp4
15
- - bnb_4bit_use_double_quant: True
16
- - bnb_4bit_compute_dtype: float16
17
-
18
- The following `bitsandbytes` quantization config was used during training:
19
- - load_in_8bit: False
20
- - load_in_4bit: True
21
- - llm_int8_threshold: 6.0
22
- - llm_int8_skip_modules: None
23
- - llm_int8_enable_fp32_cpu_offload: False
24
- - llm_int8_has_fp16_weight: False
25
- - bnb_4bit_quant_type: fp4
26
- - bnb_4bit_use_double_quant: True
27
- - bnb_4bit_compute_dtype: float16
28
-
29
- The following `bitsandbytes` quantization config was used during training:
30
- - load_in_8bit: False
31
- - load_in_4bit: True
32
- - llm_int8_threshold: 6.0
33
- - llm_int8_skip_modules: None
34
- - llm_int8_enable_fp32_cpu_offload: False
35
- - llm_int8_has_fp16_weight: False
36
- - bnb_4bit_quant_type: fp4
37
- - bnb_4bit_use_double_quant: True
38
- - bnb_4bit_compute_dtype: float16
39
-
40
- The following `bitsandbytes` quantization config was used during training:
41
- - load_in_8bit: False
42
- - load_in_4bit: True
43
- - llm_int8_threshold: 6.0
44
- - llm_int8_skip_modules: None
45
- - llm_int8_enable_fp32_cpu_offload: False
46
- - llm_int8_has_fp16_weight: False
47
- - bnb_4bit_quant_type: fp4
48
- - bnb_4bit_use_double_quant: True
49
- - bnb_4bit_compute_dtype: float16
50
-
51
- The following `bitsandbytes` quantization config was used during training:
52
- - load_in_8bit: False
53
- - load_in_4bit: True
54
- - llm_int8_threshold: 6.0
55
- - llm_int8_skip_modules: None
56
- - llm_int8_enable_fp32_cpu_offload: False
57
- - llm_int8_has_fp16_weight: False
58
- - bnb_4bit_quant_type: fp4
59
- - bnb_4bit_use_double_quant: True
60
- - bnb_4bit_compute_dtype: float16
61
-
62
- The following `bitsandbytes` quantization config was used during training:
63
- - load_in_8bit: False
64
- - load_in_4bit: True
65
- - llm_int8_threshold: 6.0
66
- - llm_int8_skip_modules: None
67
- - llm_int8_enable_fp32_cpu_offload: False
68
- - llm_int8_has_fp16_weight: False
69
- - bnb_4bit_quant_type: fp4
70
- - bnb_4bit_use_double_quant: True
71
- - bnb_4bit_compute_dtype: float16
72
-
73
- The following `bitsandbytes` quantization config was used during training:
74
- - load_in_8bit: False
75
- - load_in_4bit: True
76
- - llm_int8_threshold: 6.0
77
- - llm_int8_skip_modules: None
78
- - llm_int8_enable_fp32_cpu_offload: False
79
- - llm_int8_has_fp16_weight: False
80
- - bnb_4bit_quant_type: fp4
81
- - bnb_4bit_use_double_quant: True
82
- - bnb_4bit_compute_dtype: float16
83
-
84
- The following `bitsandbytes` quantization config was used during training:
85
- - load_in_8bit: False
86
- - load_in_4bit: True
87
- - llm_int8_threshold: 6.0
88
- - llm_int8_skip_modules: None
89
- - llm_int8_enable_fp32_cpu_offload: False
90
- - llm_int8_has_fp16_weight: False
91
- - bnb_4bit_quant_type: fp4
92
- - bnb_4bit_use_double_quant: True
93
- - bnb_4bit_compute_dtype: float16
94
-
95
- The following `bitsandbytes` quantization config was used during training:
96
- - load_in_8bit: False
97
- - load_in_4bit: True
98
- - llm_int8_threshold: 6.0
99
- - llm_int8_skip_modules: None
100
- - llm_int8_enable_fp32_cpu_offload: False
101
- - llm_int8_has_fp16_weight: False
102
- - bnb_4bit_quant_type: fp4
103
- - bnb_4bit_use_double_quant: True
104
- - bnb_4bit_compute_dtype: float16
105
-
106
- The following `bitsandbytes` quantization config was used during training:
107
- - load_in_8bit: False
108
- - load_in_4bit: True
109
- - llm_int8_threshold: 6.0
110
- - llm_int8_skip_modules: None
111
- - llm_int8_enable_fp32_cpu_offload: False
112
- - llm_int8_has_fp16_weight: False
113
- - bnb_4bit_quant_type: fp4
114
- - bnb_4bit_use_double_quant: True
115
- - bnb_4bit_compute_dtype: float16
116
-
117
- The following `bitsandbytes` quantization config was used during training:
118
- - load_in_8bit: False
119
- - load_in_4bit: True
120
- - llm_int8_threshold: 6.0
121
- - llm_int8_skip_modules: None
122
- - llm_int8_enable_fp32_cpu_offload: False
123
- - llm_int8_has_fp16_weight: False
124
- - bnb_4bit_quant_type: fp4
125
- - bnb_4bit_use_double_quant: True
126
- - bnb_4bit_compute_dtype: float16
127
-
128
- The following `bitsandbytes` quantization config was used during training:
129
- - load_in_8bit: False
130
- - load_in_4bit: True
131
- - llm_int8_threshold: 6.0
132
- - llm_int8_skip_modules: None
133
- - llm_int8_enable_fp32_cpu_offload: False
134
- - llm_int8_has_fp16_weight: False
135
- - bnb_4bit_quant_type: fp4
136
- - bnb_4bit_use_double_quant: True
137
- - bnb_4bit_compute_dtype: float16
138
-
139
- The following `bitsandbytes` quantization config was used during training:
140
- - load_in_8bit: False
141
- - load_in_4bit: True
142
- - llm_int8_threshold: 6.0
143
- - llm_int8_skip_modules: None
144
- - llm_int8_enable_fp32_cpu_offload: False
145
- - llm_int8_has_fp16_weight: False
146
- - bnb_4bit_quant_type: fp4
147
- - bnb_4bit_use_double_quant: True
148
- - bnb_4bit_compute_dtype: float16
149
-
150
- The following `bitsandbytes` quantization config was used during training:
151
- - load_in_8bit: False
152
- - load_in_4bit: True
153
- - llm_int8_threshold: 6.0
154
- - llm_int8_skip_modules: None
155
- - llm_int8_enable_fp32_cpu_offload: False
156
- - llm_int8_has_fp16_weight: False
157
- - bnb_4bit_quant_type: fp4
158
- - bnb_4bit_use_double_quant: True
159
- - bnb_4bit_compute_dtype: float16
160
-
161
- The following `bitsandbytes` quantization config was used during training:
162
- - load_in_8bit: False
163
- - load_in_4bit: True
164
- - llm_int8_threshold: 6.0
165
- - llm_int8_skip_modules: None
166
- - llm_int8_enable_fp32_cpu_offload: False
167
- - llm_int8_has_fp16_weight: False
168
- - bnb_4bit_quant_type: fp4
169
- - bnb_4bit_use_double_quant: True
170
- - bnb_4bit_compute_dtype: float16
171
-
172
- The following `bitsandbytes` quantization config was used during training:
173
- - load_in_8bit: False
174
- - load_in_4bit: True
175
- - llm_int8_threshold: 6.0
176
- - llm_int8_skip_modules: None
177
- - llm_int8_enable_fp32_cpu_offload: False
178
- - llm_int8_has_fp16_weight: False
179
- - bnb_4bit_quant_type: fp4
180
- - bnb_4bit_use_double_quant: True
181
- - bnb_4bit_compute_dtype: float16
182
-
183
- The following `bitsandbytes` quantization config was used during training:
184
- - load_in_8bit: False
185
- - load_in_4bit: True
186
- - llm_int8_threshold: 6.0
187
- - llm_int8_skip_modules: None
188
- - llm_int8_enable_fp32_cpu_offload: False
189
- - llm_int8_has_fp16_weight: False
190
- - bnb_4bit_quant_type: fp4
191
- - bnb_4bit_use_double_quant: True
192
- - bnb_4bit_compute_dtype: float16
193
-
194
- The following `bitsandbytes` quantization config was used during training:
195
- - load_in_8bit: False
196
- - load_in_4bit: True
197
- - llm_int8_threshold: 6.0
198
- - llm_int8_skip_modules: None
199
- - llm_int8_enable_fp32_cpu_offload: False
200
- - llm_int8_has_fp16_weight: False
201
- - bnb_4bit_quant_type: fp4
202
- - bnb_4bit_use_double_quant: True
203
- - bnb_4bit_compute_dtype: float16
204
-
205
- The following `bitsandbytes` quantization config was used during training:
206
- - load_in_8bit: False
207
- - load_in_4bit: True
208
- - llm_int8_threshold: 6.0
209
- - llm_int8_skip_modules: None
210
- - llm_int8_enable_fp32_cpu_offload: False
211
- - llm_int8_has_fp16_weight: False
212
- - bnb_4bit_quant_type: fp4
213
- - bnb_4bit_use_double_quant: True
214
- - bnb_4bit_compute_dtype: float16
215
-
216
- The following `bitsandbytes` quantization config was used during training:
217
- - load_in_8bit: False
218
- - load_in_4bit: True
219
- - llm_int8_threshold: 6.0
220
- - llm_int8_skip_modules: None
221
- - llm_int8_enable_fp32_cpu_offload: False
222
- - llm_int8_has_fp16_weight: False
223
- - bnb_4bit_quant_type: fp4
224
- - bnb_4bit_use_double_quant: True
225
- - bnb_4bit_compute_dtype: float16
226
-
227
- The following `bitsandbytes` quantization config was used during training:
228
- - load_in_8bit: False
229
- - load_in_4bit: True
230
- - llm_int8_threshold: 6.0
231
- - llm_int8_skip_modules: None
232
- - llm_int8_enable_fp32_cpu_offload: False
233
- - llm_int8_has_fp16_weight: False
234
- - bnb_4bit_quant_type: fp4
235
- - bnb_4bit_use_double_quant: True
236
- - bnb_4bit_compute_dtype: float16
237
-
238
- The following `bitsandbytes` quantization config was used during training:
239
- - load_in_8bit: False
240
- - load_in_4bit: True
241
- - llm_int8_threshold: 6.0
242
- - llm_int8_skip_modules: None
243
- - llm_int8_enable_fp32_cpu_offload: False
244
- - llm_int8_has_fp16_weight: False
245
- - bnb_4bit_quant_type: fp4
246
- - bnb_4bit_use_double_quant: True
247
- - bnb_4bit_compute_dtype: float16
248
-
249
- The following `bitsandbytes` quantization config was used during training:
250
- - load_in_8bit: False
251
- - load_in_4bit: True
252
- - llm_int8_threshold: 6.0
253
- - llm_int8_skip_modules: None
254
- - llm_int8_enable_fp32_cpu_offload: False
255
- - llm_int8_has_fp16_weight: False
256
- - bnb_4bit_quant_type: fp4
257
- - bnb_4bit_use_double_quant: True
258
- - bnb_4bit_compute_dtype: float16
259
-
260
- The following `bitsandbytes` quantization config was used during training:
261
- - load_in_8bit: False
262
- - load_in_4bit: True
263
- - llm_int8_threshold: 6.0
264
- - llm_int8_skip_modules: None
265
- - llm_int8_enable_fp32_cpu_offload: False
266
- - llm_int8_has_fp16_weight: False
267
- - bnb_4bit_quant_type: fp4
268
- - bnb_4bit_use_double_quant: True
269
- - bnb_4bit_compute_dtype: float16
270
-
271
- The following `bitsandbytes` quantization config was used during training:
272
- - load_in_8bit: False
273
- - load_in_4bit: True
274
- - llm_int8_threshold: 6.0
275
- - llm_int8_skip_modules: None
276
- - llm_int8_enable_fp32_cpu_offload: False
277
- - llm_int8_has_fp16_weight: False
278
- - bnb_4bit_quant_type: fp4
279
- - bnb_4bit_use_double_quant: True
280
- - bnb_4bit_compute_dtype: float16
281
-
282
- The following `bitsandbytes` quantization config was used during training:
283
- - load_in_8bit: False
284
- - load_in_4bit: True
285
- - llm_int8_threshold: 6.0
286
- - llm_int8_skip_modules: None
287
- - llm_int8_enable_fp32_cpu_offload: False
288
- - llm_int8_has_fp16_weight: False
289
- - bnb_4bit_quant_type: fp4
290
- - bnb_4bit_use_double_quant: True
291
- - bnb_4bit_compute_dtype: float16
292
- ### Framework versions
293
-
294
- - PEFT 0.4.0.dev0
295
- - PEFT 0.4.0.dev0
296
- - PEFT 0.4.0.dev0
297
- - PEFT 0.4.0.dev0
298
- - PEFT 0.4.0.dev0
299
- - PEFT 0.4.0.dev0
300
- - PEFT 0.4.0.dev0
301
- - PEFT 0.4.0.dev0
302
- - PEFT 0.4.0.dev0
303
- - PEFT 0.4.0.dev0
304
- - PEFT 0.4.0.dev0
305
- - PEFT 0.4.0.dev0
306
- - PEFT 0.4.0.dev0
307
- - PEFT 0.4.0.dev0
308
- - PEFT 0.4.0.dev0
309
- - PEFT 0.4.0.dev0
310
- - PEFT 0.4.0.dev0
311
- - PEFT 0.4.0.dev0
312
- - PEFT 0.4.0.dev0
313
- - PEFT 0.4.0.dev0
314
- - PEFT 0.4.0.dev0
315
- - PEFT 0.4.0.dev0
316
- - PEFT 0.4.0.dev0
317
- - PEFT 0.4.0.dev0
318
- - PEFT 0.4.0.dev0
319
-
320
- - PEFT 0.4.0.dev0