m4lw4r3exe commited on
Commit
371f79e
·
1 Parent(s): fbba9d2

Upload manual_upload with huggingface_hub

Browse files
manual_upload/.gitattributes ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tflite filter=lfs diff=lfs merge=lfs -text
29
+ *.tgz filter=lfs diff=lfs merge=lfs -text
30
+ *.wasm filter=lfs diff=lfs merge=lfs -text
31
+ *.xz filter=lfs diff=lfs merge=lfs -text
32
+ *.zip filter=lfs diff=lfs merge=lfs -text
33
+ *.zst filter=lfs diff=lfs merge=lfs -text
34
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
manual_upload/.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
manual_upload/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "layer_norm_epsilon": 1e-05,
12
+ "model_type": "gpt2",
13
+ "n_embd": 512,
14
+ "n_head": 8,
15
+ "n_inner": null,
16
+ "n_layer": 6,
17
+ "n_positions": 2048,
18
+ "pad_token_id": 1,
19
+ "reorder_and_upcast_attn": false,
20
+ "resid_pdrop": 0.1,
21
+ "scale_attn_by_inverse_layer_idx": false,
22
+ "scale_attn_weights": true,
23
+ "summary_activation": null,
24
+ "summary_first_dropout": 0.1,
25
+ "summary_proj_to_labels": true,
26
+ "summary_type": "cls_index",
27
+ "summary_use_proj": true,
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.26.0.dev0",
30
+ "use_cache": true,
31
+ "vocab_size": 301
32
+ }
manual_upload/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54af5d8875e2c2fd3cc37c56d33cad185fa27c7098ef23bdcb9ec77ecf847f0e
3
+ size 105666297
manual_upload/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "pad_token": "[PAD]"
3
+ }
manual_upload/tokenizer.json ADDED
@@ -0,0 +1,347 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "[UNK]",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "[PAD]",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 2,
26
+ "content": "[MASK]",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ }
33
+ ],
34
+ "normalizer": null,
35
+ "pre_tokenizer": {
36
+ "type": "WhitespaceSplit"
37
+ },
38
+ "post_processor": null,
39
+ "decoder": null,
40
+ "model": {
41
+ "type": "WordLevel",
42
+ "vocab": {
43
+ "[UNK]": 0,
44
+ "[PAD]": 1,
45
+ "[MASK]": 2,
46
+ "TIME_DELTA=1": 3,
47
+ "TIME_DELTA=2": 4,
48
+ "BAR_END": 5,
49
+ "BAR_START": 6,
50
+ "NOTE_OFF=42": 7,
51
+ "NOTE_ON=42": 8,
52
+ "NOTE_OFF=36": 9,
53
+ "NOTE_ON=36": 10,
54
+ "TIME_DELTA=4": 11,
55
+ "NOTE_OFF=38": 12,
56
+ "NOTE_ON=38": 13,
57
+ "NOTE_OFF=40": 14,
58
+ "NOTE_ON=40": 15,
59
+ "NOTE_OFF=35": 16,
60
+ "NOTE_ON=35": 17,
61
+ "TIME_DELTA=3": 18,
62
+ "NOTE_OFF=64": 19,
63
+ "NOTE_ON=64": 20,
64
+ "NOTE_OFF=57": 21,
65
+ "NOTE_ON=57": 22,
66
+ "NOTE_OFF=62": 23,
67
+ "NOTE_ON=62": 24,
68
+ "NOTE_OFF=54": 25,
69
+ "NOTE_ON=54": 26,
70
+ "NOTE_OFF=59": 27,
71
+ "NOTE_ON=59": 28,
72
+ "NOTE_OFF=55": 29,
73
+ "NOTE_ON=55": 30,
74
+ "NOTE_OFF=69": 31,
75
+ "NOTE_ON=69": 32,
76
+ "NOTE_OFF=60": 33,
77
+ "NOTE_ON=60": 34,
78
+ "NOTE_OFF=46": 35,
79
+ "NOTE_ON=46": 36,
80
+ "NOTE_OFF=44": 37,
81
+ "NOTE_ON=44": 38,
82
+ "NOTE_OFF=50": 39,
83
+ "NOTE_ON=50": 40,
84
+ "NOTE_OFF=67": 41,
85
+ "NOTE_ON=67": 42,
86
+ "NOTE_OFF=45": 43,
87
+ "NOTE_ON=45": 44,
88
+ "NOTE_OFF=52": 45,
89
+ "NOTE_ON=52": 46,
90
+ "NOTE_OFF=70": 47,
91
+ "NOTE_ON=70": 48,
92
+ "NOTE_OFF=51": 49,
93
+ "NOTE_ON=51": 50,
94
+ "TRACK_END": 51,
95
+ "TRACK_START": 52,
96
+ "NOTE_OFF=61": 53,
97
+ "NOTE_ON=61": 54,
98
+ "NOTE_OFF=66": 55,
99
+ "NOTE_ON=66": 56,
100
+ "NOTE_OFF=48": 57,
101
+ "NOTE_ON=48": 58,
102
+ "NOTE_OFF=65": 59,
103
+ "NOTE_ON=65": 60,
104
+ "NOTE_OFF=43": 61,
105
+ "NOTE_ON=43": 62,
106
+ "NOTE_OFF=63": 63,
107
+ "NOTE_ON=63": 64,
108
+ "NOTE_OFF=53": 65,
109
+ "NOTE_ON=53": 66,
110
+ "NOTE_OFF=47": 67,
111
+ "NOTE_ON=47": 68,
112
+ "NOTE_OFF=49": 69,
113
+ "NOTE_ON=49": 70,
114
+ "NOTE_OFF=58": 71,
115
+ "NOTE_ON=58": 72,
116
+ "NOTE_OFF=56": 73,
117
+ "NOTE_ON=56": 74,
118
+ "NOTE_OFF=39": 75,
119
+ "NOTE_ON=39": 76,
120
+ "NOTE_OFF=41": 77,
121
+ "NOTE_ON=41": 78,
122
+ "NOTE_OFF=71": 79,
123
+ "NOTE_ON=71": 80,
124
+ "TIME_DELTA=6": 81,
125
+ "NOTE_OFF=72": 82,
126
+ "NOTE_ON=72": 83,
127
+ "NOTE_OFF=33": 84,
128
+ "NOTE_ON=33": 85,
129
+ "NOTE_OFF=68": 86,
130
+ "NOTE_ON=68": 87,
131
+ "NOTE_OFF=74": 88,
132
+ "NOTE_ON=74": 89,
133
+ "TIME_DELTA=8": 90,
134
+ "TIME_DELTA=16": 91,
135
+ "NOTE_OFF=76": 92,
136
+ "NOTE_ON=76": 93,
137
+ "NOTE_OFF=82": 94,
138
+ "NOTE_ON=82": 95,
139
+ "NOTE_OFF=37": 96,
140
+ "NOTE_ON=37": 97,
141
+ "NOTE_OFF=31": 98,
142
+ "NOTE_ON=31": 99,
143
+ "NOTE_OFF=73": 100,
144
+ "NOTE_ON=73": 101,
145
+ "NOTE_OFF=28": 102,
146
+ "NOTE_ON=28": 103,
147
+ "NOTE_OFF=34": 104,
148
+ "NOTE_ON=34": 105,
149
+ "NOTE_OFF=75": 106,
150
+ "NOTE_ON=75": 107,
151
+ "TIME_DELTA=5": 108,
152
+ "NOTE_OFF=29": 109,
153
+ "NOTE_ON=29": 110,
154
+ "NOTE_OFF=32": 111,
155
+ "NOTE_ON=32": 112,
156
+ "NOTE_OFF=79": 113,
157
+ "NOTE_ON=79": 114,
158
+ "DENSITY=3": 115,
159
+ "NOTE_OFF=81": 116,
160
+ "NOTE_ON=81": 117,
161
+ "NOTE_OFF=77": 118,
162
+ "NOTE_ON=77": 119,
163
+ "NOTE_OFF=78": 120,
164
+ "NOTE_ON=78": 121,
165
+ "INST=3": 122,
166
+ "NOTE_OFF=30": 123,
167
+ "NOTE_ON=30": 124,
168
+ "DENSITY=2": 125,
169
+ "DENSITY=1": 126,
170
+ "TIME_DELTA=7": 127,
171
+ "DENSITY=0": 128,
172
+ "NOTE_OFF=26": 129,
173
+ "NOTE_ON=26": 130,
174
+ "INST=DRUMS": 131,
175
+ "NOTE_OFF=80": 132,
176
+ "NOTE_ON=80": 133,
177
+ "TIME_DELTA=12": 134,
178
+ "NOTE_OFF=27": 135,
179
+ "NOTE_ON=27": 136,
180
+ "PIECE_START": 137,
181
+ "TIME_DELTA=10": 138,
182
+ "NOTE_OFF=83": 139,
183
+ "NOTE_ON=83": 140,
184
+ "INST=4": 141,
185
+ "NOTE_OFF=84": 142,
186
+ "NOTE_ON=84": 143,
187
+ "NOTE_OFF=86": 144,
188
+ "NOTE_ON=86": 145,
189
+ "TIME_DELTA=14": 146,
190
+ "INST=0": 147,
191
+ "TIME_DELTA=15": 148,
192
+ "INST=6": 149,
193
+ "NOTE_OFF=85": 150,
194
+ "NOTE_ON=85": 151,
195
+ "NOTE_OFF=88": 152,
196
+ "NOTE_ON=88": 153,
197
+ "TIME_DELTA=9": 154,
198
+ "NOTE_OFF=24": 155,
199
+ "NOTE_ON=24": 156,
200
+ "NOTE_OFF=87": 157,
201
+ "NOTE_ON=87": 158,
202
+ "TIME_DELTA=11": 159,
203
+ "NOTE_OFF=91": 160,
204
+ "NOTE_ON=91": 161,
205
+ "INST=10": 162,
206
+ "TIME_DELTA=13": 163,
207
+ "NOTE_OFF=25": 164,
208
+ "NOTE_ON=25": 165,
209
+ "NOTE_OFF=89": 166,
210
+ "NOTE_ON=89": 167,
211
+ "INST=8": 168,
212
+ "NOTE_OFF=93": 169,
213
+ "NOTE_ON=93": 170,
214
+ "INST=7": 171,
215
+ "INST=11": 172,
216
+ "NOTE_OFF=90": 173,
217
+ "NOTE_ON=90": 174,
218
+ "NOTE_OFF=22": 175,
219
+ "NOTE_ON=22": 176,
220
+ "INST=2": 177,
221
+ "NOTE_OFF=23": 178,
222
+ "NOTE_ON=23": 179,
223
+ "NOTE_OFF=0": 180,
224
+ "NOTE_ON=0": 181,
225
+ "NOTE_OFF=94": 182,
226
+ "NOTE_ON=94": 183,
227
+ "NOTE_OFF=95": 184,
228
+ "NOTE_ON=95": 185,
229
+ "NOTE_OFF=96": 186,
230
+ "NOTE_ON=96": 187,
231
+ "INST=9": 188,
232
+ "INST=5": 189,
233
+ "INST=12": 190,
234
+ "NOTE_OFF=92": 191,
235
+ "NOTE_ON=92": 192,
236
+ "INST=1": 193,
237
+ "NOTE_OFF=98": 194,
238
+ "NOTE_ON=98": 195,
239
+ "INST=14": 196,
240
+ "NOTE_OFF=20": 197,
241
+ "NOTE_ON=20": 198,
242
+ "NOTE_OFF=100": 199,
243
+ "NOTE_ON=100": 200,
244
+ "INST=13": 201,
245
+ "INST=15": 202,
246
+ "NOTE_OFF=21": 203,
247
+ "NOTE_ON=21": 204,
248
+ "NOTE_OFF=99": 205,
249
+ "NOTE_ON=99": 206,
250
+ "NOTE_OFF=101": 207,
251
+ "NOTE_ON=101": 208,
252
+ "NOTE_OFF=16": 209,
253
+ "NOTE_ON=16": 210,
254
+ "NOTE_OFF=97": 211,
255
+ "NOTE_ON=97": 212,
256
+ "NOTE_OFF=102": 213,
257
+ "NOTE_ON=102": 214,
258
+ "NOTE_OFF=17": 215,
259
+ "NOTE_ON=17": 216,
260
+ "NOTE_OFF=19": 217,
261
+ "NOTE_ON=19": 218,
262
+ "NOTE_OFF=103": 219,
263
+ "NOTE_ON=103": 220,
264
+ "NOTE_OFF=107": 221,
265
+ "NOTE_ON=107": 222,
266
+ "NOTE_OFF=105": 223,
267
+ "NOTE_ON=105": 224,
268
+ "NOTE_OFF=18": 225,
269
+ "NOTE_ON=18": 226,
270
+ "NOTE_OFF=109": 227,
271
+ "NOTE_ON=109": 228,
272
+ "NOTE_OFF=110": 229,
273
+ "NOTE_ON=110": 230,
274
+ "NOTE_OFF=112": 231,
275
+ "NOTE_ON=112": 232,
276
+ "NOTE_OFF=15": 233,
277
+ "NOTE_ON=15": 234,
278
+ "NOTE_OFF=6": 235,
279
+ "NOTE_ON=6": 236,
280
+ "NOTE_OFF=108": 237,
281
+ "NOTE_ON=108": 238,
282
+ "NOTE_OFF=12": 239,
283
+ "NOTE_ON=12": 240,
284
+ "NOTE_OFF=126": 241,
285
+ "NOTE_ON=126": 242,
286
+ "NOTE_OFF=14": 243,
287
+ "NOTE_ON=14": 244,
288
+ "NOTE_OFF=104": 245,
289
+ "NOTE_ON=104": 246,
290
+ "NOTE_OFF=7": 247,
291
+ "NOTE_ON=7": 248,
292
+ "NOTE_OFF=8": 249,
293
+ "NOTE_ON=8": 250,
294
+ "NOTE_OFF=123": 251,
295
+ "NOTE_ON=123": 252,
296
+ "NOTE_OFF=106": 253,
297
+ "NOTE_ON=106": 254,
298
+ "NOTE_OFF=1": 255,
299
+ "NOTE_ON=1": 256,
300
+ "NOTE_OFF=122": 257,
301
+ "NOTE_ON=122": 258,
302
+ "NOTE_OFF=124": 259,
303
+ "NOTE_ON=124": 260,
304
+ "NOTE_OFF=127": 261,
305
+ "NOTE_ON=127": 262,
306
+ "NOTE_OFF=9": 263,
307
+ "NOTE_ON=9": 264,
308
+ "NOTE_OFF=125": 265,
309
+ "NOTE_ON=125": 266,
310
+ "NOTE_OFF=114": 267,
311
+ "NOTE_ON=114": 268,
312
+ "NOTE_OFF=117": 269,
313
+ "NOTE_ON=117": 270,
314
+ "NOTE_OFF=113": 271,
315
+ "NOTE_ON=113": 272,
316
+ "NOTE_OFF=119": 273,
317
+ "NOTE_ON=119": 274,
318
+ "NOTE_OFF=120": 275,
319
+ "NOTE_OFF=2": 276,
320
+ "NOTE_ON=120": 277,
321
+ "NOTE_ON=2": 278,
322
+ "NOTE_OFF=115": 279,
323
+ "NOTE_ON=115": 280,
324
+ "NOTE_OFF=111": 281,
325
+ "NOTE_ON=111": 282,
326
+ "NOTE_OFF=4": 283,
327
+ "NOTE_ON=4": 284,
328
+ "NOTE_OFF=116": 285,
329
+ "NOTE_ON=116": 286,
330
+ "NOTE_OFF=5": 287,
331
+ "NOTE_ON=5": 288,
332
+ "NOTE_OFF=3": 289,
333
+ "NOTE_ON=3": 290,
334
+ "NOTE_OFF=11": 291,
335
+ "NOTE_ON=11": 292,
336
+ "NOTE_OFF=10": 293,
337
+ "NOTE_OFF=118": 294,
338
+ "NOTE_OFF=121": 295,
339
+ "NOTE_ON=10": 296,
340
+ "NOTE_ON=118": 297,
341
+ "NOTE_ON=121": 298,
342
+ "NOTE_OFF=13": 299,
343
+ "NOTE_ON=13": 300
344
+ },
345
+ "unk_token": "[UNK]"
346
+ }
347
+ }
manual_upload/tokenizer_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "model_max_length": 1000000000000000019884624838656,
3
+ "name_or_path": "JammyMachina/elec-gmusic-familized-model-13-12__17-35-53",
4
+ "special_tokens_map_file": "/root/.cache/huggingface/hub/models--JammyMachina--elec-gmusic-familized-model-13-12__17-35-53/snapshots/fbba9d2ac598a2e0fbec338593aceff49347aff4/special_tokens_map.json",
5
+ "tokenizer_class": "PreTrainedTokenizerFast"
6
+ }
manual_upload/trainer_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": null,
5
+ "global_step": 0,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [],
10
+ "max_steps": 0,
11
+ "num_train_epochs": 0,
12
+ "total_flos": 0,
13
+ "trial_name": null,
14
+ "trial_params": null
15
+ }
manual_upload/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0206a7605fed29c8321de91dfcf793fb6150f09dc4e519e1121bb6e0b17b29fc
3
+ size 3515
manual_upload/training_args.json ADDED
@@ -0,0 +1,109 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_dir": "models/elec-gmusic-familized",
3
+ "overwrite_output_dir": true,
4
+ "do_train": false,
5
+ "do_eval": true,
6
+ "do_predict": false,
7
+ "evaluation_strategy": "steps",
8
+ "prediction_loss_only": false,
9
+ "per_device_train_batch_size": 10,
10
+ "per_device_eval_batch_size": 8,
11
+ "per_gpu_train_batch_size": null,
12
+ "per_gpu_eval_batch_size": null,
13
+ "gradient_accumulation_steps": 1,
14
+ "eval_accumulation_steps": null,
15
+ "eval_delay": 0,
16
+ "learning_rate": 0.0005,
17
+ "weight_decay": 0.1,
18
+ "adam_beta1": 0.9,
19
+ "adam_beta2": 0.999,
20
+ "adam_epsilon": 1e-08,
21
+ "max_grad_norm": 1.0,
22
+ "num_train_epochs": 6,
23
+ "max_steps": -1,
24
+ "lr_scheduler_type": "cosine",
25
+ "warmup_ratio": 0.0,
26
+ "warmup_steps": 200,
27
+ "log_level": "passive",
28
+ "log_level_replica": "passive",
29
+ "log_on_each_node": true,
30
+ "logging_dir": "models/elec-gmusic-familized/logs",
31
+ "logging_strategy": "steps",
32
+ "logging_first_step": false,
33
+ "logging_steps": 4096,
34
+ "logging_nan_inf_filter": true,
35
+ "save_strategy": "steps",
36
+ "save_steps": 16384,
37
+ "save_total_limit": 5,
38
+ "save_on_each_node": false,
39
+ "no_cuda": false,
40
+ "use_mps_device": false,
41
+ "seed": 42,
42
+ "data_seed": null,
43
+ "jit_mode_eval": false,
44
+ "use_ipex": false,
45
+ "bf16": false,
46
+ "fp16": true,
47
+ "fp16_opt_level": "O1",
48
+ "half_precision_backend": "cuda_amp",
49
+ "bf16_full_eval": false,
50
+ "fp16_full_eval": false,
51
+ "tf32": null,
52
+ "local_rank": -1,
53
+ "xpu_backend": null,
54
+ "tpu_num_cores": null,
55
+ "tpu_metrics_debug": false,
56
+ "debug": [],
57
+ "dataloader_drop_last": false,
58
+ "eval_steps": 4096,
59
+ "dataloader_num_workers": 0,
60
+ "past_index": -1,
61
+ "run_name": "models/elec-gmusic-familized",
62
+ "disable_tqdm": false,
63
+ "remove_unused_columns": true,
64
+ "label_names": null,
65
+ "load_best_model_at_end": false,
66
+ "metric_for_best_model": null,
67
+ "greater_is_better": null,
68
+ "ignore_data_skip": false,
69
+ "sharded_ddp": [],
70
+ "fsdp": [],
71
+ "fsdp_min_num_params": 0,
72
+ "fsdp_transformer_layer_cls_to_wrap": null,
73
+ "deepspeed": null,
74
+ "label_smoothing_factor": 0.0,
75
+ "optim": "adamw_hf",
76
+ "optim_args": null,
77
+ "adafactor": false,
78
+ "group_by_length": false,
79
+ "length_column_name": "length",
80
+ "report_to": [
81
+ "wandb"
82
+ ],
83
+ "ddp_find_unused_parameters": null,
84
+ "ddp_bucket_cap_mb": null,
85
+ "dataloader_pin_memory": true,
86
+ "skip_memory_metrics": true,
87
+ "use_legacy_prediction_loop": false,
88
+ "push_to_hub": true,
89
+ "resume_from_checkpoint": null,
90
+ "hub_model_id": "JammyMachina/elec-gmusic-familized-model-13-12__17-35-53",
91
+ "hub_strategy": "every_save",
92
+ "hub_token": "<HUB_TOKEN>",
93
+ "hub_private_repo": false,
94
+ "gradient_checkpointing": false,
95
+ "include_inputs_for_metrics": false,
96
+ "fp16_backend": "auto",
97
+ "push_to_hub_model_id": null,
98
+ "push_to_hub_organization": null,
99
+ "push_to_hub_token": "<PUSH_TO_HUB_TOKEN>",
100
+ "mp_parameters": "",
101
+ "auto_find_batch_size": false,
102
+ "full_determinism": false,
103
+ "torchdynamo": null,
104
+ "ray_scope": "last",
105
+ "ddp_timeout": 1800,
106
+ "torch_compile": false,
107
+ "torch_compile_backend": null,
108
+ "torch_compile_mode": null
109
+ }