NeMo
English
nvidia
rlhf
llama3
zhilinw commited on
Commit
e5873af
1 Parent(s): 45faab9

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +0 -0
  2. model_config.yaml +315 -0
  3. model_weights/common.pt +3 -0
  4. model_weights/metadata.json +1 -0
  5. model_weights/model.decoder.final_layernorm.weight/.zarray +14 -0
  6. model_weights/model.decoder.final_layernorm.weight/0 +0 -0
  7. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_0_80.pt +3 -0
  8. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_10_80.pt +3 -0
  9. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_11_80.pt +3 -0
  10. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_12_80.pt +3 -0
  11. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_13_80.pt +3 -0
  12. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_14_80.pt +3 -0
  13. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_15_80.pt +3 -0
  14. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_16_80.pt +3 -0
  15. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_17_80.pt +3 -0
  16. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_18_80.pt +3 -0
  17. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_19_80.pt +3 -0
  18. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_1_80.pt +3 -0
  19. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_20_80.pt +3 -0
  20. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_21_80.pt +3 -0
  21. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_22_80.pt +3 -0
  22. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_23_80.pt +3 -0
  23. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_24_80.pt +3 -0
  24. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_25_80.pt +3 -0
  25. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_26_80.pt +3 -0
  26. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_27_80.pt +3 -0
  27. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_28_80.pt +3 -0
  28. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_29_80.pt +3 -0
  29. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_2_80.pt +3 -0
  30. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_30_80.pt +3 -0
  31. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_31_80.pt +3 -0
  32. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_32_80.pt +3 -0
  33. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_33_80.pt +3 -0
  34. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_34_80.pt +3 -0
  35. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_35_80.pt +3 -0
  36. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_36_80.pt +3 -0
  37. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_37_80.pt +3 -0
  38. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_38_80.pt +3 -0
  39. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_39_80.pt +3 -0
  40. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_3_80.pt +3 -0
  41. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_40_80.pt +3 -0
  42. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_41_80.pt +3 -0
  43. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_42_80.pt +3 -0
  44. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_43_80.pt +3 -0
  45. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_44_80.pt +3 -0
  46. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_45_80.pt +3 -0
  47. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_46_80.pt +3 -0
  48. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_47_80.pt +3 -0
  49. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_48_80.pt +3 -0
  50. model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_49_80.pt +3 -0
.gitattributes CHANGED
The diff for this file is too large to render. See raw diff
 
model_config.yaml ADDED
@@ -0,0 +1,315 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ mcore_gpt: true
2
+ micro_batch_size: 1
3
+ global_batch_size: 128
4
+ tensor_model_parallel_size: 8
5
+ pipeline_model_parallel_size: 8
6
+ virtual_pipeline_model_parallel_size: null
7
+ encoder_seq_length: 8192
8
+ max_position_embeddings: 8192
9
+ num_layers: 80
10
+ hidden_size: 8192
11
+ ffn_hidden_size: 28672
12
+ num_attention_heads: 64
13
+ init_method_std: 0.02
14
+ use_scaled_init_method: true
15
+ hidden_dropout: 0.0
16
+ attention_dropout: 0.0
17
+ ffn_dropout: 0.0
18
+ kv_channels: null
19
+ apply_query_key_layer_scaling: true
20
+ normalization: rmsnorm
21
+ layernorm_epsilon: 1.0e-05
22
+ do_layer_norm_weight_decay: false
23
+ make_vocab_size_divisible_by: 128
24
+ pre_process: true
25
+ post_process: true
26
+ persist_layer_norm: true
27
+ bias: false
28
+ activation: fast-swiglu
29
+ headscale: false
30
+ transformer_block_type: pre_ln
31
+ openai_gelu: false
32
+ normalize_attention_scores: true
33
+ position_embedding_type: rope
34
+ rotary_percentage: 1.0
35
+ attention_type: multihead
36
+ share_embeddings_and_output_weights: false
37
+ overlap_p2p_comm: false
38
+ batch_p2p_comm: true
39
+ num_query_groups: 8
40
+ tokenizer:
41
+ library: huggingface
42
+ type: meta-llama/Meta-Llama-3-70B
43
+ model: /dataset/models/llama2-13b/llama-tokenizer.model
44
+ vocab_file: null
45
+ merge_file: null
46
+ tokenizer_model: /dataset/models/llama2-13b/llama-tokenizer.model
47
+ sentencepiece_legacy: false
48
+ use_fast: true
49
+ native_amp_init_scale: 4294967296
50
+ native_amp_growth_interval: 1000
51
+ hysteresis: 2
52
+ fp32_residual_connection: false
53
+ fp16_lm_cross_entropy: false
54
+ megatron_amp_O2: true
55
+ grad_allreduce_chunk_size_mb: 125
56
+ grad_div_ar_fusion: true
57
+ gradient_accumulation_fusion: false
58
+ bias_activation_fusion: false
59
+ bias_dropout_add_fusion: false
60
+ masked_softmax_fusion: true
61
+ get_attention_mask_from_fusion: true
62
+ apply_rope_fusion: false
63
+ seed: 1234
64
+ resume_from_checkpoint: null
65
+ use_cpu_initialization: false
66
+ onnx_safe: false
67
+ apex_transformer_log_level: 30
68
+ gradient_as_bucket_view: false
69
+ sync_batch_comm: false
70
+ activations_checkpoint_granularity: full
71
+ activations_checkpoint_method: uniform
72
+ activations_checkpoint_num_layers: 1
73
+ num_micro_batches_with_partial_activation_checkpoints: null
74
+ activations_checkpoint_layers_per_pipeline: null
75
+ sequence_parallel: false
76
+ transformer_engine: true
77
+ fp8: false
78
+ fp8_e4m3: false
79
+ fp8_hybrid: true
80
+ fp8_margin: 0
81
+ fp8_interval: 1
82
+ fp8_amax_history_len: 1024
83
+ fp8_amax_compute_algo: max
84
+ reduce_amax: true
85
+ use_emha: false
86
+ data:
87
+ chat: true
88
+ chat_prompt_tokens:
89
+ system_turn_start: <extra_id_0>
90
+ turn_start: <extra_id_1>
91
+ label_start: <extra_id_2>
92
+ end_of_turn: '
93
+
94
+ '
95
+ end_of_name: '
96
+
97
+ '
98
+ sample: true
99
+ num_workers: 2
100
+ dataloader_type: single
101
+ train_ds:
102
+ file_path: /dataset/train.jsonl
103
+ global_batch_size: 384
104
+ micro_batch_size: 1
105
+ shuffle: true
106
+ memmap_workers: null
107
+ max_seq_length: 4096
108
+ min_seq_length: 1
109
+ drop_last: true
110
+ concat_sampling_probabilities: null
111
+ label_key: output
112
+ add_eos: false
113
+ add_sep: false
114
+ add_bos: false
115
+ truncation_field: input
116
+ index_mapping_dir: /indexmap_dir
117
+ prompt_template: '<extra_id_0>System
118
+
119
+ {system message}
120
+
121
+ <extra_id_1>User
122
+
123
+ {turn 1 user message}
124
+
125
+ <extra_id_1>Assistant
126
+
127
+ <extra_id_2>{turn 1 assistant label}
128
+
129
+ {turn 1 assistant message}
130
+
131
+ <extra_id_1>User
132
+
133
+ {turn 2 user message}
134
+
135
+ <extra_id_1>Assistant
136
+
137
+ <extra_id_2>{turn 2 assistant label}
138
+
139
+ {turn 2 assistant message}
140
+
141
+ <extra_id_1>'
142
+ hf_dataset: true
143
+ truncation_method: right
144
+ validation_ds:
145
+ file_path: /dataset/train.jsonl
146
+ names: null
147
+ global_batch_size: 384
148
+ micro_batch_size: 1
149
+ shuffle: false
150
+ memmap_workers: null
151
+ max_seq_length: 4096
152
+ min_seq_length: 1
153
+ drop_last: false
154
+ label_key: output
155
+ add_eos: false
156
+ add_sep: false
157
+ add_bos: false
158
+ write_predictions_to_file: false
159
+ output_file_path_prefix: null
160
+ truncation_field: input
161
+ index_mapping_dir: /indexmap_dir
162
+ prompt_template: '<extra_id_0>System
163
+
164
+ {system message}
165
+
166
+ <extra_id_1>User
167
+
168
+ {turn 1 user message}
169
+
170
+ <extra_id_1>Assistant
171
+
172
+ <extra_id_2>{turn 1 assistant label}
173
+
174
+ {turn 1 assistant message}
175
+
176
+ <extra_id_1>User
177
+
178
+ {turn 2 user message}
179
+
180
+ <extra_id_1>Assistant
181
+
182
+ <extra_id_2>{turn 2 assistant label}
183
+
184
+ {turn 2 assistant message}
185
+
186
+ <extra_id_1>'
187
+ tokens_to_generate: 32
188
+ hf_dataset: true
189
+ truncation_method: right
190
+ metric:
191
+ name: loss
192
+ average: null
193
+ num_classes: null
194
+ test_ds:
195
+ prompt_template: '<extra_id_0>System
196
+
197
+ {system message}
198
+
199
+ <extra_id_1>User
200
+
201
+ {turn 1 user message}
202
+
203
+ <extra_id_1>Assistant
204
+
205
+ <extra_id_2>{turn 1 assistant label}
206
+
207
+ {turn 1 assistant message}
208
+
209
+ <extra_id_1>User
210
+
211
+ {turn 2 user message}
212
+
213
+ <extra_id_1>Assistant
214
+
215
+ <extra_id_2>{turn 2 assistant label}
216
+
217
+ {turn 2 assistant message}
218
+
219
+ <extra_id_1>'
220
+ data_impl: jsonl
221
+ splits_string: null
222
+ seq_length: 8192
223
+ skip_warmup: true
224
+ reset_position_ids: false
225
+ reset_attention_mask: false
226
+ eod_mask_loss: false
227
+ index_mapping_dir: null
228
+ data_prefix:
229
+ train:
230
+ - /lustre/fsw/coreai_dlalgo_llm/geshen/trt_llm/helpsteer_data/helpsteer_11_train_prompts.jsonl
231
+ validation:
232
+ - /lustre/fsw/coreai_dlalgo_llm/geshen/trt_llm/helpsteer_data/helpsteer_11_val_prompts.jsonl
233
+ test:
234
+ - /lustre/fsw/coreai_dlalgo_llm/geshen/trt_llm/helpsteer_data/helpsteer_11_val_prompts.jsonl
235
+ nsys_profile:
236
+ enabled: false
237
+ start_step: 10
238
+ end_step: 10
239
+ ranks:
240
+ - 0
241
+ gen_shape: false
242
+ optim:
243
+ name: distributed_fused_adam
244
+ lr: 2.0e-07
245
+ weight_decay: 0.1
246
+ betas:
247
+ - 0.9
248
+ - 0.98
249
+ sched:
250
+ name: CosineAnnealing
251
+ warmup_steps: 10
252
+ constant_steps: 1000
253
+ min_lr: 1.9999e-07
254
+ max_steps: 88
255
+ bucket_cap_mb: 200
256
+ overlap_grad_sync: false
257
+ contiguous_grad_buffer: true
258
+ rotary_base: 500000.0
259
+ precision: bf16
260
+ answer_only_loss: true
261
+ restore_from_path: /models/unpacked_llama3_70b_base
262
+ save_nemo_on_validation_end: true
263
+ use_flash_attention: null
264
+ pipeline_model_parallel_split_rank: 0
265
+ ppo:
266
+ trt_llm:
267
+ enable: true
268
+ reshard: true
269
+ max_input_len: 4096
270
+ max_input_tokens: 32768
271
+ model_type: LLaMAForCausalLM
272
+ unload_engine_train: false
273
+ rollout_micro_batch_size: 8
274
+ num_rollout_samples: 128
275
+ forward_micro_batch_size: 4
276
+ val_rollout_micro_batch_size: 8
277
+ num_val_samples: 128
278
+ offload_adam_states: true
279
+ entropy_bonus: 0.0
280
+ ratio_eps: 0.2
281
+ sampling_params:
282
+ use_greedy: false
283
+ temperature: 1.0
284
+ top_k: 0
285
+ top_p: 1.0
286
+ repetition_penalty: 1.0
287
+ add_BOS: false
288
+ all_probs: false
289
+ compute_logprob: false
290
+ end_strings:
291
+ - <|endoftext|>
292
+ - <extra_id_1>
293
+ length_params:
294
+ max_length: 2048
295
+ min_length: 1
296
+ peft:
297
+ peft_scheme: none
298
+ restore_from_path: null
299
+ restore_from_ckpt:
300
+ checkpoint_dir: null
301
+ checkpoint_name: null
302
+ lora_tuning:
303
+ target_modules:
304
+ - attention_qkv
305
+ adapter_dim: 32
306
+ adapter_dropout: 0.0
307
+ column_init_method: xavier
308
+ row_init_method: zero
309
+ layer_selection: null
310
+ weight_tying: false
311
+ position_embedding_strategy: null
312
+ offload_adam_states: true
313
+ enable_nge: true
314
+ target: nemo_aligner.models.nlp.gpt.megatron_gpt_ppo_actor.MegatronGPTActorModel
315
+ nemo_version: 2.0.0rc0
model_weights/common.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d508deac1f3085ed140e04c6620939154ae96f07278be1363e143c0a083e3a3f
3
+ size 35715
model_weights/metadata.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"sharded_backend": "zarr", "sharded_backend_version": 1, "common_backend": "torch", "common_backend_version": 1}
model_weights/model.decoder.final_layernorm.weight/.zarray ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunks": [
3
+ 8192
4
+ ],
5
+ "compressor": null,
6
+ "dtype": "bfloat16",
7
+ "fill_value": null,
8
+ "filters": null,
9
+ "order": "C",
10
+ "shape": [
11
+ 8192
12
+ ],
13
+ "zarr_format": 2
14
+ }
model_weights/model.decoder.final_layernorm.weight/0 ADDED
Binary file (16.4 kB). View file
 
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_0_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e9a729975a47eb62643f3ac6a8bf145c719b877b2ce5b967be11c2b18801913
3
+ size 1836
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_10_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faac9f0f6c1329624b1cad1daa596a7833bf7475949da3e7e4ec86709dca4952
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_11_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:158179c28c9e196e80f0eb646eb219fce5fff77cd50a03ab445e028d2ac8bf76
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_12_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87fe00dd30d5e94f76b7c60e50c4fac2b19594743a26da4d5081a1d56fa54c47
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_13_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9a51a0705943fe66ca2db9ac36923dee52b1dadb1b32efcf1c50c1e62eb54f0
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_14_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d96cc54ddbca63e537764caffb8b3602a65f652827354e19907c5980de821b9
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_15_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41cb45e944859eb6dcedcd8e7e56c7332ae31753f6daa3f61289ca04368db067
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_16_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34e4d66af34f6390b2bc1b83973f1ebe6ad6b32b391e0f1795f426ddcbdda783
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_17_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5732c2a0c0f07c0d4d11616be40ad87ac2a80986400cdc1f7d55a98f1d094d4
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_18_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66833ba47bee314a4c94098e985aa831ca98533ee2ec794b1c5dd95149f9f82c
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_19_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01b45b3f4ef6b69193217eb1bbdd50781c76efcb176194b72f28b2caeb617799
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_1_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15b6735d1668ebc2574d56235fd254e7c2977a571e8b2aa717f53158adaa76b8
3
+ size 1836
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_20_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27682c2cd23ed3f1a7a3a5d40c5c84f0ac24ff07e1528665e3574d52ea707d71
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_21_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:040aa73c2be92a7ed400cac3319ef89eede969f585b0e12c2cddd9c01800f2cc
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_22_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02368a8b8351735ecb4d892cea21dea8c5cbaee51e9ab6e0f42994c093e42f53
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_23_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b738c1a25114960e407fa1d35e3417e1b9f9e4a5d4d7e45c777f4ddffaa24a2b
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_24_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c72f2341f3ef6ebc07f2d7f409c29e85a27730089b09fd2c30e2cc1c9e30b17b
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_25_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5130873b235257ae1630a2f83dfb3bfb073e49d469ad1e802cbe797fb4ea1bb
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_26_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:249b210a061b725c547cb1855b4afd0feaef02593d6e8c371fc8d96c7ae019f4
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_27_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1266642f65d8d695707e2eaa9631ea2472f16f4f97d3e22e0065da6875daf252
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_28_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e5b09e3f9f634bf19a80bb9abb0b45f49ee4b3c7f007e35cd8e7925daabbe32
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_29_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba77cb7bb7ee730d9e91fe35e1084a89c45eb32ff363a94133b8f30c50c229f1
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_2_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:026af90a314db680105fea6a1aae98989ebe2d6f52eaf062006550d077bda05a
3
+ size 1836
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_30_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78b87e344c6ce15fc3190ed5741823cc913defc8da276302cd704de63ddf3f85
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_31_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ae569fddb6d7b9d201e4e935ea33b736dee2487c533fe330b1f4b20edbf0175
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_32_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:433659103399b2f5ea6533ab08d9cd213c58e13501de3f3f049a84b8d7a58ac0
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_33_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9169d2efa0cc9cc2c46021569a1140179409233c658576499376f5db9d1d8c8
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_34_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abebd91eaa58bf1577d69ce1c0bfb8b91f05002d3b26edf434c0b7eb55f7475d
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_35_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3d95be77323d1eb8c0716e39d9ddb8d06cfa3cda9c5189d82afb91e459a9146
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_36_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e38c7ed5c1127ea72da184b561f154c2393e512769ee59e77512c6c85f7517a7
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_37_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc79248144069a2751c43f604917cca98519c9fe514dd53be67d2a955d48c7b1
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_38_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c894260cb35b9c59a8501e743bee84de53188279a4d31791f63a33f4ad56f229
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_39_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed58173cd24c9b37daef9ef2c3ceac4da6f2927c2320d3eefb3c222b6b68d8f6
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_3_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28ae5501b97631ba9aff2524f117edd3eb863dfebb29395610f689c851282d58
3
+ size 1836
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_40_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01530e6ee0f9dd2007409746855c10469ff5e93da59ccc5eb4e678cd478a9c17
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_41_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3477e4945e17e402f8190e1d9937627097783ba125917bb0df1760ed25844bda
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_42_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7f9dcbf8f93adbcf221e0ae8117118a069ce4eadeb0c11cc44cf661eb1c860f
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_43_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74147b0cf9114f4966c081b94393a9ada2b3fc27eed18e0ac34d580ed8f5d2d2
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_44_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bc4cbaf569529f6b36bfc4ab323c0581d32135a021837ee0e932ff41f1c5e42
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_45_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56f96e00f6542f620b6ee51e64bc9aa76e3907c9e117e27825af4ef10f4d3cd7
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_46_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36515ee57e2e8b16b0ee22dc8b9176a29aa212a901eadb82a6d133c2487939b9
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_47_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce91aff8cb5ebdeeb73dd871ed4aa4b19efb94861f55757cfc2d3ebb037dca78
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_48_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc878f893ca0b2bd6a4556a1b8ef3e13dabbf7c2001e556a4939383886e2f857
3
+ size 1840
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_49_80.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9187866aa84481d462ec5d43082b88c1c92363c6207de7a1a7b43a7744b2fda6
3
+ size 1840