Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +0 -0
- model_config.yaml +315 -0
- model_weights/common.pt +3 -0
- model_weights/metadata.json +1 -0
- model_weights/model.decoder.final_layernorm.weight/.zarray +14 -0
- model_weights/model.decoder.final_layernorm.weight/0 +0 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_0_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_10_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_11_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_12_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_13_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_14_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_15_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_16_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_17_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_18_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_19_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_1_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_20_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_21_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_22_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_23_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_24_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_25_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_26_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_27_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_28_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_29_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_2_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_30_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_31_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_32_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_33_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_34_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_35_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_36_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_37_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_38_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_39_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_3_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_40_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_41_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_42_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_43_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_44_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_45_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_46_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_47_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_48_80.pt +3 -0
- model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_49_80.pt +3 -0
.gitattributes
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
model_config.yaml
ADDED
@@ -0,0 +1,315 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
mcore_gpt: true
|
2 |
+
micro_batch_size: 1
|
3 |
+
global_batch_size: 128
|
4 |
+
tensor_model_parallel_size: 8
|
5 |
+
pipeline_model_parallel_size: 8
|
6 |
+
virtual_pipeline_model_parallel_size: null
|
7 |
+
encoder_seq_length: 8192
|
8 |
+
max_position_embeddings: 8192
|
9 |
+
num_layers: 80
|
10 |
+
hidden_size: 8192
|
11 |
+
ffn_hidden_size: 28672
|
12 |
+
num_attention_heads: 64
|
13 |
+
init_method_std: 0.02
|
14 |
+
use_scaled_init_method: true
|
15 |
+
hidden_dropout: 0.0
|
16 |
+
attention_dropout: 0.0
|
17 |
+
ffn_dropout: 0.0
|
18 |
+
kv_channels: null
|
19 |
+
apply_query_key_layer_scaling: true
|
20 |
+
normalization: rmsnorm
|
21 |
+
layernorm_epsilon: 1.0e-05
|
22 |
+
do_layer_norm_weight_decay: false
|
23 |
+
make_vocab_size_divisible_by: 128
|
24 |
+
pre_process: true
|
25 |
+
post_process: true
|
26 |
+
persist_layer_norm: true
|
27 |
+
bias: false
|
28 |
+
activation: fast-swiglu
|
29 |
+
headscale: false
|
30 |
+
transformer_block_type: pre_ln
|
31 |
+
openai_gelu: false
|
32 |
+
normalize_attention_scores: true
|
33 |
+
position_embedding_type: rope
|
34 |
+
rotary_percentage: 1.0
|
35 |
+
attention_type: multihead
|
36 |
+
share_embeddings_and_output_weights: false
|
37 |
+
overlap_p2p_comm: false
|
38 |
+
batch_p2p_comm: true
|
39 |
+
num_query_groups: 8
|
40 |
+
tokenizer:
|
41 |
+
library: huggingface
|
42 |
+
type: meta-llama/Meta-Llama-3-70B
|
43 |
+
model: /dataset/models/llama2-13b/llama-tokenizer.model
|
44 |
+
vocab_file: null
|
45 |
+
merge_file: null
|
46 |
+
tokenizer_model: /dataset/models/llama2-13b/llama-tokenizer.model
|
47 |
+
sentencepiece_legacy: false
|
48 |
+
use_fast: true
|
49 |
+
native_amp_init_scale: 4294967296
|
50 |
+
native_amp_growth_interval: 1000
|
51 |
+
hysteresis: 2
|
52 |
+
fp32_residual_connection: false
|
53 |
+
fp16_lm_cross_entropy: false
|
54 |
+
megatron_amp_O2: true
|
55 |
+
grad_allreduce_chunk_size_mb: 125
|
56 |
+
grad_div_ar_fusion: true
|
57 |
+
gradient_accumulation_fusion: false
|
58 |
+
bias_activation_fusion: false
|
59 |
+
bias_dropout_add_fusion: false
|
60 |
+
masked_softmax_fusion: true
|
61 |
+
get_attention_mask_from_fusion: true
|
62 |
+
apply_rope_fusion: false
|
63 |
+
seed: 1234
|
64 |
+
resume_from_checkpoint: null
|
65 |
+
use_cpu_initialization: false
|
66 |
+
onnx_safe: false
|
67 |
+
apex_transformer_log_level: 30
|
68 |
+
gradient_as_bucket_view: false
|
69 |
+
sync_batch_comm: false
|
70 |
+
activations_checkpoint_granularity: full
|
71 |
+
activations_checkpoint_method: uniform
|
72 |
+
activations_checkpoint_num_layers: 1
|
73 |
+
num_micro_batches_with_partial_activation_checkpoints: null
|
74 |
+
activations_checkpoint_layers_per_pipeline: null
|
75 |
+
sequence_parallel: false
|
76 |
+
transformer_engine: true
|
77 |
+
fp8: false
|
78 |
+
fp8_e4m3: false
|
79 |
+
fp8_hybrid: true
|
80 |
+
fp8_margin: 0
|
81 |
+
fp8_interval: 1
|
82 |
+
fp8_amax_history_len: 1024
|
83 |
+
fp8_amax_compute_algo: max
|
84 |
+
reduce_amax: true
|
85 |
+
use_emha: false
|
86 |
+
data:
|
87 |
+
chat: true
|
88 |
+
chat_prompt_tokens:
|
89 |
+
system_turn_start: <extra_id_0>
|
90 |
+
turn_start: <extra_id_1>
|
91 |
+
label_start: <extra_id_2>
|
92 |
+
end_of_turn: '
|
93 |
+
|
94 |
+
'
|
95 |
+
end_of_name: '
|
96 |
+
|
97 |
+
'
|
98 |
+
sample: true
|
99 |
+
num_workers: 2
|
100 |
+
dataloader_type: single
|
101 |
+
train_ds:
|
102 |
+
file_path: /dataset/train.jsonl
|
103 |
+
global_batch_size: 384
|
104 |
+
micro_batch_size: 1
|
105 |
+
shuffle: true
|
106 |
+
memmap_workers: null
|
107 |
+
max_seq_length: 4096
|
108 |
+
min_seq_length: 1
|
109 |
+
drop_last: true
|
110 |
+
concat_sampling_probabilities: null
|
111 |
+
label_key: output
|
112 |
+
add_eos: false
|
113 |
+
add_sep: false
|
114 |
+
add_bos: false
|
115 |
+
truncation_field: input
|
116 |
+
index_mapping_dir: /indexmap_dir
|
117 |
+
prompt_template: '<extra_id_0>System
|
118 |
+
|
119 |
+
{system message}
|
120 |
+
|
121 |
+
<extra_id_1>User
|
122 |
+
|
123 |
+
{turn 1 user message}
|
124 |
+
|
125 |
+
<extra_id_1>Assistant
|
126 |
+
|
127 |
+
<extra_id_2>{turn 1 assistant label}
|
128 |
+
|
129 |
+
{turn 1 assistant message}
|
130 |
+
|
131 |
+
<extra_id_1>User
|
132 |
+
|
133 |
+
{turn 2 user message}
|
134 |
+
|
135 |
+
<extra_id_1>Assistant
|
136 |
+
|
137 |
+
<extra_id_2>{turn 2 assistant label}
|
138 |
+
|
139 |
+
{turn 2 assistant message}
|
140 |
+
|
141 |
+
<extra_id_1>'
|
142 |
+
hf_dataset: true
|
143 |
+
truncation_method: right
|
144 |
+
validation_ds:
|
145 |
+
file_path: /dataset/train.jsonl
|
146 |
+
names: null
|
147 |
+
global_batch_size: 384
|
148 |
+
micro_batch_size: 1
|
149 |
+
shuffle: false
|
150 |
+
memmap_workers: null
|
151 |
+
max_seq_length: 4096
|
152 |
+
min_seq_length: 1
|
153 |
+
drop_last: false
|
154 |
+
label_key: output
|
155 |
+
add_eos: false
|
156 |
+
add_sep: false
|
157 |
+
add_bos: false
|
158 |
+
write_predictions_to_file: false
|
159 |
+
output_file_path_prefix: null
|
160 |
+
truncation_field: input
|
161 |
+
index_mapping_dir: /indexmap_dir
|
162 |
+
prompt_template: '<extra_id_0>System
|
163 |
+
|
164 |
+
{system message}
|
165 |
+
|
166 |
+
<extra_id_1>User
|
167 |
+
|
168 |
+
{turn 1 user message}
|
169 |
+
|
170 |
+
<extra_id_1>Assistant
|
171 |
+
|
172 |
+
<extra_id_2>{turn 1 assistant label}
|
173 |
+
|
174 |
+
{turn 1 assistant message}
|
175 |
+
|
176 |
+
<extra_id_1>User
|
177 |
+
|
178 |
+
{turn 2 user message}
|
179 |
+
|
180 |
+
<extra_id_1>Assistant
|
181 |
+
|
182 |
+
<extra_id_2>{turn 2 assistant label}
|
183 |
+
|
184 |
+
{turn 2 assistant message}
|
185 |
+
|
186 |
+
<extra_id_1>'
|
187 |
+
tokens_to_generate: 32
|
188 |
+
hf_dataset: true
|
189 |
+
truncation_method: right
|
190 |
+
metric:
|
191 |
+
name: loss
|
192 |
+
average: null
|
193 |
+
num_classes: null
|
194 |
+
test_ds:
|
195 |
+
prompt_template: '<extra_id_0>System
|
196 |
+
|
197 |
+
{system message}
|
198 |
+
|
199 |
+
<extra_id_1>User
|
200 |
+
|
201 |
+
{turn 1 user message}
|
202 |
+
|
203 |
+
<extra_id_1>Assistant
|
204 |
+
|
205 |
+
<extra_id_2>{turn 1 assistant label}
|
206 |
+
|
207 |
+
{turn 1 assistant message}
|
208 |
+
|
209 |
+
<extra_id_1>User
|
210 |
+
|
211 |
+
{turn 2 user message}
|
212 |
+
|
213 |
+
<extra_id_1>Assistant
|
214 |
+
|
215 |
+
<extra_id_2>{turn 2 assistant label}
|
216 |
+
|
217 |
+
{turn 2 assistant message}
|
218 |
+
|
219 |
+
<extra_id_1>'
|
220 |
+
data_impl: jsonl
|
221 |
+
splits_string: null
|
222 |
+
seq_length: 8192
|
223 |
+
skip_warmup: true
|
224 |
+
reset_position_ids: false
|
225 |
+
reset_attention_mask: false
|
226 |
+
eod_mask_loss: false
|
227 |
+
index_mapping_dir: null
|
228 |
+
data_prefix:
|
229 |
+
train:
|
230 |
+
- /lustre/fsw/coreai_dlalgo_llm/geshen/trt_llm/helpsteer_data/helpsteer_11_train_prompts.jsonl
|
231 |
+
validation:
|
232 |
+
- /lustre/fsw/coreai_dlalgo_llm/geshen/trt_llm/helpsteer_data/helpsteer_11_val_prompts.jsonl
|
233 |
+
test:
|
234 |
+
- /lustre/fsw/coreai_dlalgo_llm/geshen/trt_llm/helpsteer_data/helpsteer_11_val_prompts.jsonl
|
235 |
+
nsys_profile:
|
236 |
+
enabled: false
|
237 |
+
start_step: 10
|
238 |
+
end_step: 10
|
239 |
+
ranks:
|
240 |
+
- 0
|
241 |
+
gen_shape: false
|
242 |
+
optim:
|
243 |
+
name: distributed_fused_adam
|
244 |
+
lr: 2.0e-07
|
245 |
+
weight_decay: 0.1
|
246 |
+
betas:
|
247 |
+
- 0.9
|
248 |
+
- 0.98
|
249 |
+
sched:
|
250 |
+
name: CosineAnnealing
|
251 |
+
warmup_steps: 10
|
252 |
+
constant_steps: 1000
|
253 |
+
min_lr: 1.9999e-07
|
254 |
+
max_steps: 88
|
255 |
+
bucket_cap_mb: 200
|
256 |
+
overlap_grad_sync: false
|
257 |
+
contiguous_grad_buffer: true
|
258 |
+
rotary_base: 500000.0
|
259 |
+
precision: bf16
|
260 |
+
answer_only_loss: true
|
261 |
+
restore_from_path: /models/unpacked_llama3_70b_base
|
262 |
+
save_nemo_on_validation_end: true
|
263 |
+
use_flash_attention: null
|
264 |
+
pipeline_model_parallel_split_rank: 0
|
265 |
+
ppo:
|
266 |
+
trt_llm:
|
267 |
+
enable: true
|
268 |
+
reshard: true
|
269 |
+
max_input_len: 4096
|
270 |
+
max_input_tokens: 32768
|
271 |
+
model_type: LLaMAForCausalLM
|
272 |
+
unload_engine_train: false
|
273 |
+
rollout_micro_batch_size: 8
|
274 |
+
num_rollout_samples: 128
|
275 |
+
forward_micro_batch_size: 4
|
276 |
+
val_rollout_micro_batch_size: 8
|
277 |
+
num_val_samples: 128
|
278 |
+
offload_adam_states: true
|
279 |
+
entropy_bonus: 0.0
|
280 |
+
ratio_eps: 0.2
|
281 |
+
sampling_params:
|
282 |
+
use_greedy: false
|
283 |
+
temperature: 1.0
|
284 |
+
top_k: 0
|
285 |
+
top_p: 1.0
|
286 |
+
repetition_penalty: 1.0
|
287 |
+
add_BOS: false
|
288 |
+
all_probs: false
|
289 |
+
compute_logprob: false
|
290 |
+
end_strings:
|
291 |
+
- <|endoftext|>
|
292 |
+
- <extra_id_1>
|
293 |
+
length_params:
|
294 |
+
max_length: 2048
|
295 |
+
min_length: 1
|
296 |
+
peft:
|
297 |
+
peft_scheme: none
|
298 |
+
restore_from_path: null
|
299 |
+
restore_from_ckpt:
|
300 |
+
checkpoint_dir: null
|
301 |
+
checkpoint_name: null
|
302 |
+
lora_tuning:
|
303 |
+
target_modules:
|
304 |
+
- attention_qkv
|
305 |
+
adapter_dim: 32
|
306 |
+
adapter_dropout: 0.0
|
307 |
+
column_init_method: xavier
|
308 |
+
row_init_method: zero
|
309 |
+
layer_selection: null
|
310 |
+
weight_tying: false
|
311 |
+
position_embedding_strategy: null
|
312 |
+
offload_adam_states: true
|
313 |
+
enable_nge: true
|
314 |
+
target: nemo_aligner.models.nlp.gpt.megatron_gpt_ppo_actor.MegatronGPTActorModel
|
315 |
+
nemo_version: 2.0.0rc0
|
model_weights/common.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d508deac1f3085ed140e04c6620939154ae96f07278be1363e143c0a083e3a3f
|
3 |
+
size 35715
|
model_weights/metadata.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"sharded_backend": "zarr", "sharded_backend_version": 1, "common_backend": "torch", "common_backend_version": 1}
|
model_weights/model.decoder.final_layernorm.weight/.zarray
ADDED
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"chunks": [
|
3 |
+
8192
|
4 |
+
],
|
5 |
+
"compressor": null,
|
6 |
+
"dtype": "bfloat16",
|
7 |
+
"fill_value": null,
|
8 |
+
"filters": null,
|
9 |
+
"order": "C",
|
10 |
+
"shape": [
|
11 |
+
8192
|
12 |
+
],
|
13 |
+
"zarr_format": 2
|
14 |
+
}
|
model_weights/model.decoder.final_layernorm.weight/0
ADDED
Binary file (16.4 kB). View file
|
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_0_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e9a729975a47eb62643f3ac6a8bf145c719b877b2ce5b967be11c2b18801913
|
3 |
+
size 1836
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_10_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:faac9f0f6c1329624b1cad1daa596a7833bf7475949da3e7e4ec86709dca4952
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_11_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:158179c28c9e196e80f0eb646eb219fce5fff77cd50a03ab445e028d2ac8bf76
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_12_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:87fe00dd30d5e94f76b7c60e50c4fac2b19594743a26da4d5081a1d56fa54c47
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_13_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c9a51a0705943fe66ca2db9ac36923dee52b1dadb1b32efcf1c50c1e62eb54f0
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_14_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d96cc54ddbca63e537764caffb8b3602a65f652827354e19907c5980de821b9
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_15_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41cb45e944859eb6dcedcd8e7e56c7332ae31753f6daa3f61289ca04368db067
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_16_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:34e4d66af34f6390b2bc1b83973f1ebe6ad6b32b391e0f1795f426ddcbdda783
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_17_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e5732c2a0c0f07c0d4d11616be40ad87ac2a80986400cdc1f7d55a98f1d094d4
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_18_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:66833ba47bee314a4c94098e985aa831ca98533ee2ec794b1c5dd95149f9f82c
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_19_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01b45b3f4ef6b69193217eb1bbdd50781c76efcb176194b72f28b2caeb617799
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_1_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15b6735d1668ebc2574d56235fd254e7c2977a571e8b2aa717f53158adaa76b8
|
3 |
+
size 1836
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_20_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:27682c2cd23ed3f1a7a3a5d40c5c84f0ac24ff07e1528665e3574d52ea707d71
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_21_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:040aa73c2be92a7ed400cac3319ef89eede969f585b0e12c2cddd9c01800f2cc
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_22_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02368a8b8351735ecb4d892cea21dea8c5cbaee51e9ab6e0f42994c093e42f53
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_23_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b738c1a25114960e407fa1d35e3417e1b9f9e4a5d4d7e45c777f4ddffaa24a2b
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_24_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c72f2341f3ef6ebc07f2d7f409c29e85a27730089b09fd2c30e2cc1c9e30b17b
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_25_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e5130873b235257ae1630a2f83dfb3bfb073e49d469ad1e802cbe797fb4ea1bb
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_26_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:249b210a061b725c547cb1855b4afd0feaef02593d6e8c371fc8d96c7ae019f4
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_27_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1266642f65d8d695707e2eaa9631ea2472f16f4f97d3e22e0065da6875daf252
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_28_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3e5b09e3f9f634bf19a80bb9abb0b45f49ee4b3c7f007e35cd8e7925daabbe32
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_29_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba77cb7bb7ee730d9e91fe35e1084a89c45eb32ff363a94133b8f30c50c229f1
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_2_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:026af90a314db680105fea6a1aae98989ebe2d6f52eaf062006550d077bda05a
|
3 |
+
size 1836
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_30_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78b87e344c6ce15fc3190ed5741823cc913defc8da276302cd704de63ddf3f85
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_31_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0ae569fddb6d7b9d201e4e935ea33b736dee2487c533fe330b1f4b20edbf0175
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_32_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:433659103399b2f5ea6533ab08d9cd213c58e13501de3f3f049a84b8d7a58ac0
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_33_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9169d2efa0cc9cc2c46021569a1140179409233c658576499376f5db9d1d8c8
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_34_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:abebd91eaa58bf1577d69ce1c0bfb8b91f05002d3b26edf434c0b7eb55f7475d
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_35_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3d95be77323d1eb8c0716e39d9ddb8d06cfa3cda9c5189d82afb91e459a9146
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_36_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e38c7ed5c1127ea72da184b561f154c2393e512769ee59e77512c6c85f7517a7
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_37_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc79248144069a2751c43f604917cca98519c9fe514dd53be67d2a955d48c7b1
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_38_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c894260cb35b9c59a8501e743bee84de53188279a4d31791f63a33f4ad56f229
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_39_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed58173cd24c9b37daef9ef2c3ceac4da6f2927c2320d3eefb3c222b6b68d8f6
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_3_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28ae5501b97631ba9aff2524f117edd3eb863dfebb29395610f689c851282d58
|
3 |
+
size 1836
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_40_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01530e6ee0f9dd2007409746855c10469ff5e93da59ccc5eb4e678cd478a9c17
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_41_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3477e4945e17e402f8190e1d9937627097783ba125917bb0df1760ed25844bda
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_42_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a7f9dcbf8f93adbcf221e0ae8117118a069ce4eadeb0c11cc44cf661eb1c860f
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_43_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74147b0cf9114f4966c081b94393a9ada2b3fc27eed18e0ac34d580ed8f5d2d2
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_44_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2bc4cbaf569529f6b36bfc4ab323c0581d32135a021837ee0e932ff41f1c5e42
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_45_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:56f96e00f6542f620b6ee51e64bc9aa76e3907c9e117e27825af4ef10f4d3cd7
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_46_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36515ee57e2e8b16b0ee22dc8b9176a29aa212a901eadb82a6d133c2487939b9
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_47_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ce91aff8cb5ebdeeb73dd871ed4aa4b19efb94861f55757cfc2d3ebb037dca78
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_48_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc878f893ca0b2bd6a4556a1b8ef3e13dabbf7c2001e556a4939383886e2f857
|
3 |
+
size 1840
|
model_weights/model.decoder.layers.mlp.linear_fc1._extra_state/shard_49_80.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9187866aa84481d462ec5d43082b88c1c92363c6207de7a1a7b43a7744b2fda6
|
3 |
+
size 1840
|