Muennighoff
commited on
Commit
•
6134937
1
Parent(s):
6381e32
Add
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +6 -0
- 2b855b11bc4/transformers/config.json +1 -0
- 2b855b11bc4/transformers/pytorch_model.bin +3 -0
- 2b855b14bc4/transformers/config.json +1 -0
- 2b855b14bc4/transformers/pytorch_model.bin +3 -0
- 2b855b18bc4/transformers/config.json +1 -0
- 2b855b18bc4/transformers/pytorch_model.bin +3 -0
- 2b855b28bc4/transformers/config.json +1 -0
- 2b855b28bc4/transformers/pytorch_model.bin +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_100_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_101_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_102_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_103_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_104_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_105_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_106_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_107_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_108_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_109_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_110_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_111_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_112_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_113_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_114_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_115_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_116_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_117_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_118_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_119_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_120_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_121_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_122_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_123_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_124_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_125_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_126_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_127_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_128_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_129_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_130_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_131_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_132_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_133_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_134_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_135_mp_rank_00_optim_states.pt +3 -0
- 2b855b55bc4/global_step52452/bf16_zero_pp_rank_136_mp_rank_00_optim_states.pt +3 -0
.gitattributes
CHANGED
@@ -32,3 +32,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
33 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
34 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
35 |
+
evaluation/2b855b11bc4/examples.limited=3000.model=2b855b11bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:13:42.jsonl filter=lfs diff=lfs merge=lfs -text
|
36 |
+
evaluation/2b855b14bc4/examples.limited=3000.model=2b855b14bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:13:28.jsonl filter=lfs diff=lfs merge=lfs -text
|
37 |
+
evaluation/2b855b18bc4/examples.limited=3000.model=2b855b18bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:13:22.jsonl filter=lfs diff=lfs merge=lfs -text
|
38 |
+
evaluation/2b855b9bc4/examples.limited=3000.model=transformers.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T11:32:35.jsonl filter=lfs diff=lfs merge=lfs -text
|
39 |
+
evaluation/2b855b55bc4/examples.limited=3000.model=transformers.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:09:38.jsonl filter=lfs diff=lfs merge=lfs -text
|
40 |
+
evaluation/2b855b28bc4/examples.limited=3000.model=2b855b28bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:12:52.jsonl filter=lfs diff=lfs merge=lfs -text
|
2b855b11bc4/transformers/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
|
2b855b11bc4/transformers/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9255e28a376bbd40430cd2a7ff646d4d94e7e2bd336133ea878e826ff2e761b1
|
3 |
+
size 5903417421
|
2b855b14bc4/transformers/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
|
2b855b14bc4/transformers/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d527f439a2f6307b8b7cf255e3e0f18aa20309e4d3543eabc934e1ae3a37b7d
|
3 |
+
size 5903417421
|
2b855b18bc4/transformers/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
|
2b855b18bc4/transformers/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:737e9b0a4da70f2e23a26943c5c1d9547b9a6bc23bbbb2fa114668df1f0481b2
|
3 |
+
size 5903417421
|
2b855b28bc4/transformers/config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
|
2b855b28bc4/transformers/pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80a56c8859b0f25a8869efbfd5352bda19feb30a4f625af89785c4e56233dc06
|
3 |
+
size 5903417421
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28b947805769ce4b095bf46c21d3bbea85975614669e481ecb4efde5b2762a18
|
3 |
+
size 131677719
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_100_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35fc22bdd46b755ad67e9e7a540e3e402b7589bd79cf733db35b71e79e5c36d9
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_101_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cba32213ef29dde42406d453b48f030f2c90b27bf4f6d5f1decef949af774f70
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_102_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:82bc641f8a0eab8278abd8d5bfdd3cb8a7bf37ade190d6082268d8a0fe4a992a
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_103_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26b6c89df6123cd71022a7ba20f16f5e25898906c1dfa7a81f412495f2f3070c
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_104_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d80ef6e7e2f90a091aa40350189d057efa3acdc9c6bf8557829164f409172358
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_105_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9bef9a9d047faaac418831cd8a9cefb6185a4f43ae863304cd54f796188b0a14
|
3 |
+
size 131677677
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_106_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7de6b2f1d197c4a39536283df33b2bedaa87cdc6b19f7dc9ada808b27b68e80b
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_107_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28e90a1ebe25a038b6baacf865d34fb6849031c0307e11d991ba78d5a21d03ab
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_108_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:147655c0529a2c37212ed0039c24c1439fc6ba4ac0404f0a3236a92f63da7e2a
|
3 |
+
size 131677869
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_109_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed6443a88a75db6174d69cdc0bba56de01ca02f8bf2ee2fcc5515fcc605be368
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7fd74f4287e6c48e8eb7fb1dd650d761dd10e7b14eaac566e8dfd45eab83ff00
|
3 |
+
size 131677794
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_110_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e0121d8867148adb205cf26baadc27e20096e837fae6e64ab288842f4d8384c7
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_111_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d458cfd0017d283e453013abfdd9011025b25961ad4c6303b27ab9f2acd78a63
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_112_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f396570291f16242ee1abcb166fb6f1ed15915e55b1d89bd48d45bdd61d8dae
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_113_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae6e7a0d9467c4593b82fb4e578baadffce84c2b5b7ec63dc6c98b992b2612e6
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_114_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:afca160365fdce5370fbe40b2ef4dbc68f52e3e2f61ea9a410a8aad353d9346f
|
3 |
+
size 131677677
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_115_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:624601c13baf354c7e23b66a15b131fee0a9e5f751d90e3b32926a295cfe88e6
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_116_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab6c0c9c19d4c348fcd8d084c45310b2a0bb9206d32365ccbd1829e61e4450d8
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_117_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1f5b36f215b3f8981c79527bb2706bc3503465d9477c8bcf415ff159ec9409e1
|
3 |
+
size 131677677
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_118_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1476f4057acc3d42aa618a4fe4e9d9cefc46f154b2902da9a4447e0ee65c9ea
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_119_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1c27ff3d40d98a134f94fb411ca49a8c2ff42765bd3f7e2c23b1c94a7a32dd4
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5545c184838c2d4dd6a1621e370ea3ea6fe55dc88d124212babcca4e897530e
|
3 |
+
size 131677730
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_120_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cc5301bc12510a88381d37f9826df93ba495905620821f49cf06a6094224fc70
|
3 |
+
size 131677677
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_121_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bddbd274aa0bcdd4ab2b5f898f86b57be824e9d362545724249084329c70966a
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_122_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa4a355553a8bd26c930542a19239621330609fe0c73f67b6fe626b241dd5fc8
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_123_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dfe4e2348f7c482db041641be5b38d5e718cd0dd0fe9498b502831af85b8bd02
|
3 |
+
size 131677869
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_124_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08ffd5a9a6c1ac9741db247e0e6f63a0ec90683d3b82afb053430353c43d94a7
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_125_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9a613bacd59f87cdae87bd23b84fdb57f5a71e544336f942b6450242da9becda
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_126_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ea73a7fb74c469de7fa2784007b891a1f4790c8b6272d8390ebf2341a12c2d0
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_127_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:346577ecdff05cf304104966558bc389b32b84b7f0a335515bc297fde510d00d
|
3 |
+
size 131677677
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_128_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eab76fcf1eecaebb0fe2cc0b4a693e43881c4542cac18cddc3e78d9d5e069d51
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_129_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2bd7c3b658dc862b24d166a8b3320fa2be1bb215172567edac2e02b9bb85ea38
|
3 |
+
size 131677677
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e384fbaffe3ab864a62916568060b9254f0462ef1efe843a4f55b98b72874ca5
|
3 |
+
size 131677730
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_130_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7be774c032701feba3530277d2bf386e67c12849692d27c0a699c9fee95884a4
|
3 |
+
size 131677869
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_131_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4b7900a0c06848b3b5ab55985a6d4835023846b9dc0a6f566bedeb9a01c93d59
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_132_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7355d772023964fb364ea6ac33ab1b224cf7ad40f1603c8bc3c12dc1d386f583
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_133_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3005451dba6d3ba3dee2ed457ecf0742d58f21df7d0e3db15fe8208a9ca95272
|
3 |
+
size 131677677
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_134_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f07068654a7f9e449d53fb02339b550ab0ff761d52f6139837ec40238dff465b
|
3 |
+
size 131677741
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_135_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d714b97e6edc45f957667729a03a7328fb3b49aef86ef3147a7db47008d125e8
|
3 |
+
size 131677805
|
2b855b55bc4/global_step52452/bf16_zero_pp_rank_136_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ddc0f84014d7e379037ad26626b824ee4461b4ba7e7b13b57da9a9411caab118
|
3 |
+
size 131677741
|