Muennighoff commited on
Commit
6134937
1 Parent(s): 6381e32
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. 2b855b11bc4/transformers/config.json +1 -0
  3. 2b855b11bc4/transformers/pytorch_model.bin +3 -0
  4. 2b855b14bc4/transformers/config.json +1 -0
  5. 2b855b14bc4/transformers/pytorch_model.bin +3 -0
  6. 2b855b18bc4/transformers/config.json +1 -0
  7. 2b855b18bc4/transformers/pytorch_model.bin +3 -0
  8. 2b855b28bc4/transformers/config.json +1 -0
  9. 2b855b28bc4/transformers/pytorch_model.bin +3 -0
  10. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  11. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_100_mp_rank_00_optim_states.pt +3 -0
  12. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_101_mp_rank_00_optim_states.pt +3 -0
  13. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_102_mp_rank_00_optim_states.pt +3 -0
  14. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_103_mp_rank_00_optim_states.pt +3 -0
  15. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_104_mp_rank_00_optim_states.pt +3 -0
  16. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_105_mp_rank_00_optim_states.pt +3 -0
  17. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_106_mp_rank_00_optim_states.pt +3 -0
  18. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_107_mp_rank_00_optim_states.pt +3 -0
  19. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_108_mp_rank_00_optim_states.pt +3 -0
  20. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_109_mp_rank_00_optim_states.pt +3 -0
  21. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt +3 -0
  22. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_110_mp_rank_00_optim_states.pt +3 -0
  23. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_111_mp_rank_00_optim_states.pt +3 -0
  24. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_112_mp_rank_00_optim_states.pt +3 -0
  25. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_113_mp_rank_00_optim_states.pt +3 -0
  26. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_114_mp_rank_00_optim_states.pt +3 -0
  27. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_115_mp_rank_00_optim_states.pt +3 -0
  28. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_116_mp_rank_00_optim_states.pt +3 -0
  29. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_117_mp_rank_00_optim_states.pt +3 -0
  30. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_118_mp_rank_00_optim_states.pt +3 -0
  31. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_119_mp_rank_00_optim_states.pt +3 -0
  32. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt +3 -0
  33. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_120_mp_rank_00_optim_states.pt +3 -0
  34. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_121_mp_rank_00_optim_states.pt +3 -0
  35. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_122_mp_rank_00_optim_states.pt +3 -0
  36. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_123_mp_rank_00_optim_states.pt +3 -0
  37. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_124_mp_rank_00_optim_states.pt +3 -0
  38. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_125_mp_rank_00_optim_states.pt +3 -0
  39. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_126_mp_rank_00_optim_states.pt +3 -0
  40. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_127_mp_rank_00_optim_states.pt +3 -0
  41. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_128_mp_rank_00_optim_states.pt +3 -0
  42. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_129_mp_rank_00_optim_states.pt +3 -0
  43. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt +3 -0
  44. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_130_mp_rank_00_optim_states.pt +3 -0
  45. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_131_mp_rank_00_optim_states.pt +3 -0
  46. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_132_mp_rank_00_optim_states.pt +3 -0
  47. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_133_mp_rank_00_optim_states.pt +3 -0
  48. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_134_mp_rank_00_optim_states.pt +3 -0
  49. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_135_mp_rank_00_optim_states.pt +3 -0
  50. 2b855b55bc4/global_step52452/bf16_zero_pp_rank_136_mp_rank_00_optim_states.pt +3 -0
.gitattributes CHANGED
@@ -32,3 +32,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ evaluation/2b855b11bc4/examples.limited=3000.model=2b855b11bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:13:42.jsonl filter=lfs diff=lfs merge=lfs -text
36
+ evaluation/2b855b14bc4/examples.limited=3000.model=2b855b14bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:13:28.jsonl filter=lfs diff=lfs merge=lfs -text
37
+ evaluation/2b855b18bc4/examples.limited=3000.model=2b855b18bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:13:22.jsonl filter=lfs diff=lfs merge=lfs -text
38
+ evaluation/2b855b9bc4/examples.limited=3000.model=transformers.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T11:32:35.jsonl filter=lfs diff=lfs merge=lfs -text
39
+ evaluation/2b855b55bc4/examples.limited=3000.model=transformers.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:09:38.jsonl filter=lfs diff=lfs merge=lfs -text
40
+ evaluation/2b855b28bc4/examples.limited=3000.model=2b855b28bc4.task=GEM-wiki_lingua_en.templates=tldr_en.fewshot=1.batchsize=16.seed=1234.timestamp=2022-12-28T12:12:52.jsonl filter=lfs diff=lfs merge=lfs -text
2b855b11bc4/transformers/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
2b855b11bc4/transformers/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9255e28a376bbd40430cd2a7ff646d4d94e7e2bd336133ea878e826ff2e761b1
3
+ size 5903417421
2b855b14bc4/transformers/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
2b855b14bc4/transformers/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d527f439a2f6307b8b7cf255e3e0f18aa20309e4d3543eabc934e1ae3a37b7d
3
+ size 5903417421
2b855b18bc4/transformers/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
2b855b18bc4/transformers/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:737e9b0a4da70f2e23a26943c5c1d9547b9a6bc23bbbb2fa114668df1f0481b2
3
+ size 5903417421
2b855b28bc4/transformers/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"vocab_size": 50304, "n_positions": 2048, "n_embd": 2560, "n_layer": 34, "n_head": 20, "n_inner": 10240, "activation_function": "gelu", "resid_pdrop": 0.1, "embd_pdrop": 0.1, "attn_pdrop": 0.1, "layer_norm_epsilon": 1e-05, "initializer_range": 0.02, "summary_type": "cls_index", "summary_use_proj": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "scale_attn_weights": true, "use_cache": true, "scale_attn_by_inverse_layer_idx": false, "reorder_and_upcast_attn": false, "bos_token_id": 50256, "eos_token_id": 50256, "return_dict": true, "output_hidden_states": false, "output_attentions": false, "torchscript": false, "torch_dtype": null, "use_bfloat16": false, "tf_legacy_loss": false, "pruned_heads": {}, "tie_word_embeddings": true, "is_encoder_decoder": false, "is_decoder": false, "cross_attention_hidden_size": null, "add_cross_attention": false, "tie_encoder_decoder": false, "max_length": 20, "min_length": 0, "do_sample": false, "early_stopping": false, "num_beams": 1, "num_beam_groups": 1, "diversity_penalty": 0.0, "temperature": 1.0, "top_k": 50, "top_p": 1.0, "typical_p": 1.0, "repetition_penalty": 1.0, "length_penalty": 1.0, "no_repeat_ngram_size": 0, "encoder_no_repeat_ngram_size": 0, "bad_words_ids": null, "num_return_sequences": 1, "chunk_size_feed_forward": 0, "output_scores": false, "return_dict_in_generate": false, "forced_bos_token_id": null, "forced_eos_token_id": null, "remove_invalid_values": false, "exponential_decay_length_penalty": null, "suppress_tokens": null, "begin_suppress_tokens": null, "architectures": ["GPT2LMHeadModel"], "finetuning_task": null, "id2label": {"0": "LABEL_0", "1": "LABEL_1"}, "label2id": {"LABEL_0": 0, "LABEL_1": 1}, "tokenizer_class": null, "prefix": null, "pad_token_id": null, "sep_token_id": null, "decoder_start_token_id": null, "task_specific_params": null, "problem_type": null, "_name_or_path": "", "transformers_version": "4.25.0.dev0", "n_ctx": 1024, "gradient_checkpointing": false, "model_type": "gpt2"}
2b855b28bc4/transformers/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80a56c8859b0f25a8869efbfd5352bda19feb30a4f625af89785c4e56233dc06
3
+ size 5903417421
2b855b55bc4/global_step52452/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28b947805769ce4b095bf46c21d3bbea85975614669e481ecb4efde5b2762a18
3
+ size 131677719
2b855b55bc4/global_step52452/bf16_zero_pp_rank_100_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35fc22bdd46b755ad67e9e7a540e3e402b7589bd79cf733db35b71e79e5c36d9
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_101_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cba32213ef29dde42406d453b48f030f2c90b27bf4f6d5f1decef949af774f70
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_102_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82bc641f8a0eab8278abd8d5bfdd3cb8a7bf37ade190d6082268d8a0fe4a992a
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_103_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26b6c89df6123cd71022a7ba20f16f5e25898906c1dfa7a81f412495f2f3070c
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_104_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d80ef6e7e2f90a091aa40350189d057efa3acdc9c6bf8557829164f409172358
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_105_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bef9a9d047faaac418831cd8a9cefb6185a4f43ae863304cd54f796188b0a14
3
+ size 131677677
2b855b55bc4/global_step52452/bf16_zero_pp_rank_106_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7de6b2f1d197c4a39536283df33b2bedaa87cdc6b19f7dc9ada808b27b68e80b
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_107_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28e90a1ebe25a038b6baacf865d34fb6849031c0307e11d991ba78d5a21d03ab
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_108_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:147655c0529a2c37212ed0039c24c1439fc6ba4ac0404f0a3236a92f63da7e2a
3
+ size 131677869
2b855b55bc4/global_step52452/bf16_zero_pp_rank_109_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed6443a88a75db6174d69cdc0bba56de01ca02f8bf2ee2fcc5515fcc605be368
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fd74f4287e6c48e8eb7fb1dd650d761dd10e7b14eaac566e8dfd45eab83ff00
3
+ size 131677794
2b855b55bc4/global_step52452/bf16_zero_pp_rank_110_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0121d8867148adb205cf26baadc27e20096e837fae6e64ab288842f4d8384c7
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_111_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d458cfd0017d283e453013abfdd9011025b25961ad4c6303b27ab9f2acd78a63
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_112_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f396570291f16242ee1abcb166fb6f1ed15915e55b1d89bd48d45bdd61d8dae
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_113_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae6e7a0d9467c4593b82fb4e578baadffce84c2b5b7ec63dc6c98b992b2612e6
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_114_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afca160365fdce5370fbe40b2ef4dbc68f52e3e2f61ea9a410a8aad353d9346f
3
+ size 131677677
2b855b55bc4/global_step52452/bf16_zero_pp_rank_115_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:624601c13baf354c7e23b66a15b131fee0a9e5f751d90e3b32926a295cfe88e6
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_116_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab6c0c9c19d4c348fcd8d084c45310b2a0bb9206d32365ccbd1829e61e4450d8
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_117_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f5b36f215b3f8981c79527bb2706bc3503465d9477c8bcf415ff159ec9409e1
3
+ size 131677677
2b855b55bc4/global_step52452/bf16_zero_pp_rank_118_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1476f4057acc3d42aa618a4fe4e9d9cefc46f154b2902da9a4447e0ee65c9ea
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_119_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1c27ff3d40d98a134f94fb411ca49a8c2ff42765bd3f7e2c23b1c94a7a32dd4
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5545c184838c2d4dd6a1621e370ea3ea6fe55dc88d124212babcca4e897530e
3
+ size 131677730
2b855b55bc4/global_step52452/bf16_zero_pp_rank_120_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc5301bc12510a88381d37f9826df93ba495905620821f49cf06a6094224fc70
3
+ size 131677677
2b855b55bc4/global_step52452/bf16_zero_pp_rank_121_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bddbd274aa0bcdd4ab2b5f898f86b57be824e9d362545724249084329c70966a
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_122_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa4a355553a8bd26c930542a19239621330609fe0c73f67b6fe626b241dd5fc8
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_123_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfe4e2348f7c482db041641be5b38d5e718cd0dd0fe9498b502831af85b8bd02
3
+ size 131677869
2b855b55bc4/global_step52452/bf16_zero_pp_rank_124_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08ffd5a9a6c1ac9741db247e0e6f63a0ec90683d3b82afb053430353c43d94a7
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_125_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a613bacd59f87cdae87bd23b84fdb57f5a71e544336f942b6450242da9becda
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_126_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ea73a7fb74c469de7fa2784007b891a1f4790c8b6272d8390ebf2341a12c2d0
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_127_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:346577ecdff05cf304104966558bc389b32b84b7f0a335515bc297fde510d00d
3
+ size 131677677
2b855b55bc4/global_step52452/bf16_zero_pp_rank_128_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eab76fcf1eecaebb0fe2cc0b4a693e43881c4542cac18cddc3e78d9d5e069d51
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_129_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bd7c3b658dc862b24d166a8b3320fa2be1bb215172567edac2e02b9bb85ea38
3
+ size 131677677
2b855b55bc4/global_step52452/bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e384fbaffe3ab864a62916568060b9254f0462ef1efe843a4f55b98b72874ca5
3
+ size 131677730
2b855b55bc4/global_step52452/bf16_zero_pp_rank_130_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7be774c032701feba3530277d2bf386e67c12849692d27c0a699c9fee95884a4
3
+ size 131677869
2b855b55bc4/global_step52452/bf16_zero_pp_rank_131_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b7900a0c06848b3b5ab55985a6d4835023846b9dc0a6f566bedeb9a01c93d59
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_132_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7355d772023964fb364ea6ac33ab1b224cf7ad40f1603c8bc3c12dc1d386f583
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_133_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3005451dba6d3ba3dee2ed457ecf0742d58f21df7d0e3db15fe8208a9ca95272
3
+ size 131677677
2b855b55bc4/global_step52452/bf16_zero_pp_rank_134_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f07068654a7f9e449d53fb02339b550ab0ff761d52f6139837ec40238dff465b
3
+ size 131677741
2b855b55bc4/global_step52452/bf16_zero_pp_rank_135_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d714b97e6edc45f957667729a03a7328fb3b49aef86ef3147a7db47008d125e8
3
+ size 131677805
2b855b55bc4/global_step52452/bf16_zero_pp_rank_136_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddc0f84014d7e379037ad26626b824ee4461b4ba7e7b13b57da9a9411caab118
3
+ size 131677741