diff --git a/checkpoint-100/README.md b/checkpoint-100/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-100/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-100/adapter_config.json b/checkpoint-100/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-100/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-100/adapter_model.safetensors b/checkpoint-100/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0177af077b4d9b3df3fd65935f3163e725cf4f56 --- /dev/null +++ b/checkpoint-100/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7eda84bbe6be924cf996a444e72570b8d95cbdd618390d2be1685707b2e46d1 +size 340225224 diff --git a/checkpoint-100/optimizer.pt b/checkpoint-100/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..a913b2897e4b5f7d9cb1c4d6ad17972b50a4e481 --- /dev/null +++ b/checkpoint-100/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c2ffe3a81285945457f55b2444f10bc16e192d00c648a6de881648f0790214a +size 170950631 diff --git a/checkpoint-100/rng_state.pth b/checkpoint-100/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..8ab4ba5654447a9c0d2dfe9389cabcf3629d5bbd --- /dev/null +++ b/checkpoint-100/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27518dfe9949a7520c7abb821434d1758c6d5d5309854c90562fe8b87079b805 +size 14575 diff --git a/checkpoint-100/scheduler.pt b/checkpoint-100/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f609720b93637ecb76b3ff6404e42443f525242f --- /dev/null +++ b/checkpoint-100/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4c4c7301447dd5b860b54ad1354b925268062eb32249c904ad26e49337b80cb +size 627 diff --git a/checkpoint-100/trainer_state.json b/checkpoint-100/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..fbf046b07d20d4843dff00fb1f10679e97bd0fa7 --- /dev/null +++ b/checkpoint-100/trainer_state.json @@ -0,0 +1,75 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.5384615384615383, + "eval_steps": 25, + "global_step": 100, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 552629855846400.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-100/training_args.bin b/checkpoint-100/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-100/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-125/README.md b/checkpoint-125/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-125/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-125/adapter_config.json b/checkpoint-125/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-125/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-125/adapter_model.safetensors b/checkpoint-125/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c2c97e25ce55acd27237b266dcba6dd542863de0 --- /dev/null +++ b/checkpoint-125/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3556a4604c9cb47737743f441998cea7efbe9735b897f861120b939788df3ee +size 340225224 diff --git a/checkpoint-125/optimizer.pt b/checkpoint-125/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f4af2b92a9e0786e426cf000d7530da4dba72bfb --- /dev/null +++ b/checkpoint-125/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc8ee28a7c9eb0f1bd54e7015acc1645ea07f4d69cec97fcf6897af7f504cc41 +size 170950631 diff --git a/checkpoint-125/rng_state.pth b/checkpoint-125/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..d40bbf666a04f5fae6cb1140720d509a1551f040 --- /dev/null +++ b/checkpoint-125/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d428637a83ded27766037fec61c5f735ba258814ee89c028f3ba7c1ae398d48 +size 14575 diff --git a/checkpoint-125/scheduler.pt b/checkpoint-125/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..5ffacb9fb4ab42f3eb4c04c815cb9ecf139dd78c --- /dev/null +++ b/checkpoint-125/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06608bebd3a3ccb8d93ed73ecb3c8cc7445e126a6ac14352ea8ab88ea4bb5a06 +size 627 diff --git a/checkpoint-125/trainer_state.json b/checkpoint-125/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..7cfde73f661176450fae107427f0838efa56f517 --- /dev/null +++ b/checkpoint-125/trainer_state.json @@ -0,0 +1,89 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9230769230769231, + "eval_steps": 25, + "global_step": 125, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 690787319808000.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-125/training_args.bin b/checkpoint-125/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-125/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-150/README.md b/checkpoint-150/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-150/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-150/adapter_config.json b/checkpoint-150/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-150/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-150/adapter_model.safetensors b/checkpoint-150/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4a4397ab6cbaddade049ae74e356042469d16a84 --- /dev/null +++ b/checkpoint-150/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41f31aa13964a43d5e7ec116514a7b7a33bd2d0fb6bca0c34271761aa8cb495f +size 340225224 diff --git a/checkpoint-150/optimizer.pt b/checkpoint-150/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f36bcd58f33013eeaad6b2a6776a629a2c0c62b3 --- /dev/null +++ b/checkpoint-150/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb77702bb1f6dead385e7e89549b8cd1e002c2d63ae7f37468d8264d2e35c7df +size 170950631 diff --git a/checkpoint-150/rng_state.pth b/checkpoint-150/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..017e591e58f64b21e8bf31dbdf490c701ded2637 --- /dev/null +++ b/checkpoint-150/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ccc00f6b6529ee716cc744199b3013674203fff3fa19586e073dd0e4a374b2b +size 14575 diff --git a/checkpoint-150/scheduler.pt b/checkpoint-150/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..910400a9c3edb224317468b4a45ddf44e5d1bb2b --- /dev/null +++ b/checkpoint-150/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a160737956e1db74337f9aa888786aae9df82ea584fae309f088519ea06eec32 +size 627 diff --git a/checkpoint-150/trainer_state.json b/checkpoint-150/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e04a01aeb7b906475c03c7f775621a343826f4a1 --- /dev/null +++ b/checkpoint-150/trainer_state.json @@ -0,0 +1,103 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.3076923076923075, + "eval_steps": 25, + "global_step": 150, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + }, + { + "epoch": 2.31, + "learning_rate": 1.254180602006689e-05, + "loss": 0.4767, + "step": 150 + }, + { + "epoch": 2.31, + "eval_loss": 1.5637826919555664, + "eval_runtime": 4.4377, + "eval_samples_per_second": 4.507, + "eval_steps_per_second": 0.676, + "step": 150 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 828944783769600.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-150/training_args.bin b/checkpoint-150/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-150/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-175/README.md b/checkpoint-175/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-175/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-175/adapter_config.json b/checkpoint-175/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-175/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-175/adapter_model.safetensors b/checkpoint-175/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d10f60c1a1c8833e4da5b509a23d34ea863c5d8d --- /dev/null +++ b/checkpoint-175/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41cb7f8e503af4eb85810558537ff83e2fdda2d31a9b9708c2975002661b4b9b +size 340225224 diff --git a/checkpoint-175/optimizer.pt b/checkpoint-175/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e4100cf62b38ac8d8e63d65ff8cd1600ee7cf746 --- /dev/null +++ b/checkpoint-175/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab9719f301cf967bc01033d62351cb5aa02e3162efcd88a6539bac289eca0b27 +size 170950631 diff --git a/checkpoint-175/rng_state.pth b/checkpoint-175/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..50f1fead1f99b2af6c303cb651b58f245bfa006e --- /dev/null +++ b/checkpoint-175/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64d800d4a6a21fa2708a25d1876d0056ca6fbf4cab4e2ce4bd537079c4b59160 +size 14575 diff --git a/checkpoint-175/scheduler.pt b/checkpoint-175/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..96696ed6577db3b02c548c3d5f65b96ae59943db --- /dev/null +++ b/checkpoint-175/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:115c67b78db4e6d66498d92ba3b6ba66ac66a066e3961e34490608a562322e6d +size 627 diff --git a/checkpoint-175/trainer_state.json b/checkpoint-175/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d7a56568a70cdc263a7dc7adf3a9501707d12588 --- /dev/null +++ b/checkpoint-175/trainer_state.json @@ -0,0 +1,117 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.6923076923076925, + "eval_steps": 25, + "global_step": 175, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + }, + { + "epoch": 2.31, + "learning_rate": 1.254180602006689e-05, + "loss": 0.4767, + "step": 150 + }, + { + "epoch": 2.31, + "eval_loss": 1.5637826919555664, + "eval_runtime": 4.4377, + "eval_samples_per_second": 4.507, + "eval_steps_per_second": 0.676, + "step": 150 + }, + { + "epoch": 2.69, + "learning_rate": 1.045150501672241e-05, + "loss": 0.4167, + "step": 175 + }, + { + "epoch": 2.69, + "eval_loss": 1.6361162662506104, + "eval_runtime": 4.4339, + "eval_samples_per_second": 4.511, + "eval_steps_per_second": 0.677, + "step": 175 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 967102247731200.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-175/training_args.bin b/checkpoint-175/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-175/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-200/README.md b/checkpoint-200/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-200/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-200/adapter_config.json b/checkpoint-200/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-200/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-200/adapter_model.safetensors b/checkpoint-200/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9f6589aadbb54e4888eafbe97ac655a96300149b --- /dev/null +++ b/checkpoint-200/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97e108c2abbd378c03de9eb36ae9836382f6e2b309d1d1fc02369856534b31bb +size 340225224 diff --git a/checkpoint-200/optimizer.pt b/checkpoint-200/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..27dc480916c8cbfff83a391aa38ef4d94a85eb56 --- /dev/null +++ b/checkpoint-200/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b14e7a8a39630161ef250739c283ff8d3d85f784929cbe714742fcaa8f9171ff +size 170950631 diff --git a/checkpoint-200/rng_state.pth b/checkpoint-200/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..914a2e9ce6d3777a4eb33a8be7416284fc840bd3 --- /dev/null +++ b/checkpoint-200/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7a8897479d1dbd9a672321d73fe2a3d09370584f791ba1329e0e35cf353353f +size 14575 diff --git a/checkpoint-200/scheduler.pt b/checkpoint-200/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f9f65371c29c1bed0aa4821c98d781cbb247f77f --- /dev/null +++ b/checkpoint-200/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef9808a7b4bea21800e28f810a599c2cdcd987806b6f84e2c304ad57f3fa2c20 +size 627 diff --git a/checkpoint-200/trainer_state.json b/checkpoint-200/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9f4020a272aed396c8fa511d121f462610c4eff5 --- /dev/null +++ b/checkpoint-200/trainer_state.json @@ -0,0 +1,131 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.076923076923077, + "eval_steps": 25, + "global_step": 200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + }, + { + "epoch": 2.31, + "learning_rate": 1.254180602006689e-05, + "loss": 0.4767, + "step": 150 + }, + { + "epoch": 2.31, + "eval_loss": 1.5637826919555664, + "eval_runtime": 4.4377, + "eval_samples_per_second": 4.507, + "eval_steps_per_second": 0.676, + "step": 150 + }, + { + "epoch": 2.69, + "learning_rate": 1.045150501672241e-05, + "loss": 0.4167, + "step": 175 + }, + { + "epoch": 2.69, + "eval_loss": 1.6361162662506104, + "eval_runtime": 4.4339, + "eval_samples_per_second": 4.511, + "eval_steps_per_second": 0.677, + "step": 175 + }, + { + "epoch": 3.08, + "learning_rate": 8.361204013377926e-06, + "loss": 0.8085, + "step": 200 + }, + { + "epoch": 3.08, + "eval_loss": 1.869662880897522, + "eval_runtime": 4.4356, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 200 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 1105259711692800.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-200/training_args.bin b/checkpoint-200/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-200/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-225/README.md b/checkpoint-225/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-225/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-225/adapter_config.json b/checkpoint-225/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-225/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-225/adapter_model.safetensors b/checkpoint-225/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3a90d5805187407ecd79252516bd9bfe5d58d216 --- /dev/null +++ b/checkpoint-225/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6802c280029ff5b4f806016938fd9683c3548f5d0f921bb837f530d464766aa6 +size 340225224 diff --git a/checkpoint-225/optimizer.pt b/checkpoint-225/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..52db34e277e59b67d91efc2e98079c32e617d64e --- /dev/null +++ b/checkpoint-225/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc5d8552637b7719ece05dfbf9e12e6bea2f426bf34c30962e9c0334c34afffb +size 170950631 diff --git a/checkpoint-225/rng_state.pth b/checkpoint-225/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..ff21ef8c2b0de3520f93bfb8f9be9dbcabd745fe --- /dev/null +++ b/checkpoint-225/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c8fb3a8ddfd6d723e4dd3fb6915ac5ed5ae95a67fc27b22b9e70a0754e6b0ba +size 14575 diff --git a/checkpoint-225/scheduler.pt b/checkpoint-225/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..682090b13610062925660411f25361ed67d01f86 --- /dev/null +++ b/checkpoint-225/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02b77ed6e1b9aa49a03aa42fe79eddbeedfe61436dcd4ff94410eba42305fff6 +size 627 diff --git a/checkpoint-225/trainer_state.json b/checkpoint-225/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..adcf77a4978165be79dafcb38d33a0c8e8e4df71 --- /dev/null +++ b/checkpoint-225/trainer_state.json @@ -0,0 +1,145 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.4615384615384617, + "eval_steps": 25, + "global_step": 225, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + }, + { + "epoch": 2.31, + "learning_rate": 1.254180602006689e-05, + "loss": 0.4767, + "step": 150 + }, + { + "epoch": 2.31, + "eval_loss": 1.5637826919555664, + "eval_runtime": 4.4377, + "eval_samples_per_second": 4.507, + "eval_steps_per_second": 0.676, + "step": 150 + }, + { + "epoch": 2.69, + "learning_rate": 1.045150501672241e-05, + "loss": 0.4167, + "step": 175 + }, + { + "epoch": 2.69, + "eval_loss": 1.6361162662506104, + "eval_runtime": 4.4339, + "eval_samples_per_second": 4.511, + "eval_steps_per_second": 0.677, + "step": 175 + }, + { + "epoch": 3.08, + "learning_rate": 8.361204013377926e-06, + "loss": 0.8085, + "step": 200 + }, + { + "epoch": 3.08, + "eval_loss": 1.869662880897522, + "eval_runtime": 4.4356, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 200 + }, + { + "epoch": 3.46, + "learning_rate": 6.270903010033445e-06, + "loss": 1.1709, + "step": 225 + }, + { + "epoch": 3.46, + "eval_loss": 2.216012954711914, + "eval_runtime": 4.4388, + "eval_samples_per_second": 4.506, + "eval_steps_per_second": 0.676, + "step": 225 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 1243417175654400.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-225/training_args.bin b/checkpoint-225/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-225/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-25/README.md b/checkpoint-25/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-25/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-25/adapter_config.json b/checkpoint-25/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-25/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-25/adapter_model.safetensors b/checkpoint-25/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8242438a7f5b2a53d08c579b0fbb5b6aaaf3b8ae --- /dev/null +++ b/checkpoint-25/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e92161a00bfe2de2da02e8c674198e38ca099a2e3b802c96b77125e3dc65a668 +size 340225224 diff --git a/checkpoint-25/optimizer.pt b/checkpoint-25/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f512645def76874f687c3e746b4a68c254e21639 --- /dev/null +++ b/checkpoint-25/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a851d63d2e9f1418ba2ab16476bfedcb5c6f333501a9e883411364a386c6989c +size 170950631 diff --git a/checkpoint-25/rng_state.pth b/checkpoint-25/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..be190c5bc6d5b29c136bdaf813c9b4aec862f943 --- /dev/null +++ b/checkpoint-25/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4909159544ef64a34bf97aa2f7d267838a79561785919e397bb9ec41713229ad +size 14575 diff --git a/checkpoint-25/scheduler.pt b/checkpoint-25/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..290a66b62c657e05fff9f5430690fa4ccd0a6673 --- /dev/null +++ b/checkpoint-25/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33624aa46adb1713e06028397961e7bbe5ff1dbddd6567d3734e3c5864c42aaa +size 627 diff --git a/checkpoint-25/trainer_state.json b/checkpoint-25/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e79f0c189ccecc827b1a6324044d187d1b5268b2 --- /dev/null +++ b/checkpoint-25/trainer_state.json @@ -0,0 +1,33 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.38461538461538464, + "eval_steps": 25, + "global_step": 25, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 138157463961600.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-25/training_args.bin b/checkpoint-25/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-25/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-250/README.md b/checkpoint-250/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-250/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-250/adapter_config.json b/checkpoint-250/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-250/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-250/adapter_model.safetensors b/checkpoint-250/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b3a3bdd17c16a9fbe69be33331858224f947d284 --- /dev/null +++ b/checkpoint-250/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43cf9b3ff101c03e53318ea6780786fe24763a0db05aeb3ed1e618e4ab583f4d +size 340225224 diff --git a/checkpoint-250/optimizer.pt b/checkpoint-250/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..6fec8bbbef7c953151d268ebbaf858da7250f235 --- /dev/null +++ b/checkpoint-250/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5eeec84428c08ff1819446603a071b463447d9cb3eaafed11b38c44b0004ce1e +size 170950631 diff --git a/checkpoint-250/rng_state.pth b/checkpoint-250/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..617e208aeb4f5079bda16f759cc56786aaf5c8c6 --- /dev/null +++ b/checkpoint-250/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca3a5a75df42087b58b7553787abb245c54211b77304b11cfa5910f8f17f45d6 +size 14575 diff --git a/checkpoint-250/scheduler.pt b/checkpoint-250/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ffdbd2e3e5791fe27705fa384fdc8890aeb1fc68 --- /dev/null +++ b/checkpoint-250/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77e0eb3c14a5daa8cc5619f198ddfd3729ca52b6b3ee11adcd795b75f4bc93cc +size 627 diff --git a/checkpoint-250/trainer_state.json b/checkpoint-250/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..df9da0e4ad4b4202ca7ff695c8e1191ba5737df3 --- /dev/null +++ b/checkpoint-250/trainer_state.json @@ -0,0 +1,159 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.8461538461538463, + "eval_steps": 25, + "global_step": 250, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + }, + { + "epoch": 2.31, + "learning_rate": 1.254180602006689e-05, + "loss": 0.4767, + "step": 150 + }, + { + "epoch": 2.31, + "eval_loss": 1.5637826919555664, + "eval_runtime": 4.4377, + "eval_samples_per_second": 4.507, + "eval_steps_per_second": 0.676, + "step": 150 + }, + { + "epoch": 2.69, + "learning_rate": 1.045150501672241e-05, + "loss": 0.4167, + "step": 175 + }, + { + "epoch": 2.69, + "eval_loss": 1.6361162662506104, + "eval_runtime": 4.4339, + "eval_samples_per_second": 4.511, + "eval_steps_per_second": 0.677, + "step": 175 + }, + { + "epoch": 3.08, + "learning_rate": 8.361204013377926e-06, + "loss": 0.8085, + "step": 200 + }, + { + "epoch": 3.08, + "eval_loss": 1.869662880897522, + "eval_runtime": 4.4356, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 200 + }, + { + "epoch": 3.46, + "learning_rate": 6.270903010033445e-06, + "loss": 1.1709, + "step": 225 + }, + { + "epoch": 3.46, + "eval_loss": 2.216012954711914, + "eval_runtime": 4.4388, + "eval_samples_per_second": 4.506, + "eval_steps_per_second": 0.676, + "step": 225 + }, + { + "epoch": 3.85, + "learning_rate": 4.180602006688963e-06, + "loss": 1.4037, + "step": 250 + }, + { + "epoch": 3.85, + "eval_loss": 2.300891399383545, + "eval_runtime": 4.436, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 250 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 1381574639616000.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-250/training_args.bin b/checkpoint-250/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-250/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-275/README.md b/checkpoint-275/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-275/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-275/adapter_config.json b/checkpoint-275/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-275/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-275/adapter_model.safetensors b/checkpoint-275/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4c13a2715d3a760fa797eec45e0bf10708399409 --- /dev/null +++ b/checkpoint-275/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c7a15355b300256adf9e592ec5a5b05106ff623d3363851ef4694ef1358dd0d +size 340225224 diff --git a/checkpoint-275/optimizer.pt b/checkpoint-275/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4e939def17e8a08d9501b9f53a6c340a66f54bb9 --- /dev/null +++ b/checkpoint-275/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42c1177cb54f0c39ee4e85c75fdd4c0f14bc379694f4f3f497f2e9c5264b7ed7 +size 170951079 diff --git a/checkpoint-275/rng_state.pth b/checkpoint-275/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..1749af7b5066f420cccde363e8611f10d6b1bb1a --- /dev/null +++ b/checkpoint-275/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a936e97df371d92e527960290e42770e2f94b6b3e9cfb7aa088b9f9a27d6375 +size 14575 diff --git a/checkpoint-275/scheduler.pt b/checkpoint-275/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..374cd26436bfe9aec07d7968b2ba8098f5e3f464 --- /dev/null +++ b/checkpoint-275/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cc2df9dffdf8513a391211f95b0fb3a1d5a329398c0f06fb99c00dc39b09a13 +size 627 diff --git a/checkpoint-275/trainer_state.json b/checkpoint-275/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..81370baeb689a007f0580521a8966556dd15427f --- /dev/null +++ b/checkpoint-275/trainer_state.json @@ -0,0 +1,173 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.230769230769231, + "eval_steps": 25, + "global_step": 275, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + }, + { + "epoch": 2.31, + "learning_rate": 1.254180602006689e-05, + "loss": 0.4767, + "step": 150 + }, + { + "epoch": 2.31, + "eval_loss": 1.5637826919555664, + "eval_runtime": 4.4377, + "eval_samples_per_second": 4.507, + "eval_steps_per_second": 0.676, + "step": 150 + }, + { + "epoch": 2.69, + "learning_rate": 1.045150501672241e-05, + "loss": 0.4167, + "step": 175 + }, + { + "epoch": 2.69, + "eval_loss": 1.6361162662506104, + "eval_runtime": 4.4339, + "eval_samples_per_second": 4.511, + "eval_steps_per_second": 0.677, + "step": 175 + }, + { + "epoch": 3.08, + "learning_rate": 8.361204013377926e-06, + "loss": 0.8085, + "step": 200 + }, + { + "epoch": 3.08, + "eval_loss": 1.869662880897522, + "eval_runtime": 4.4356, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 200 + }, + { + "epoch": 3.46, + "learning_rate": 6.270903010033445e-06, + "loss": 1.1709, + "step": 225 + }, + { + "epoch": 3.46, + "eval_loss": 2.216012954711914, + "eval_runtime": 4.4388, + "eval_samples_per_second": 4.506, + "eval_steps_per_second": 0.676, + "step": 225 + }, + { + "epoch": 3.85, + "learning_rate": 4.180602006688963e-06, + "loss": 1.4037, + "step": 250 + }, + { + "epoch": 3.85, + "eval_loss": 2.300891399383545, + "eval_runtime": 4.436, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 250 + }, + { + "epoch": 4.23, + "learning_rate": 2.0903010033444816e-06, + "loss": 1.5494, + "step": 275 + }, + { + "epoch": 4.23, + "eval_loss": 2.494781494140625, + "eval_runtime": 4.4383, + "eval_samples_per_second": 4.506, + "eval_steps_per_second": 0.676, + "step": 275 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 1519732103577600.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-275/training_args.bin b/checkpoint-275/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-275/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-300/README.md b/checkpoint-300/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-300/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-300/adapter_config.json b/checkpoint-300/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-300/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-300/adapter_model.safetensors b/checkpoint-300/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..88681288291d508a9ea7c0a6367def07989e55f2 --- /dev/null +++ b/checkpoint-300/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e9eda395952b68af7f86a47f1eab64fa9f9f7d3f256cf15d7231373ed4c6aa7 +size 340225224 diff --git a/checkpoint-300/optimizer.pt b/checkpoint-300/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4bd49dbefdb8e10f4a54f234bb52fc81f895252b --- /dev/null +++ b/checkpoint-300/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:062c424feb59874705870921b46728cb3f4447afb9dff143f38b7f48058c805c +size 170951079 diff --git a/checkpoint-300/rng_state.pth b/checkpoint-300/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..de72f9294eaac87f1500a34745906769980f48ac --- /dev/null +++ b/checkpoint-300/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae1e3c97e826ec188f2d3622b31f1e50fe96d5050e1d89188d99992ae505fe49 +size 14575 diff --git a/checkpoint-300/scheduler.pt b/checkpoint-300/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..b90f497d1964ab533a00cf1f82ed5ac177a9ab76 --- /dev/null +++ b/checkpoint-300/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780f4b688518fbdf7f8cbc3eb69c0632d3e2760d34259b259badb69c9b2efe24 +size 627 diff --git a/checkpoint-300/trainer_state.json b/checkpoint-300/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..fc4ba232c8041323d35a1b307f5f1d9875ebd908 --- /dev/null +++ b/checkpoint-300/trainer_state.json @@ -0,0 +1,187 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.615384615384615, + "eval_steps": 25, + "global_step": 300, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + }, + { + "epoch": 1.54, + "learning_rate": 1.6722408026755853e-05, + "loss": 0.6915, + "step": 100 + }, + { + "epoch": 1.54, + "eval_loss": 1.3842780590057373, + "eval_runtime": 4.44, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 100 + }, + { + "epoch": 1.92, + "learning_rate": 1.4632107023411373e-05, + "loss": 0.706, + "step": 125 + }, + { + "epoch": 1.92, + "eval_loss": 1.4492868185043335, + "eval_runtime": 4.4412, + "eval_samples_per_second": 4.503, + "eval_steps_per_second": 0.675, + "step": 125 + }, + { + "epoch": 2.31, + "learning_rate": 1.254180602006689e-05, + "loss": 0.4767, + "step": 150 + }, + { + "epoch": 2.31, + "eval_loss": 1.5637826919555664, + "eval_runtime": 4.4377, + "eval_samples_per_second": 4.507, + "eval_steps_per_second": 0.676, + "step": 150 + }, + { + "epoch": 2.69, + "learning_rate": 1.045150501672241e-05, + "loss": 0.4167, + "step": 175 + }, + { + "epoch": 2.69, + "eval_loss": 1.6361162662506104, + "eval_runtime": 4.4339, + "eval_samples_per_second": 4.511, + "eval_steps_per_second": 0.677, + "step": 175 + }, + { + "epoch": 3.08, + "learning_rate": 8.361204013377926e-06, + "loss": 0.8085, + "step": 200 + }, + { + "epoch": 3.08, + "eval_loss": 1.869662880897522, + "eval_runtime": 4.4356, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 200 + }, + { + "epoch": 3.46, + "learning_rate": 6.270903010033445e-06, + "loss": 1.1709, + "step": 225 + }, + { + "epoch": 3.46, + "eval_loss": 2.216012954711914, + "eval_runtime": 4.4388, + "eval_samples_per_second": 4.506, + "eval_steps_per_second": 0.676, + "step": 225 + }, + { + "epoch": 3.85, + "learning_rate": 4.180602006688963e-06, + "loss": 1.4037, + "step": 250 + }, + { + "epoch": 3.85, + "eval_loss": 2.300891399383545, + "eval_runtime": 4.436, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 250 + }, + { + "epoch": 4.23, + "learning_rate": 2.0903010033444816e-06, + "loss": 1.5494, + "step": 275 + }, + { + "epoch": 4.23, + "eval_loss": 2.494781494140625, + "eval_runtime": 4.4383, + "eval_samples_per_second": 4.506, + "eval_steps_per_second": 0.676, + "step": 275 + }, + { + "epoch": 4.62, + "learning_rate": 0.0, + "loss": 1.4895, + "step": 300 + }, + { + "epoch": 4.62, + "eval_loss": 2.576362133026123, + "eval_runtime": 4.4364, + "eval_samples_per_second": 4.508, + "eval_steps_per_second": 0.676, + "step": 300 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 1657889567539200.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-300/training_args.bin b/checkpoint-300/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-300/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-50/README.md b/checkpoint-50/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-50/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-50/adapter_config.json b/checkpoint-50/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-50/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-50/adapter_model.safetensors b/checkpoint-50/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f75d2af6900a0429df30de84bd26ddb18173ced7 --- /dev/null +++ b/checkpoint-50/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85f059c9e732071d1ef44dbc8240c4b42a0634ac24f1e2aec457a2e69c836aa8 +size 340225224 diff --git a/checkpoint-50/optimizer.pt b/checkpoint-50/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..86bb87e2c0fe0734d8306fd602204c6921ba9dfc --- /dev/null +++ b/checkpoint-50/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7930c2b70930a408673deb032e6abe037bdc697045753ebd5217b27200e65f5f +size 170950631 diff --git a/checkpoint-50/rng_state.pth b/checkpoint-50/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..1f8ce5e7fd42995690b284ca224f4372c87d7a02 --- /dev/null +++ b/checkpoint-50/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a50f642f6f99c3f59c9e322dda23642879e9618c4c0d0ca297f0913b3073741 +size 14575 diff --git a/checkpoint-50/scheduler.pt b/checkpoint-50/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..3c635f5c5adecdb1ac426e21010b6c4c38b6a25f --- /dev/null +++ b/checkpoint-50/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55a58a131dcbcb304da7d0b32934fccc762e2f989c2e7533853ffe8295eb908c +size 627 diff --git a/checkpoint-50/trainer_state.json b/checkpoint-50/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9be285166fc23e9e900f8f93d9bdfd66349d22ad --- /dev/null +++ b/checkpoint-50/trainer_state.json @@ -0,0 +1,47 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.7692307692307693, + "eval_steps": 25, + "global_step": 50, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 276314927923200.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-50/training_args.bin b/checkpoint-50/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-50/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155 diff --git a/checkpoint-75/README.md b/checkpoint-75/README.md new file mode 100644 index 0000000000000000000000000000000000000000..89f4a1813feacaaf110bfa5244b7ca80556ea4d4 --- /dev/null +++ b/checkpoint-75/README.md @@ -0,0 +1,219 @@ +--- +library_name: peft +base_model: mistralai/Mistral-7B-v0.1 +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + +## Training procedure + + +The following `bitsandbytes` quantization config was used during training: +- quant_method: bitsandbytes +- load_in_8bit: False +- load_in_4bit: True +- llm_int8_threshold: 6.0 +- llm_int8_skip_modules: None +- llm_int8_enable_fp32_cpu_offload: False +- llm_int8_has_fp16_weight: False +- bnb_4bit_quant_type: nf4 +- bnb_4bit_use_double_quant: True +- bnb_4bit_compute_dtype: bfloat16 + +### Framework versions + + +- PEFT 0.6.0 diff --git a/checkpoint-75/adapter_config.json b/checkpoint-75/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7f5806b2c588a0859f7da5fb9a851fdc3c6f3f49 --- /dev/null +++ b/checkpoint-75/adapter_config.json @@ -0,0 +1,29 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "mistralai/Mistral-7B-v0.1", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "lora_alpha": 64, + "lora_dropout": 0.05, + "modules_to_save": null, + "peft_type": "LORA", + "r": 32, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "down_proj", + "up_proj", + "lm_head", + "o_proj", + "v_proj", + "k_proj", + "gate_proj" + ], + "task_type": "CAUSAL_LM" +} \ No newline at end of file diff --git a/checkpoint-75/adapter_model.safetensors b/checkpoint-75/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f5e4a2a00b6dcc2a58a75bcba6133be1d053207 --- /dev/null +++ b/checkpoint-75/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bf699752d8c41351955d373e44ba8567eab4be4f6fa4d3a0509b9e7615bf279 +size 340225224 diff --git a/checkpoint-75/optimizer.pt b/checkpoint-75/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..4711f7a241c10a85ca04677ecff1815927298160 --- /dev/null +++ b/checkpoint-75/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bf758418d9479b1ceb40c61b84fe1622361d7b577ad5998198052a99d55946c +size 170950631 diff --git a/checkpoint-75/rng_state.pth b/checkpoint-75/rng_state.pth new file mode 100644 index 0000000000000000000000000000000000000000..a47465f4e8705321d4f383d2b2328cb610e41820 --- /dev/null +++ b/checkpoint-75/rng_state.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebcbb02e034024a606097bc54463ddca6cbd1b95db06baf28639bc24acb56d6f +size 14575 diff --git a/checkpoint-75/scheduler.pt b/checkpoint-75/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..4de962e7b6d8301d1c5a3beded574cbf950ea418 --- /dev/null +++ b/checkpoint-75/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b59e43443a01ea6c78d5d075c1b9bfdd03eb098b7fcb3702e30f72a83450546 +size 627 diff --git a/checkpoint-75/trainer_state.json b/checkpoint-75/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..50e80d4a49d14d625f2f9df19e67f568e01b42d5 --- /dev/null +++ b/checkpoint-75/trainer_state.json @@ -0,0 +1,61 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.1538461538461537, + "eval_steps": 25, + "global_step": 75, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.38, + "learning_rate": 2.29933110367893e-05, + "loss": 1.9196, + "step": 25 + }, + { + "epoch": 0.38, + "eval_loss": 1.4835141897201538, + "eval_runtime": 4.4399, + "eval_samples_per_second": 4.505, + "eval_steps_per_second": 0.676, + "step": 25 + }, + { + "epoch": 0.77, + "learning_rate": 2.090301003344482e-05, + "loss": 0.9398, + "step": 50 + }, + { + "epoch": 0.77, + "eval_loss": 1.353461503982544, + "eval_runtime": 4.4353, + "eval_samples_per_second": 4.509, + "eval_steps_per_second": 0.676, + "step": 50 + }, + { + "epoch": 1.15, + "learning_rate": 1.8812709030100337e-05, + "loss": 0.7795, + "step": 75 + }, + { + "epoch": 1.15, + "eval_loss": 1.4102122783660889, + "eval_runtime": 4.4451, + "eval_samples_per_second": 4.499, + "eval_steps_per_second": 0.675, + "step": 75 + } + ], + "logging_steps": 25, + "max_steps": 300, + "num_train_epochs": 5, + "save_steps": 25, + "total_flos": 414472391884800.0, + "trial_name": null, + "trial_params": null +} diff --git a/checkpoint-75/training_args.bin b/checkpoint-75/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..6ee43fbd43335fa37cbd514edf5fe2c55ba9ae92 --- /dev/null +++ b/checkpoint-75/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e93936e735452d40dba2683877f022f826972f6b34855bfaa1e7eec4b5f8299 +size 4155