munish0838 commited on
Commit
d72ea8d
1 Parent(s): 0977b44

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +131 -0
README.md ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3
3
+ base_model: Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-200K-v0.1
4
+ tags:
5
+ - axolotl
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: Llama-3-8B-Magpie-Pro-SFT-200K-v0.1
9
+ results: []
10
+ library_name: transformers
11
+ pipeline_tag: text-generation
12
+ ---
13
+
14
+ # QuantFactory/Llama-3-8B-Magpie-Pro-SFT-200K-v0.1-GGUF
15
+ This is quantized version of [Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-200K-v0.1](https://huggingface.co/Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-200K-v0.1) created using llama.cpp
16
+
17
+ # Model Description
18
+
19
+ Project Web: [https://magpie-align.github.io/](https://magpie-align.github.io/)
20
+
21
+ Arxiv Technical Report: [https://arxiv.org/abs/2406.08464](https://arxiv.org/abs/2406.08464)
22
+
23
+ Codes: [https://github.com/magpie-align/magpie](https://github.com/magpie-align/magpie)
24
+
25
+ ## About This Model
26
+
27
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on **First 200K data** of [Magpie-Align/Magpie-Pro-300K-Filtered](https://huggingface.co/datasets/Magpie-Align/Magpie-Pro-300K-Filtered) dataset.
28
+
29
+ Please use [Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-v0.1](https://huggingface.co/Magpie-Align/Llama-3-8B-Magpie-Pro-SFT-v0.1) with better performance.
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 2e-05
37
+ - train_batch_size: 1
38
+ - eval_batch_size: 1
39
+ - seed: 42
40
+ - distributed_type: multi-GPU
41
+ - num_devices: 4
42
+ - gradient_accumulation_steps: 8
43
+ - total_train_batch_size: 32
44
+ - total_eval_batch_size: 4
45
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
+ - lr_scheduler_type: cosine
47
+ - lr_scheduler_warmup_steps: 100
48
+ - num_epochs: 2
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss |
53
+ |:-------------:|:------:|:----:|:---------------:|
54
+ | 0.8686 | 0.0018 | 1 | 0.8670 |
55
+ | 0.514 | 0.3342 | 184 | 0.5190 |
56
+ | 0.4769 | 0.6685 | 368 | 0.4684 |
57
+ | 0.4394 | 1.0027 | 552 | 0.4440 |
58
+ | 0.3399 | 1.3224 | 736 | 0.4436 |
59
+ | 0.3394 | 1.6567 | 920 | 0.4413 |
60
+
61
+
62
+ ### Framework versions
63
+
64
+ - Transformers 4.40.2
65
+ - Pytorch 2.3.0+cu121
66
+ - Datasets 2.19.1
67
+ - Tokenizers 0.19.1
68
+
69
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
70
+ <details><summary>See axolotl config</summary>
71
+
72
+ axolotl version: `0.4.0`
73
+ ```yaml
74
+ base_model: meta-llama/Meta-Llama-3-8B
75
+ model_type: LlamaForCausalLM
76
+ tokenizer_type: AutoTokenizer
77
+
78
+ load_in_8bit: false
79
+ load_in_4bit: false
80
+ strict: false
81
+
82
+ datasets:
83
+ - path: Magpie-Align/Magpie-Pro-300K-Filtered-First200K
84
+ type: sharegpt
85
+ conversation: llama3
86
+ dataset_prepared_path: last_run_prepared
87
+ val_set_size: 0.001
88
+ output_dir: ./out_Llama-3-8B-Magpie-Pro-200K-FilteredL
89
+
90
+ sequence_len: 8192
91
+ sample_packing: true
92
+ eval_sample_packing: false
93
+ pad_to_sequence_len: true
94
+
95
+ gradient_accumulation_steps: 8
96
+ micro_batch_size: 1
97
+ num_epochs: 2
98
+ optimizer: paged_adamw_8bit
99
+ lr_scheduler: cosine
100
+ learning_rate: 2e-5
101
+
102
+ train_on_inputs: false
103
+ group_by_length: false
104
+ bf16: auto
105
+ fp16:
106
+ tf32: false
107
+
108
+ gradient_checkpointing: true
109
+ gradient_checkpointing_kwargs:
110
+ use_reentrant: false
111
+ early_stopping_patience:
112
+ resume_from_checkpoint:
113
+ logging_steps: 1
114
+ xformers_attention:
115
+ flash_attention: true
116
+
117
+ warmup_steps: 100
118
+ evals_per_epoch: 3
119
+ eval_table_size:
120
+ saves_per_epoch: 1
121
+ debug:
122
+ deepspeed:
123
+ weight_decay: 0.0
124
+ fsdp:
125
+ fsdp_config:
126
+ special_tokens:
127
+ pad_token: <|end_of_text|>
128
+
129
+ ```
130
+
131
+ </details><br>