andysalerno commited on
Commit
3610162
1 Parent(s): d9d0141

End of training

Browse files
Files changed (3) hide show
  1. README.md +181 -0
  2. adapter_model.bin +3 -0
  3. adapter_model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - axolotl
6
+ - generated_from_trainer
7
+ base_model: andysalerno/mistral-sft-v3
8
+ model-index:
9
+ - name: rainbowfish-v7
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
17
+ <details><summary>See axolotl config</summary>
18
+
19
+ axolotl version: `0.4.0`
20
+ ```yaml
21
+ base_model: andysalerno/mistral-sft-v3
22
+ model_type: AutoModelForCausalLM
23
+
24
+ load_in_8bit: true
25
+ load_in_4bit: false
26
+ strict: false
27
+
28
+ datasets:
29
+ - path: andysalerno/rainbowfish-v1
30
+ type:
31
+ system_prompt: ""
32
+ field_system: system
33
+ field_instruction: input
34
+ field_output: output
35
+ format: "{instruction}"
36
+ no_input_format: "{instruction}"
37
+ dataset_prepared_path: last_run_prepared
38
+ val_set_size: 0.005
39
+ output_dir: ./lora-out-rainbow7
40
+
41
+ adapter: lora
42
+ lora_model_dir:
43
+
44
+ sequence_len: 2048
45
+ sample_packing: false # was true
46
+ eval_sample_packing: false
47
+ pad_to_sequence_len: false
48
+ padding_side: left
49
+
50
+ lora_r: 64
51
+ lora_alpha: 16
52
+ lora_dropout: 0.05
53
+ lora_target_linear: true
54
+ lora_fan_in_fan_out:
55
+ lora_target_modules:
56
+ - gate_proj
57
+ - down_proj
58
+ - up_proj
59
+ - q_proj
60
+ - v_proj
61
+ - k_proj
62
+ - o_proj
63
+
64
+ lora_modules_to_save:
65
+ - embed_tokens
66
+ - lm_head
67
+
68
+ wandb_project: axolotl
69
+ wandb_entity:
70
+ wandb_watch:
71
+ wandb_name:
72
+ wandb_log_model:
73
+
74
+ gradient_accumulation_steps: 4
75
+ micro_batch_size: 4
76
+ optimizer: paged_adamw_8bit
77
+ lr_scheduler: cosine
78
+ learning_rate: 2e-5
79
+
80
+ train_on_inputs: false
81
+ group_by_length: false
82
+ bf16: true
83
+ fp16:
84
+ tf32: false
85
+
86
+ gradient_checkpointing: true
87
+ gradient_checkpointing_kwargs:
88
+ use_reentrant: false
89
+ # early_stopping_patience: 3
90
+ local_rank:
91
+ logging_steps: 1
92
+ xformers_attention:
93
+ flash_attention: true
94
+
95
+ loss_watchdog_threshold: 5.0
96
+ loss_watchdog_patience: 3
97
+
98
+ hub_strategy: "every_save"
99
+ hub_model_id: andysalerno/rainbowfish-v7
100
+
101
+ num_epochs: 2
102
+ warmup_steps: 100
103
+ # warmup_ratio: 0.1
104
+ eval_steps: 200
105
+ eval_table_size:
106
+ eval_table_max_new_tokens: 128
107
+ # save_steps: 5
108
+ # max_steps: 400
109
+ saves_per_epoch: 2
110
+ debug:
111
+ weight_decay: 0.1
112
+ fsdp:
113
+ fsdp_config:
114
+ special_tokens:
115
+ bos_token: "<|im_start|>"
116
+ eos_token: "<|im_end|>"
117
+ unk_token: "<unk>"
118
+ ```
119
+
120
+ </details><br>
121
+
122
+ # rainbowfish-v7
123
+
124
+ This model is a fine-tuned version of [andysalerno/mistral-sft-v3](https://huggingface.co/andysalerno/mistral-sft-v3) on the None dataset.
125
+ It achieves the following results on the evaluation set:
126
+ - Loss: 0.6464
127
+
128
+ ## Model description
129
+
130
+ More information needed
131
+
132
+ ## Intended uses & limitations
133
+
134
+ More information needed
135
+
136
+ ## Training and evaluation data
137
+
138
+ More information needed
139
+
140
+ ## Training procedure
141
+
142
+ ### Training hyperparameters
143
+
144
+ The following hyperparameters were used during training:
145
+ - learning_rate: 2e-05
146
+ - train_batch_size: 4
147
+ - eval_batch_size: 4
148
+ - seed: 42
149
+ - distributed_type: multi-GPU
150
+ - num_devices: 4
151
+ - gradient_accumulation_steps: 4
152
+ - total_train_batch_size: 64
153
+ - total_eval_batch_size: 16
154
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
155
+ - lr_scheduler_type: cosine
156
+ - lr_scheduler_warmup_steps: 100
157
+ - num_epochs: 2
158
+
159
+ ### Training results
160
+
161
+ | Training Loss | Epoch | Step | Validation Loss |
162
+ |:-------------:|:-----:|:----:|:---------------:|
163
+ | 0.6514 | 0.18 | 200 | 0.6828 |
164
+ | 0.6875 | 0.37 | 400 | 0.6691 |
165
+ | 0.6626 | 0.55 | 600 | 0.6625 |
166
+ | 0.688 | 0.74 | 800 | 0.6558 |
167
+ | 0.7143 | 0.92 | 1000 | 0.6520 |
168
+ | 0.5243 | 1.11 | 1200 | 0.6495 |
169
+ | 0.6205 | 1.29 | 1400 | 0.6482 |
170
+ | 0.6159 | 1.47 | 1600 | 0.6469 |
171
+ | 0.6287 | 1.66 | 1800 | 0.6465 |
172
+ | 0.6606 | 1.84 | 2000 | 0.6464 |
173
+
174
+
175
+ ### Framework versions
176
+
177
+ - PEFT 0.8.2
178
+ - Transformers 4.38.0.dev0
179
+ - Pytorch 2.1.2+cu118
180
+ - Datasets 2.16.1
181
+ - Tokenizers 0.15.0
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ed9da7686387f0f355118a91bb0ccdcc19c7c97c1650e25dffa452f1ce44a6b
3
+ size 1195572114
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:464349a0ce2245b7dd1207ce4367922df51d680c918d9eb153c8b233001c26ca
3
  size 1195470168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ae78c76066af8084ee22333b297b2a4a0fcb1ed5858e0b040dbe83265928509
3
  size 1195470168