NobodyExistsOnTheInternet commited on
Commit
07c881d
1 Parent(s): 969a5aa

End of training

Browse files
README.md ADDED
@@ -0,0 +1,249 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - axolotl
6
+ - dpo
7
+ - trl
8
+ - dpo
9
+ - generated_from_trainer
10
+ base_model: mistralai/Mistral-7B-v0.1
11
+ model-index:
12
+ - name: mistral-7b-base-dpo-run
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
20
+ <details><summary>See axolotl config</summary>
21
+
22
+ axolotl version: `0.4.0`
23
+ ```yaml
24
+ base_model: mistralai/Mistral-7B-v0.1
25
+ base_model_ignore_patterns: []
26
+ base_model_config: mistralai/Mistral-7B-v0.1
27
+ model_revision:
28
+ tokenizer_config:
29
+ model_type: AutoModelForCausalLM
30
+ tokenizer_type: AutoTokenizer
31
+ trust_remote_code: true
32
+ tokenizer_use_fast: true
33
+ tokenizer_legacy: true
34
+ resize_token_embeddings_to_32x: false
35
+
36
+ is_falcon_derived_model: false
37
+ is_llama_derived_model: false
38
+ is_mistral_derived_model: true
39
+ is_qwen_derived_model: false
40
+
41
+ model_config:
42
+ rope_scaling:
43
+
44
+ bnb_config_kwargs:
45
+
46
+ gptq: false
47
+ gptq_groupsize:
48
+ gptq_model_v1: false
49
+
50
+ load_in_8bit: false
51
+ load_in_4bit: true
52
+
53
+ fp16: true
54
+
55
+ lora_on_cpu: false
56
+
57
+ rl: dpo
58
+ datasets:
59
+ - path: NobodyExistsOnTheInternet/Fixed-gutenberg-dpo-v0.1
60
+ split: train
61
+ type: chatml.intel
62
+ - path: NobodyExistsOnTheInternet/Fixed-Distilabel-intel-orca-dpo-pairs
63
+ split: train
64
+ type: chatml.intel
65
+ - path: NobodyExistsOnTheInternet/ToxicDPOqa
66
+ split: train
67
+ type: chatml.intel
68
+ - path: NobodyExistsOnTheInternet/system-message-DPO
69
+ split: train
70
+ type: chatml.intel
71
+ - path: NobodyExistsOnTheInternet/alpaca-intel-data-dpo
72
+ split: train
73
+ type: chatml.intel
74
+ - path: NobodyExistsOnTheInternet/ToxicDPOqa
75
+ split: train
76
+ type: chatml.intel
77
+
78
+
79
+ chat_template: chatml
80
+ default_system_message: Generate a preferable answer.
81
+ dataset_prepared_path: data/last_run_prepared
82
+ push_dataset_to_hub:
83
+ dataset_processes:
84
+ dataset_keep_in_memory:
85
+ hub_model_id: NobodyExistsOnTheInternet/mistral-7b-base-dpo-run
86
+ hub_strategy: every_save
87
+ hf_use_auth_token: true
88
+ val_set_size: 0
89
+ dataset_shard_num:
90
+ dataset_shard_idx:
91
+
92
+ sequence_len: 1024
93
+ sample_packing: false
94
+ eval_sample_packing:
95
+ sample_packing_eff_est:
96
+ total_num_tokens:
97
+
98
+ device_map:
99
+ max_memory:
100
+
101
+ adapter: qlora
102
+ lora_model_dir:
103
+
104
+ lora_r: 32
105
+ lora_alpha: 64
106
+ lora_dropout: 0.05
107
+ lora_target_linear: true
108
+ lora_target_module:
109
+
110
+ lora_modules_to_save:
111
+ - embed_tokens
112
+ - lm_head
113
+ lora_fan_in_fan_out:
114
+
115
+ wandb_project: dpo-hermes-2.5
116
+ wandb_entity:
117
+ wandb_watch:
118
+ wandb_name:
119
+ wandb_run_id:
120
+ wandb_log_model:
121
+
122
+ mlflow_tracking_uri:
123
+ mlflow_experiment_name:
124
+
125
+ output_dir: ./completed-model
126
+
127
+ torch_compile: true
128
+ gradient_accumulation_steps: 4
129
+ micro_batch_size: 1
130
+ eval_batch_size:
131
+ num_epochs: 2
132
+ warmup_steps: 100
133
+ warmup_ratio:
134
+ learning_rate: 0.000001
135
+ lr_quadratic_warmup:
136
+ logging_steps:
137
+ eval_steps:
138
+ evals_per_epoch:
139
+ save_strategy: steps
140
+ save_steps: 1000
141
+ saves_per_epoch:
142
+ save_total_limit:
143
+ eval_table_size:
144
+ eval_max_new_tokens:
145
+ eval_causal_lm_metrics:
146
+
147
+ loss_watchdog_threshold:
148
+ loss_watchwatchdog_patience:
149
+
150
+ train_on_inputs: false
151
+ group_by_length: false
152
+
153
+ gradient_checkpointing: true
154
+ gradient_checkpointing_kwargs:
155
+ use_reentrant: false
156
+
157
+ lr_scheduler:
158
+
159
+ optimizer: paged_adamw_8bit
160
+ weight_decay: 0.01
161
+ adam_beta1: 0.95
162
+ adam_beta2: 0.999
163
+ adam_epsilon: 0.0000001
164
+
165
+ neftune_noise_alpha: 5
166
+
167
+ flash_optimum:
168
+ xformers_attention:
169
+ flash_attention: true
170
+ flash_attn_cross_entropy:
171
+ flash_attn_rms_norm:
172
+ flash_attn_fuse_qkv:
173
+ flash_attn_fuse_mlp:
174
+ sdp_attention:
175
+ s2_attention:
176
+ resume_from_checkpoint:
177
+ auto_resume_from_checkpoints: false
178
+
179
+ local_rank:
180
+
181
+ tokens:
182
+
183
+ fsdp:
184
+ fsdp_config:
185
+
186
+ deepspeed:
187
+
188
+ ddp_timeout:
189
+ ddp_bucket_cap_mb:
190
+ ddp_broadcast_buffers:
191
+
192
+ torchdistx_path:
193
+
194
+ pretraining_dataset:
195
+
196
+ debug:
197
+
198
+ seed:
199
+
200
+ ```
201
+
202
+ </details><br>
203
+
204
+ # mistral-7b-base-dpo-run
205
+
206
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
207
+
208
+ ## Model description
209
+
210
+ More information needed
211
+
212
+ ## Intended uses & limitations
213
+
214
+ More information needed
215
+
216
+ ## Training and evaluation data
217
+
218
+ More information needed
219
+
220
+ ## Training procedure
221
+
222
+ ### Training hyperparameters
223
+
224
+ The following hyperparameters were used during training:
225
+ - learning_rate: 1e-06
226
+ - train_batch_size: 1
227
+ - eval_batch_size: 8
228
+ - seed: 42
229
+ - distributed_type: multi-GPU
230
+ - num_devices: 4
231
+ - gradient_accumulation_steps: 4
232
+ - total_train_batch_size: 16
233
+ - total_eval_batch_size: 32
234
+ - optimizer: Adam with betas=(0.95,0.999) and epsilon=1e-07
235
+ - lr_scheduler_type: cosine
236
+ - lr_scheduler_warmup_steps: 100
237
+ - training_steps: 15031
238
+
239
+ ### Training results
240
+
241
+
242
+
243
+ ### Framework versions
244
+
245
+ - PEFT 0.8.2
246
+ - Transformers 4.38.0
247
+ - Pytorch 2.2.0+cu121
248
+ - Datasets 2.16.1
249
+ - Tokenizers 0.15.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99bbea3a9f47522b8d9d6ad81b886e9e307ef855c26eeb9bf12205bfd8c6b264
3
  size 1384182104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d73df302ad2980a1dc5078e34787075785e5b639f2909ec0e1490a4e69f4a6a3
3
  size 1384182104
config.json CHANGED
@@ -34,7 +34,7 @@
34
  "rope_theta": 10000.0,
35
  "sliding_window": 4096,
36
  "tie_word_embeddings": false,
37
- "torch_dtype": "bfloat16",
38
  "transformers_version": "4.38.0",
39
  "use_cache": false,
40
  "vocab_size": 32000
 
34
  "rope_theta": 10000.0,
35
  "sliding_window": 4096,
36
  "tie_word_embeddings": false,
37
+ "torch_dtype": "float32",
38
  "transformers_version": "4.38.0",
39
  "use_cache": false,
40
  "vocab_size": 32000
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": true,
5
+ "eos_token_id": 2,
6
+ "transformers_version": "4.38.0"
7
+ }
pytorch_model-00001-of-00002.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab3ebfe135a69b54abaf5d30f63310a98312b7b2de0f849ad1895ab499789c60
3
+ size 4986578982
pytorch_model-00002-of-00002.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8ec967a5d5093595fdd83d143b93a20830bade33e2d8505358fe40f6f635e58
3
+ size 1048577750
pytorch_model.bin.index.json ADDED
The diff for this file is too large to render. See raw diff
 
runs/Mar10_11-25-20_05c27c174db0/events.out.tfevents.1710069924.05c27c174db0.832.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04cf8639924fe5849e3c8bf7eee9537e0ef92ddce99b9068bbcca8746c9e402e
3
- size 10323609
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48eb65c80c8852641bc29d68b0d669ff26d74458a7b77fc51df50baf1cce2c4d
3
+ size 10345291