gultar commited on
Commit
e5b0f05
1 Parent(s): 932f17c

End of training

Browse files
Files changed (2) hide show
  1. README.md +190 -0
  2. adapter_model.bin +3 -0
README.md CHANGED
@@ -1,3 +1,193 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - axolotl
6
+ - generated_from_trainer
7
+ base_model: teknium/OpenHermes-2-Mistral-7B
8
+ model-index:
9
+ - name: Hermes-Agent
10
+ results: []
11
  ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
17
+ <details><summary>See axolotl config</summary>
18
+
19
+ axolotl version: `0.4.0`
20
+ ```yaml
21
+ base_model: teknium/OpenHermes-2-Mistral-7B
22
+ base_model_config: teknium/OpenHermes-2-Mistral-7B
23
+ model_type: MistralForCausalLM
24
+ tokenizer_type: LlamaTokenizer
25
+ is_mistral_derived_model: true
26
+
27
+ load_in_8bit: false
28
+ load_in_4bit: true
29
+ strict: false
30
+
31
+ datasets:
32
+ - path: THUDM/AgentInstruct
33
+ type: sharegpt
34
+ conversation: llama-2
35
+ split: os
36
+ - path: THUDM/AgentInstruct
37
+ type: sharegpt
38
+ conversation: llama-2
39
+ split: db
40
+ - path: THUDM/AgentInstruct
41
+ type: sharegpt
42
+ conversation: llama-2
43
+ split: alfworld
44
+ - path: THUDM/AgentInstruct
45
+ type: sharegpt
46
+ conversation: llama-2
47
+ split: webshop
48
+ - path: THUDM/AgentInstruct
49
+ type: sharegpt
50
+ conversation: llama-2
51
+ split: kg
52
+ - path: THUDM/AgentInstruct
53
+ type: sharegpt
54
+ conversation: llama-2
55
+ split: mind2web
56
+
57
+ dataset_prepared_path: last_run_prepared
58
+ val_set_size: 0.1
59
+ output_dir: ./qlora-out
60
+
61
+ wandb_project: "Mistral-Agent"
62
+ wandb_log_model: "checkpoint"
63
+
64
+ hub_model_id: "gultar/Hermes-Agent"
65
+
66
+ chat_template: inst
67
+
68
+ adapter: qlora
69
+ lora_model_dir:
70
+
71
+ sequence_len: 8192
72
+ sample_packing: true
73
+ pad_to_sequence_len: true
74
+
75
+ lora_r: 8
76
+ lora_alpha: 16
77
+ lora_dropout: 0.05
78
+ lora_target_linear: true
79
+ lora_fan_in_fan_out:
80
+ lora_target_modules:
81
+ - gate_proj
82
+ - down_proj
83
+ - up_proj
84
+ - q_proj
85
+ - v_proj
86
+ - k_proj
87
+ - o_proj
88
+
89
+ wandb_project:
90
+ wandb_entity:
91
+ wandb_watch:
92
+ wandb_name:
93
+ wandb_log_model:
94
+
95
+ gradient_accumulation_steps: 4
96
+ micro_batch_size: 2
97
+ num_epochs: 1
98
+ optimizer: adamw_bnb_8bit
99
+ lr_scheduler: cosine
100
+ learning_rate: 0.0002
101
+
102
+ train_on_inputs: false
103
+ group_by_length: false
104
+ bf16: true
105
+ fp16: false
106
+ tf32: false
107
+
108
+ gradient_checkpointing: true
109
+ early_stopping_patience:
110
+ resume_from_checkpoint:
111
+ local_rank:
112
+ logging_steps: 1
113
+ xformers_attention:
114
+ flash_attention: true
115
+
116
+ loss_watchdog_threshold: 5.0
117
+ loss_watchdog_patience: 3
118
+
119
+ warmup_steps: 10
120
+ evals_per_epoch: 4
121
+ eval_table_size:
122
+ eval_table_max_new_tokens: 128
123
+ saves_per_epoch: 1
124
+ debug:
125
+ deepspeed:
126
+ weight_decay: 0.0
127
+ fsdp:
128
+ fsdp_config:
129
+ special_tokens:
130
+ bos_token: "<s>"
131
+ eos_token: "</s>"
132
+ unk_token: "<unk>"
133
+ lora_modules_to_save:
134
+ - lm_head
135
+ - embed_tokens
136
+
137
+
138
+
139
+ ```
140
+
141
+ </details><br>
142
+
143
+ # Hermes-Agent
144
+
145
+ This model is a fine-tuned version of [teknium/OpenHermes-2-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2-Mistral-7B) on the None dataset.
146
+ It achieves the following results on the evaluation set:
147
+ - Loss: 0.3002
148
+
149
+ ## Model description
150
+
151
+ More information needed
152
+
153
+ ## Intended uses & limitations
154
+
155
+ More information needed
156
+
157
+ ## Training and evaluation data
158
+
159
+ More information needed
160
+
161
+ ## Training procedure
162
+
163
+ ### Training hyperparameters
164
+
165
+ The following hyperparameters were used during training:
166
+ - learning_rate: 0.0002
167
+ - train_batch_size: 2
168
+ - eval_batch_size: 2
169
+ - seed: 42
170
+ - gradient_accumulation_steps: 4
171
+ - total_train_batch_size: 8
172
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
173
+ - lr_scheduler_type: cosine
174
+ - lr_scheduler_warmup_steps: 10
175
+ - num_epochs: 1
176
+
177
+ ### Training results
178
+
179
+ | Training Loss | Epoch | Step | Validation Loss |
180
+ |:-------------:|:-----:|:----:|:---------------:|
181
+ | 0.6859 | 0.03 | 1 | 0.7320 |
182
+ | 0.3448 | 0.26 | 9 | 0.3382 |
183
+ | 0.4193 | 0.53 | 18 | 0.3233 |
184
+ | 0.2986 | 0.79 | 27 | 0.3002 |
185
+
186
+
187
+ ### Framework versions
188
+
189
+ - PEFT 0.8.2
190
+ - Transformers 4.38.0.dev0
191
+ - Pytorch 2.0.1+cu118
192
+ - Datasets 2.16.1
193
+ - Tokenizers 0.15.0
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b099ddd8160d053df9012186feb0d563d31b2aaf1ddb338c64b113c58228e92
3
+ size 608368405