hamel commited on
Commit
512a32b
·
1 Parent(s): ea3421c

End of training

Browse files
README.md CHANGED
@@ -11,7 +11,38 @@ model-index:
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  # axolotl-test
16
 
17
  This model is a fine-tuned version of [openaccess-ai-collective/tiny-mistral](https://huggingface.co/openaccess-ai-collective/tiny-mistral) on the None dataset.
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)## axolotl config
15
+ ```yaml
16
+ base_model: openaccess-ai-collective/tiny-mistral
17
+ flash_attention: true
18
+ sequence_len: 1024
19
+ load_in_8bit: true
20
+ adapter: lora
21
+ lora_r: 32
22
+ lora_alpha: 64
23
+ lora_dropout: 0.05
24
+ lora_target_linear: true
25
+ val_set_size: 0.1
26
+ special_tokens:
27
+ unk_token: <unk>
28
+ bos_token: <s>
29
+ eos_token: </s>
30
+ datasets:
31
+ - path: mhenrichsen/alpaca_2k_test
32
+ type: alpaca
33
+ num_epochs: 2
34
+ micro_batch_size: 2
35
+ gradient_accumulation_steps: 1
36
+ output_dir: temp_dir
37
+ learning_rate: 0.00001
38
+ optimizer: adamw_torch
39
+ lr_scheduler: cosine
40
+ max_steps: 20
41
+ save_steps: 10
42
+ eval_steps: 10
43
+ hub_model_id: hamel/axolotl-test
44
+ dataset_processes: 1
45
+ ```
46
  # axolotl-test
47
 
48
  This model is a fine-tuned version of [openaccess-ai-collective/tiny-mistral](https://huggingface.co/openaccess-ai-collective/tiny-mistral) on the None dataset.
adapter_config.json CHANGED
@@ -16,13 +16,13 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "gate_proj",
20
  "up_proj",
 
 
 
21
  "down_proj",
22
  "v_proj",
23
- "o_proj",
24
- "k_proj",
25
- "q_proj"
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "up_proj",
20
+ "q_proj",
21
+ "gate_proj",
22
+ "o_proj",
23
  "down_proj",
24
  "v_proj",
25
+ "k_proj"
 
 
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1021d6333395951d27fb257dbbaf54629e4bc56bb7c684a2ac1f6f336d3ee80
3
  size 49035696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb0d0bf616b0ec3bd6fd2a39c4d00678d2582666fedfaa5a5ba722e303e5df34
3
  size 49035696
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4303f89612d33a141de3e34c1f741780fdec24e3dd717b397e5cb97683bde0e5
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c05a806adf74045eb5ac3ed69e66cc42c10590328fa2414dd57e9f02453b12d8
3
  size 5176