Liu-Xiang commited on
Commit
21ff036
1 Parent(s): 04a49d9

LLM-Alchemy-Chamber/mistral-instruct-generation

Browse files
README.md CHANGED
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [mistralai/Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.8556
20
 
21
  ## Model description
22
 
@@ -48,16 +48,16 @@ The following hyperparameters were used during training:
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
- | 1.1522 | 0.03 | 10 | 1.1246 |
52
- | 1.0849 | 0.07 | 20 | 1.0571 |
53
- | 1.0144 | 0.1 | 30 | 0.9943 |
54
- | 0.9654 | 0.13 | 40 | 0.9525 |
55
- | 0.9383 | 0.16 | 50 | 0.9212 |
56
- | 0.8954 | 0.2 | 60 | 0.8979 |
57
- | 0.8671 | 0.23 | 70 | 0.8798 |
58
- | 0.9088 | 0.26 | 80 | 0.8664 |
59
- | 0.8696 | 0.3 | 90 | 0.8584 |
60
- | 0.8281 | 0.33 | 100 | 0.8556 |
61
 
62
 
63
  ### Framework versions
 
16
 
17
  This model is a fine-tuned version of [mistralai/Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.8576
20
 
21
  ## Model description
22
 
 
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
+ | 1.1699 | 0.03 | 10 | 1.1218 |
52
+ | 1.0878 | 0.07 | 20 | 1.0544 |
53
+ | 1.0525 | 0.1 | 30 | 0.9935 |
54
+ | 0.9611 | 0.13 | 40 | 0.9529 |
55
+ | 0.931 | 0.16 | 50 | 0.9230 |
56
+ | 0.9212 | 0.2 | 60 | 0.8993 |
57
+ | 0.8918 | 0.23 | 70 | 0.8817 |
58
+ | 0.8808 | 0.26 | 80 | 0.8683 |
59
+ | 0.8575 | 0.3 | 90 | 0.8604 |
60
+ | 0.8848 | 0.33 | 100 | 0.8576 |
61
 
62
 
63
  ### Framework versions
adapter_config.json CHANGED
@@ -20,14 +20,14 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj",
25
- "down_proj",
26
  "o_proj",
27
- "k_proj",
 
28
  "gate_proj",
29
  "lm_head",
30
- "up_proj"
 
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "up_proj",
 
 
24
  "o_proj",
25
+ "down_proj",
26
+ "q_proj",
27
  "gate_proj",
28
  "lm_head",
29
+ "k_proj",
30
+ "v_proj"
31
  ],
32
  "task_type": "CAUSAL_LM",
33
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef037664ff1f241c9f05dbc8529a74c4b3f73f84043e90d3ed3dab3fdf3797f5
3
  size 751667752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2144d3d7ba49f9a6d25161173a1ced5aa06c73567931c64e8c7e06b9052e3cb9
3
  size 751667752
runs/Mar17_02-55-22_llm-back-project-workbench-0/events.out.tfevents.1710644123.llm-back-project-workbench-0.376.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aebc90b5ee6e9f6903e482d129d0f7b4838c56c879fa407cba00d8bd4f19e271
3
+ size 9431
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6bbec64aaf2fa220f49f1f4d5180dd6a7e764015a902bf93f9475b73fcce36fa
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1636e27ff461eb334da7d10b79596f6474584607f17ad70a1ebbafb1d09dc0
3
  size 4283