nblinh63 commited on
Commit
5d8ad07
·
verified ·
1 Parent(s): 3a07b2e

End of training

Browse files
README.md CHANGED
@@ -78,7 +78,7 @@ resume_from_checkpoint: null
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
- sequence_len: 2048
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
@@ -103,7 +103,7 @@ xformers_attention: true
103
 
104
  This model is a fine-tuned version of [unsloth/gemma-2b-it](https://huggingface.co/unsloth/gemma-2b-it) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: 1.1827
107
 
108
  ## Model description
109
 
@@ -135,7 +135,7 @@ The following hyperparameters were used during training:
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
- | 1.3076 | 0.0046 | 10 | 1.1827 |
139
 
140
 
141
  ### Framework versions
 
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
+ sequence_len: 3072
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
 
103
 
104
  This model is a fine-tuned version of [unsloth/gemma-2b-it](https://huggingface.co/unsloth/gemma-2b-it) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 1.1491
107
 
108
  ## Model description
109
 
 
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
+ | 1.3424 | 0.0046 | 10 | 1.1491 |
139
 
140
 
141
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
  "up_proj",
 
 
 
25
  "q_proj",
26
  "gate_proj",
27
- "down_proj",
28
- "o_proj",
29
- "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "up_proj",
24
+ "o_proj",
25
+ "down_proj",
26
+ "k_proj",
27
  "q_proj",
28
  "gate_proj",
29
+ "v_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee3649262999aa3c61bd48c21fc879dc2a5ce0760d6dcc09dd68b540d5458fcb
3
  size 78537274
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe68c5ecae888631cb76716cb2ac236080f417b646cefd2561e378e70cd6ba85
3
  size 78537274
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:614dd8f0204a963ebebf6e1ac06ace2de16a3ace590bb3ee3134ef7220b429cb
3
  size 78480072
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a57f75a35cea0ebeff06fc046a1c7f3075a87554f0291766141805c3e43f38c
3
  size 78480072
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b991a4e1c6dfa4ed81243d209bdc061e7a670ccd591d81686f35c699ad10a756
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abe0ebb60c3333ae384da44c98de6aa8aaa0ee15e6f147db2e4a44f6405504b1
3
  size 6776