arshiakarimian1 commited on
Commit
681e602
1 Parent(s): 3937eb1

Model save

Browse files
README.md CHANGED
@@ -17,8 +17,6 @@ should probably proofread and complete it, then remove this comment. -->
17
  # spam-llama3.1-8B-teacher-2
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the None dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 1.6740
22
 
23
  ## Model description
24
 
@@ -49,16 +47,6 @@ The following hyperparameters were used during training:
49
  - num_epochs: 1
50
  - mixed_precision_training: Native AMP
51
 
52
- ### Training results
53
-
54
- | Training Loss | Epoch | Step | Validation Loss |
55
- |:-------------:|:------:|:----:|:---------------:|
56
- | 1.6576 | 0.2009 | 91 | 1.8055 |
57
- | 1.5242 | 0.4019 | 182 | 1.7284 |
58
- | 1.5695 | 0.6028 | 273 | 1.6887 |
59
- | 1.3364 | 0.8038 | 364 | 1.6740 |
60
-
61
-
62
  ### Framework versions
63
 
64
  - PEFT 0.12.0
 
17
  # spam-llama3.1-8B-teacher-2
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the None dataset.
 
 
20
 
21
  ## Model description
22
 
 
47
  - num_epochs: 1
48
  - mixed_precision_training: Native AMP
49
 
 
 
 
 
 
 
 
 
 
 
50
  ### Framework versions
51
 
52
  - PEFT 0.12.0
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "gate_proj",
24
- "up_proj",
25
- "k_proj",
26
- "down_proj",
27
  "o_proj",
 
 
28
  "v_proj",
29
- "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "q_proj",
24
  "gate_proj",
 
 
 
25
  "o_proj",
26
+ "k_proj",
27
+ "up_proj",
28
  "v_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bab7ccfea159a00fd9baa91f267b4cd4661d5e1fb68a2bd0582dd571ba2a889
3
  size 671149168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cde1a66817dddf0d9289c1ce9fe5c154c8e8b40700cff8b94e290a6e55d9dee1
3
  size 671149168
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f67439908f4c503200f661217ea4e257664e11cc5ac1067da2d2ca1d0599f467
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dded637a70887f4e9f53837ff0525a6fa2e6f12cd3d0b05f1c26302cff590c9
3
  size 5560