imdatta0 commited on
Commit
ed6f0d5
1 Parent(s): 878db0c

End of training

Browse files
Files changed (2) hide show
  1. README.md +51 -50
  2. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -2,21 +2,22 @@
2
  license: llama2
3
  library_name: peft
4
  tags:
 
5
  - generated_from_trainer
6
  base_model: meta-llama/Llama-2-7b-hf
7
  model-index:
8
- - name: llama_2_7b_MetaMathQA_40K_downNupNgateNqNkNvNo_r8_lr0.0001_bg88_alpha8_0_41
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- # llama_2_7b_MetaMathQA_40K_downNupNgateNqNkNvNo_r8_lr0.0001_bg88_alpha8_0_41
16
 
17
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.5581
20
 
21
  ## Model description
22
 
@@ -50,53 +51,53 @@ The following hyperparameters were used during training:
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:------:|:----:|:---------------:|
53
- | 1.0537 | 0.0211 | 13 | 0.9041 |
54
- | 0.8106 | 0.0421 | 26 | 0.7414 |
55
- | 0.6994 | 0.0632 | 39 | 0.6928 |
56
- | 0.6515 | 0.0842 | 52 | 0.6703 |
57
- | 0.6348 | 0.1053 | 65 | 0.6533 |
58
- | 0.612 | 0.1264 | 78 | 0.6429 |
59
- | 0.6204 | 0.1474 | 91 | 0.6329 |
60
- | 0.5964 | 0.1685 | 104 | 0.6248 |
61
- | 0.6179 | 0.1896 | 117 | 0.6181 |
62
- | 0.6008 | 0.2106 | 130 | 0.6133 |
63
- | 0.5846 | 0.2317 | 143 | 0.6087 |
64
- | 0.5685 | 0.2527 | 156 | 0.6043 |
65
- | 0.5704 | 0.2738 | 169 | 0.6001 |
66
- | 0.5657 | 0.2949 | 182 | 0.5961 |
67
- | 0.551 | 0.3159 | 195 | 0.5937 |
68
- | 0.5658 | 0.3370 | 208 | 0.5904 |
69
- | 0.5677 | 0.3580 | 221 | 0.5876 |
70
- | 0.5589 | 0.3791 | 234 | 0.5839 |
71
- | 0.5484 | 0.4002 | 247 | 0.5823 |
72
- | 0.5514 | 0.4212 | 260 | 0.5802 |
73
- | 0.5694 | 0.4423 | 273 | 0.5782 |
74
- | 0.5551 | 0.4633 | 286 | 0.5768 |
75
- | 0.5562 | 0.4844 | 299 | 0.5744 |
76
- | 0.5596 | 0.5055 | 312 | 0.5728 |
77
- | 0.5402 | 0.5265 | 325 | 0.5714 |
78
- | 0.5461 | 0.5476 | 338 | 0.5698 |
79
- | 0.536 | 0.5687 | 351 | 0.5691 |
80
- | 0.5421 | 0.5897 | 364 | 0.5668 |
81
- | 0.5397 | 0.6108 | 377 | 0.5660 |
82
- | 0.5397 | 0.6318 | 390 | 0.5652 |
83
- | 0.547 | 0.6529 | 403 | 0.5641 |
84
- | 0.5465 | 0.6740 | 416 | 0.5634 |
85
- | 0.546 | 0.6950 | 429 | 0.5623 |
86
- | 0.5434 | 0.7161 | 442 | 0.5617 |
87
- | 0.5366 | 0.7371 | 455 | 0.5611 |
88
- | 0.5306 | 0.7582 | 468 | 0.5603 |
89
- | 0.5441 | 0.7793 | 481 | 0.5600 |
90
- | 0.5305 | 0.8003 | 494 | 0.5592 |
91
- | 0.5255 | 0.8214 | 507 | 0.5588 |
92
- | 0.546 | 0.8424 | 520 | 0.5587 |
93
- | 0.545 | 0.8635 | 533 | 0.5586 |
94
- | 0.5377 | 0.8846 | 546 | 0.5584 |
95
- | 0.5361 | 0.9056 | 559 | 0.5582 |
96
- | 0.5315 | 0.9267 | 572 | 0.5581 |
97
- | 0.5394 | 0.9478 | 585 | 0.5580 |
98
- | 0.534 | 0.9688 | 598 | 0.5579 |
99
- | 0.5362 | 0.9899 | 611 | 0.5581 |
100
 
101
 
102
  ### Framework versions
 
2
  license: llama2
3
  library_name: peft
4
  tags:
5
+ - unsloth
6
  - generated_from_trainer
7
  base_model: meta-llama/Llama-2-7b-hf
8
  model-index:
9
+ - name: llama_2_7b_MetaMathQA_40K
10
  results: []
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ # llama_2_7b_MetaMathQA_40K
17
 
18
  This model is a fine-tuned version of [meta-llama/Llama-2-7b-hf](https://huggingface.co/meta-llama/Llama-2-7b-hf) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.5058
21
 
22
  ## Model description
23
 
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:------:|:----:|:---------------:|
54
+ | 0.8364 | 0.0211 | 13 | 0.6835 |
55
+ | 0.642 | 0.0421 | 26 | 0.6490 |
56
+ | 0.6184 | 0.0632 | 39 | 0.6282 |
57
+ | 0.5964 | 0.0842 | 52 | 0.6165 |
58
+ | 0.5847 | 0.1053 | 65 | 0.6063 |
59
+ | 0.5688 | 0.1264 | 78 | 0.6001 |
60
+ | 0.5782 | 0.1474 | 91 | 0.5918 |
61
+ | 0.5552 | 0.1685 | 104 | 0.5858 |
62
+ | 0.5831 | 0.1896 | 117 | 0.5824 |
63
+ | 0.5693 | 0.2106 | 130 | 0.5779 |
64
+ | 0.5469 | 0.2317 | 143 | 0.5726 |
65
+ | 0.5318 | 0.2527 | 156 | 0.5695 |
66
+ | 0.5368 | 0.2738 | 169 | 0.5664 |
67
+ | 0.5359 | 0.2949 | 182 | 0.5610 |
68
+ | 0.5226 | 0.3159 | 195 | 0.5574 |
69
+ | 0.5341 | 0.3370 | 208 | 0.5532 |
70
+ | 0.5356 | 0.3580 | 221 | 0.5514 |
71
+ | 0.5275 | 0.3791 | 234 | 0.5479 |
72
+ | 0.5145 | 0.4002 | 247 | 0.5444 |
73
+ | 0.5177 | 0.4212 | 260 | 0.5419 |
74
+ | 0.5334 | 0.4423 | 273 | 0.5402 |
75
+ | 0.5155 | 0.4633 | 286 | 0.5369 |
76
+ | 0.5213 | 0.4844 | 299 | 0.5346 |
77
+ | 0.5211 | 0.5055 | 312 | 0.5310 |
78
+ | 0.5048 | 0.5265 | 325 | 0.5300 |
79
+ | 0.5131 | 0.5476 | 338 | 0.5277 |
80
+ | 0.4965 | 0.5687 | 351 | 0.5265 |
81
+ | 0.5053 | 0.5897 | 364 | 0.5227 |
82
+ | 0.4989 | 0.6108 | 377 | 0.5210 |
83
+ | 0.5005 | 0.6318 | 390 | 0.5190 |
84
+ | 0.5037 | 0.6529 | 403 | 0.5181 |
85
+ | 0.507 | 0.6740 | 416 | 0.5167 |
86
+ | 0.5002 | 0.6950 | 429 | 0.5154 |
87
+ | 0.498 | 0.7161 | 442 | 0.5141 |
88
+ | 0.491 | 0.7371 | 455 | 0.5121 |
89
+ | 0.4834 | 0.7582 | 468 | 0.5106 |
90
+ | 0.4971 | 0.7793 | 481 | 0.5094 |
91
+ | 0.4864 | 0.8003 | 494 | 0.5085 |
92
+ | 0.4778 | 0.8214 | 507 | 0.5076 |
93
+ | 0.4991 | 0.8424 | 520 | 0.5073 |
94
+ | 0.4951 | 0.8635 | 533 | 0.5068 |
95
+ | 0.489 | 0.8846 | 546 | 0.5064 |
96
+ | 0.4916 | 0.9056 | 559 | 0.5061 |
97
+ | 0.4841 | 0.9267 | 572 | 0.5058 |
98
+ | 0.4919 | 0.9478 | 585 | 0.5058 |
99
+ | 0.486 | 0.9688 | 598 | 0.5059 |
100
+ | 0.489 | 0.9899 | 611 | 0.5058 |
101
 
102
 
103
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4c4efc944c0853dedadae9e49f44f3403deb726c0d232f899b81da33c959ffe
3
  size 80013120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c75a96b90989969e413c72e360bec5974b35ccc0e59ce9e8d02decf5a07cb16
3
  size 80013120