Model save
Browse files- README.md +155 -0
- all_results.json +21 -0
- config.json +27 -0
- eval_results.json +16 -0
- generation_config.json +10 -0
- model-00001-of-00003.safetensors +3 -0
- model-00002-of-00003.safetensors +3 -0
- model-00003-of-00003.safetensors +3 -0
- model.safetensors.index.json +298 -0
- runs/Jan14_05-17-19_amaterasu/events.out.tfevents.1705209484.amaterasu.1284454.0 +3 -0
- runs/Jan14_05-17-19_amaterasu/events.out.tfevents.1705320794.amaterasu.1284454.1 +3 -0
- special_tokens_map.json +24 -0
- tokenizer.json +0 -0
- tokenizer_config.json +40 -0
- train_results.json +8 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
README.md
ADDED
@@ -0,0 +1,155 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
tags:
|
3 |
+
- generated_from_trainer
|
4 |
+
model-index:
|
5 |
+
- name: llama2-7b-chat-dpo-full-hydrox-safe
|
6 |
+
results: []
|
7 |
+
---
|
8 |
+
|
9 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
10 |
+
should probably proofread and complete it, then remove this comment. -->
|
11 |
+
|
12 |
+
# llama2-7b-chat-dpo-full-hydrox-safe
|
13 |
+
|
14 |
+
This model was trained from scratch on an unknown dataset.
|
15 |
+
It achieves the following results on the evaluation set:
|
16 |
+
- Loss: 0.0013
|
17 |
+
- Rewards/chosen: -0.0939
|
18 |
+
- Rewards/rejected: -28.6036
|
19 |
+
- Rewards/accuracies: 0.9992
|
20 |
+
- Rewards/margins: 28.5097
|
21 |
+
- Logps/rejected: -700.8997
|
22 |
+
- Logps/chosen: -219.0951
|
23 |
+
- Logits/rejected: -0.7196
|
24 |
+
- Logits/chosen: -0.6433
|
25 |
+
|
26 |
+
## Model description
|
27 |
+
|
28 |
+
More information needed
|
29 |
+
|
30 |
+
## Intended uses & limitations
|
31 |
+
|
32 |
+
More information needed
|
33 |
+
|
34 |
+
## Training and evaluation data
|
35 |
+
|
36 |
+
More information needed
|
37 |
+
|
38 |
+
## Training procedure
|
39 |
+
|
40 |
+
### Training hyperparameters
|
41 |
+
|
42 |
+
The following hyperparameters were used during training:
|
43 |
+
- learning_rate: 5e-07
|
44 |
+
- train_batch_size: 8
|
45 |
+
- eval_batch_size: 4
|
46 |
+
- seed: 42
|
47 |
+
- distributed_type: multi-GPU
|
48 |
+
- num_devices: 8
|
49 |
+
- total_train_batch_size: 64
|
50 |
+
- total_eval_batch_size: 32
|
51 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
52 |
+
- lr_scheduler_type: linear
|
53 |
+
- lr_scheduler_warmup_ratio: 0.1
|
54 |
+
- num_epochs: 3
|
55 |
+
|
56 |
+
### Training results
|
57 |
+
|
58 |
+
| Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
|
59 |
+
|:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
|
60 |
+
| 0.6247 | 0.03 | 100 | 0.6155 | 0.0158 | -0.1451 | 0.8577 | 0.1608 | -416.3145 | -217.9982 | -0.2810 | -0.5546 |
|
61 |
+
| 0.3738 | 0.07 | 200 | 0.3507 | 0.1699 | -0.8998 | 0.9184 | 1.0697 | -423.8618 | -216.4572 | -0.3464 | -0.5597 |
|
62 |
+
| 0.2144 | 0.1 | 300 | 0.2152 | 0.3547 | -1.8023 | 0.9394 | 2.1570 | -432.8871 | -214.6091 | -0.3537 | -0.5335 |
|
63 |
+
| 0.1567 | 0.14 | 400 | 0.1458 | 0.4781 | -2.9967 | 0.9554 | 3.4748 | -444.8309 | -213.3745 | -0.3343 | -0.4947 |
|
64 |
+
| 0.1121 | 0.17 | 500 | 0.1250 | 0.5327 | -4.1425 | 0.9689 | 4.6752 | -456.2888 | -212.8291 | -0.3020 | -0.4604 |
|
65 |
+
| 0.1003 | 0.2 | 600 | 0.0926 | 0.3744 | -5.4423 | 0.9697 | 5.8167 | -469.2868 | -214.4121 | -0.3227 | -0.4643 |
|
66 |
+
| 0.0602 | 0.24 | 700 | 0.0769 | 0.3722 | -6.6327 | 0.9739 | 7.0049 | -481.1906 | -214.4338 | -0.3134 | -0.4460 |
|
67 |
+
| 0.0584 | 0.27 | 800 | 0.0638 | 0.4037 | -7.6613 | 0.9806 | 8.0650 | -491.4773 | -214.1189 | -0.2857 | -0.4235 |
|
68 |
+
| 0.0555 | 0.31 | 900 | 0.0557 | 0.4281 | -8.2327 | 0.9848 | 8.6608 | -497.1909 | -213.8745 | -0.2914 | -0.4270 |
|
69 |
+
| 0.0471 | 0.34 | 1000 | 0.0472 | 0.4046 | -9.7769 | 0.9891 | 10.1814 | -512.6325 | -214.1102 | -0.3491 | -0.4740 |
|
70 |
+
| 0.0673 | 0.37 | 1100 | 0.0383 | 0.3282 | -11.0251 | 0.9949 | 11.3534 | -525.1152 | -214.8733 | -0.3772 | -0.4955 |
|
71 |
+
| 0.031 | 0.41 | 1200 | 0.0325 | 0.1923 | -11.8461 | 0.9958 | 12.0384 | -533.3251 | -216.2326 | -0.4115 | -0.5142 |
|
72 |
+
| 0.0242 | 0.44 | 1300 | 0.0275 | 0.2059 | -12.9425 | 0.9966 | 13.1485 | -544.2894 | -216.0965 | -0.4212 | -0.5150 |
|
73 |
+
| 0.0143 | 0.48 | 1400 | 0.0215 | 0.0180 | -11.9692 | 0.9958 | 11.9872 | -534.5560 | -217.9758 | -0.5405 | -0.5538 |
|
74 |
+
| 0.0157 | 0.51 | 1500 | 0.0181 | -0.3479 | -14.0856 | 0.9958 | 13.7377 | -555.7203 | -221.6349 | -0.5292 | -0.5576 |
|
75 |
+
| 0.0155 | 0.54 | 1600 | 0.0286 | -0.2238 | -13.8665 | 0.9958 | 13.6427 | -553.5294 | -220.3942 | -0.4943 | -0.5256 |
|
76 |
+
| 0.0148 | 0.58 | 1700 | 0.0251 | -0.2352 | -15.8803 | 0.9975 | 15.6451 | -573.6669 | -220.5081 | -0.4799 | -0.5212 |
|
77 |
+
| 0.0094 | 0.61 | 1800 | 0.0163 | -0.1817 | -16.7316 | 0.9975 | 16.5499 | -582.1795 | -219.9725 | -0.4976 | -0.5385 |
|
78 |
+
| 0.0112 | 0.65 | 1900 | 0.0159 | -0.3917 | -18.6440 | 0.9949 | 18.2523 | -601.3036 | -222.0726 | -0.5874 | -0.6174 |
|
79 |
+
| 0.007 | 0.68 | 2000 | 0.0106 | -0.1240 | -16.5280 | 0.9975 | 16.4040 | -580.1437 | -219.3957 | -0.5555 | -0.5702 |
|
80 |
+
| 0.0083 | 0.71 | 2100 | 0.0167 | -0.3388 | -18.5238 | 0.9975 | 18.1849 | -600.1016 | -221.5440 | -0.5802 | -0.5848 |
|
81 |
+
| 0.0058 | 0.75 | 2200 | 0.0166 | 0.1875 | -16.4876 | 0.9975 | 16.6751 | -579.7398 | -216.2812 | -0.5300 | -0.5517 |
|
82 |
+
| 0.0031 | 0.78 | 2300 | 0.0167 | -0.4853 | -19.1077 | 0.9966 | 18.6224 | -605.9405 | -223.0087 | -0.5945 | -0.5932 |
|
83 |
+
| 0.0041 | 0.82 | 2400 | 0.0148 | -0.1266 | -19.3544 | 0.9983 | 19.2278 | -608.4083 | -219.4222 | -0.5528 | -0.5695 |
|
84 |
+
| 0.0129 | 0.85 | 2500 | 0.0277 | -0.6526 | -21.0389 | 0.9983 | 20.3863 | -625.2532 | -224.6820 | -0.6317 | -0.6223 |
|
85 |
+
| 0.0169 | 0.88 | 2600 | 0.0158 | -0.6507 | -22.0352 | 0.9983 | 21.3845 | -635.2158 | -224.6625 | -0.6147 | -0.6148 |
|
86 |
+
| 0.005 | 0.92 | 2700 | 0.0148 | -0.7455 | -22.5637 | 0.9983 | 21.8181 | -640.5008 | -225.6113 | -0.6401 | -0.6379 |
|
87 |
+
| 0.0075 | 0.95 | 2800 | 0.0429 | -0.6179 | -21.7587 | 0.9983 | 21.1408 | -632.4512 | -224.3349 | -0.6580 | -0.6338 |
|
88 |
+
| 0.0053 | 0.99 | 2900 | 0.0452 | -0.3093 | -21.3611 | 0.9983 | 21.0518 | -628.4748 | -221.2488 | -0.6473 | -0.6362 |
|
89 |
+
| 0.0033 | 1.02 | 3000 | 0.0399 | -0.4299 | -20.6185 | 0.9992 | 20.1886 | -621.0488 | -222.4544 | -0.6812 | -0.6500 |
|
90 |
+
| 0.1239 | 1.05 | 3100 | 0.0098 | -0.4156 | -21.6528 | 0.9992 | 21.2371 | -631.3915 | -222.3120 | -0.6612 | -0.6328 |
|
91 |
+
| 0.0029 | 1.09 | 3200 | 0.0041 | -0.4823 | -24.1370 | 0.9992 | 23.6547 | -656.2342 | -222.9791 | -0.6460 | -0.6310 |
|
92 |
+
| 0.0015 | 1.12 | 3300 | 0.0037 | -0.6250 | -25.4442 | 0.9975 | 24.8192 | -669.3063 | -224.4059 | -0.6623 | -0.6482 |
|
93 |
+
| 0.002 | 1.16 | 3400 | 0.0039 | -0.1881 | -23.5637 | 0.9983 | 23.3756 | -650.5010 | -220.0367 | -0.6331 | -0.6142 |
|
94 |
+
| 0.0027 | 1.19 | 3500 | 0.0039 | -0.3251 | -24.0619 | 0.9992 | 23.7368 | -655.4830 | -221.4067 | -0.6644 | -0.6402 |
|
95 |
+
| 0.0015 | 1.22 | 3600 | 0.0031 | -0.4337 | -26.8013 | 0.9983 | 26.3676 | -682.8770 | -222.4931 | -0.6421 | -0.6330 |
|
96 |
+
| 0.0067 | 1.26 | 3700 | 0.0030 | -0.1107 | -22.8513 | 0.9992 | 22.7406 | -643.3767 | -219.2624 | -0.6412 | -0.6162 |
|
97 |
+
| 0.002 | 1.29 | 3800 | 0.0029 | -0.4330 | -24.7254 | 0.9992 | 24.2925 | -662.1182 | -222.4855 | -0.6750 | -0.6447 |
|
98 |
+
| 0.004 | 1.33 | 3900 | 0.0026 | -0.5258 | -25.6407 | 0.9992 | 25.1150 | -671.2714 | -223.4133 | -0.6613 | -0.6319 |
|
99 |
+
| 0.0032 | 1.36 | 4000 | 0.0025 | -0.8592 | -27.4389 | 0.9975 | 26.5797 | -689.2528 | -226.7478 | -0.6796 | -0.6569 |
|
100 |
+
| 0.0293 | 1.39 | 4100 | 0.0032 | -0.6286 | -26.4388 | 0.9992 | 25.8102 | -679.2518 | -224.4421 | -0.6657 | -0.6341 |
|
101 |
+
| 0.002 | 1.43 | 4200 | 0.0026 | -0.6449 | -26.1156 | 0.9992 | 25.4707 | -676.0200 | -224.6045 | -0.6907 | -0.6546 |
|
102 |
+
| 0.0007 | 1.46 | 4300 | 0.0026 | -0.4135 | -25.3743 | 0.9992 | 24.9609 | -668.6074 | -222.2907 | -0.6704 | -0.6348 |
|
103 |
+
| 0.001 | 1.5 | 4400 | 0.0025 | -0.1706 | -25.4135 | 0.9992 | 25.2428 | -668.9984 | -219.8623 | -0.6670 | -0.6312 |
|
104 |
+
| 0.0018 | 1.53 | 4500 | 0.0026 | -0.3368 | -23.9768 | 0.9992 | 23.6400 | -654.6318 | -221.5240 | -0.6866 | -0.6345 |
|
105 |
+
| 0.0035 | 1.56 | 4600 | 0.0025 | 0.0146 | -23.9455 | 0.9992 | 23.9602 | -654.3195 | -218.0095 | -0.6725 | -0.6253 |
|
106 |
+
| 0.003 | 1.6 | 4700 | 0.0024 | 0.0616 | -23.3292 | 0.9992 | 23.3908 | -648.1558 | -217.5395 | -0.6644 | -0.6168 |
|
107 |
+
| 0.0028 | 1.63 | 4800 | 0.0026 | -0.5134 | -26.9070 | 0.9992 | 26.3937 | -683.9343 | -223.2894 | -0.7161 | -0.6634 |
|
108 |
+
| 0.0047 | 1.67 | 4900 | 0.0025 | -0.0916 | -24.8206 | 0.9992 | 24.7290 | -663.0701 | -219.0718 | -0.6444 | -0.6038 |
|
109 |
+
| 0.0003 | 1.7 | 5000 | 0.0025 | 0.1584 | -23.8425 | 0.9992 | 24.0009 | -653.2887 | -216.5716 | -0.6169 | -0.5785 |
|
110 |
+
| 0.0074 | 1.73 | 5100 | 0.0026 | 0.4581 | -22.1966 | 0.9992 | 22.6546 | -636.8298 | -213.5752 | -0.6477 | -0.5976 |
|
111 |
+
| 0.002 | 1.77 | 5200 | 0.0023 | 0.1663 | -23.7774 | 0.9983 | 23.9437 | -652.6381 | -216.4931 | -0.6778 | -0.6312 |
|
112 |
+
| 0.0005 | 1.8 | 5300 | 0.0021 | 0.0885 | -24.4639 | 0.9983 | 24.5525 | -659.5032 | -217.2705 | -0.6907 | -0.6445 |
|
113 |
+
| 0.0009 | 1.84 | 5400 | 0.0020 | 0.3259 | -23.8153 | 0.9983 | 24.1412 | -653.0168 | -214.8967 | -0.6674 | -0.6177 |
|
114 |
+
| 0.0004 | 1.87 | 5500 | 0.0027 | 0.0547 | -25.4516 | 0.9992 | 25.5063 | -669.3798 | -217.6091 | -0.7239 | -0.6630 |
|
115 |
+
| 0.0078 | 1.9 | 5600 | 0.0027 | -0.2841 | -27.2416 | 0.9992 | 26.9575 | -687.2796 | -220.9968 | -0.7328 | -0.6718 |
|
116 |
+
| 0.0053 | 1.94 | 5700 | 0.0031 | 0.3394 | -23.3205 | 0.9992 | 23.6599 | -648.0685 | -214.7619 | -0.7018 | -0.6326 |
|
117 |
+
| 0.0028 | 1.97 | 5800 | 0.0022 | 0.3456 | -23.6389 | 0.9992 | 23.9845 | -651.2528 | -214.7000 | -0.6865 | -0.6247 |
|
118 |
+
| 0.0003 | 2.01 | 5900 | 0.0022 | 0.0137 | -25.1376 | 0.9992 | 25.1513 | -666.2399 | -218.0188 | -0.7179 | -0.6544 |
|
119 |
+
| 0.0003 | 2.04 | 6000 | 0.0022 | -0.0273 | -25.5899 | 0.9992 | 25.5627 | -670.7634 | -218.4287 | -0.7175 | -0.6559 |
|
120 |
+
| 0.0005 | 2.07 | 6100 | 0.0021 | -0.0506 | -26.3022 | 0.9992 | 26.2516 | -677.8860 | -218.6621 | -0.7035 | -0.6425 |
|
121 |
+
| 0.0033 | 2.11 | 6200 | 0.0020 | -0.1977 | -27.1231 | 0.9992 | 26.9254 | -686.0947 | -220.1329 | -0.6936 | -0.6406 |
|
122 |
+
| 0.0048 | 2.14 | 6300 | 0.0018 | 0.1836 | -25.2467 | 0.9992 | 25.4303 | -667.3306 | -216.3201 | -0.6888 | -0.6298 |
|
123 |
+
| 0.0007 | 2.18 | 6400 | 0.0018 | -0.0446 | -26.3563 | 0.9992 | 26.3117 | -678.4270 | -218.6022 | -0.7075 | -0.6494 |
|
124 |
+
| 0.0003 | 2.21 | 6500 | 0.0018 | -0.1020 | -27.0392 | 0.9992 | 26.9372 | -685.2560 | -219.1755 | -0.7007 | -0.6418 |
|
125 |
+
| 0.0013 | 2.24 | 6600 | 0.0017 | -0.0434 | -26.1507 | 0.9992 | 26.1073 | -676.3707 | -218.5897 | -0.7076 | -0.6401 |
|
126 |
+
| 0.0002 | 2.28 | 6700 | 0.0018 | 0.1488 | -25.4695 | 0.9992 | 25.6182 | -669.5585 | -216.6682 | -0.6911 | -0.6184 |
|
127 |
+
| 0.0003 | 2.31 | 6800 | 0.0018 | -0.0762 | -26.7830 | 0.9992 | 26.7068 | -682.6938 | -218.9181 | -0.7238 | -0.6530 |
|
128 |
+
| 0.0095 | 2.35 | 6900 | 0.0018 | -0.2520 | -27.9261 | 0.9992 | 27.6741 | -694.1253 | -220.6760 | -0.7267 | -0.6572 |
|
129 |
+
| 0.0012 | 2.38 | 7000 | 0.0017 | -0.1979 | -27.7144 | 0.9992 | 27.5165 | -692.0080 | -220.1350 | -0.7207 | -0.6516 |
|
130 |
+
| 0.0004 | 2.41 | 7100 | 0.0017 | -0.2063 | -28.2831 | 0.9992 | 28.0768 | -697.6947 | -220.2186 | -0.7147 | -0.6448 |
|
131 |
+
| 0.0002 | 2.45 | 7200 | 0.0017 | -0.2423 | -28.5426 | 0.9992 | 28.3004 | -700.2905 | -220.5785 | -0.7291 | -0.6572 |
|
132 |
+
| 0.0049 | 2.48 | 7300 | 0.0017 | -0.0938 | -27.3084 | 0.9992 | 27.2146 | -687.9479 | -219.0937 | -0.7313 | -0.6487 |
|
133 |
+
| 0.0024 | 2.52 | 7400 | 0.0016 | -0.0596 | -27.3730 | 0.9992 | 27.3134 | -688.5939 | -218.7520 | -0.7289 | -0.6467 |
|
134 |
+
| 0.0013 | 2.55 | 7500 | 0.0016 | 0.0102 | -27.3445 | 0.9992 | 27.3547 | -688.3093 | -218.0539 | -0.7271 | -0.6462 |
|
135 |
+
| 0.0014 | 2.58 | 7600 | 0.0016 | -0.1696 | -28.7332 | 0.9992 | 28.5636 | -702.1956 | -219.8516 | -0.7393 | -0.6604 |
|
136 |
+
| 0.0002 | 2.62 | 7700 | 0.0015 | -0.1083 | -28.2952 | 0.9992 | 28.1869 | -697.8158 | -219.2384 | -0.7264 | -0.6502 |
|
137 |
+
| 0.0001 | 2.65 | 7800 | 0.0015 | -0.0892 | -28.2958 | 0.9992 | 28.2066 | -697.8219 | -219.0480 | -0.7246 | -0.6479 |
|
138 |
+
| 0.0025 | 2.69 | 7900 | 0.0015 | -0.1066 | -28.4335 | 0.9992 | 28.3270 | -699.1990 | -219.2214 | -0.7196 | -0.6447 |
|
139 |
+
| 0.0002 | 2.72 | 8000 | 0.0015 | -0.1453 | -28.8184 | 0.9992 | 28.6731 | -703.0482 | -219.6090 | -0.7264 | -0.6518 |
|
140 |
+
| 0.0002 | 2.75 | 8100 | 0.0015 | -0.1058 | -28.6632 | 0.9992 | 28.5575 | -701.4964 | -219.2135 | -0.7190 | -0.6438 |
|
141 |
+
| 0.0003 | 2.79 | 8200 | 0.0015 | -0.1407 | -28.8865 | 0.9992 | 28.7459 | -703.7291 | -219.5624 | -0.7227 | -0.6488 |
|
142 |
+
| 0.0002 | 2.82 | 8300 | 0.0014 | -0.1528 | -28.9232 | 0.9992 | 28.7704 | -704.0963 | -219.6839 | -0.7272 | -0.6534 |
|
143 |
+
| 0.0001 | 2.86 | 8400 | 0.0013 | -0.1196 | -28.7573 | 0.9992 | 28.6377 | -702.4371 | -219.3522 | -0.7244 | -0.6492 |
|
144 |
+
| 0.0003 | 2.89 | 8500 | 0.0013 | -0.1542 | -28.9522 | 0.9992 | 28.7980 | -704.3861 | -219.6977 | -0.7276 | -0.6518 |
|
145 |
+
| 0.0016 | 2.92 | 8600 | 0.0013 | -0.0885 | -28.6082 | 0.9992 | 28.5197 | -700.9456 | -219.0408 | -0.7181 | -0.6426 |
|
146 |
+
| 0.0014 | 2.96 | 8700 | 0.0013 | -0.0904 | -28.5887 | 0.9992 | 28.4983 | -700.7510 | -219.0594 | -0.7190 | -0.6429 |
|
147 |
+
| 0.0005 | 2.99 | 8800 | 0.0013 | -0.0898 | -28.5857 | 0.9992 | 28.4959 | -700.7208 | -219.0538 | -0.7194 | -0.6430 |
|
148 |
+
|
149 |
+
|
150 |
+
### Framework versions
|
151 |
+
|
152 |
+
- Transformers 4.35.0
|
153 |
+
- Pytorch 2.1.1+cu121
|
154 |
+
- Datasets 2.14.6
|
155 |
+
- Tokenizers 0.14.1
|
all_results.json
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 3.0,
|
3 |
+
"eval_logits/chosen": -0.6432551741600037,
|
4 |
+
"eval_logits/rejected": -0.7195827960968018,
|
5 |
+
"eval_logps/chosen": -219.0950927734375,
|
6 |
+
"eval_logps/rejected": -700.8997192382812,
|
7 |
+
"eval_loss": 0.0013098448980599642,
|
8 |
+
"eval_rewards/accuracies": 0.9991582632064819,
|
9 |
+
"eval_rewards/chosen": -0.09392862766981125,
|
10 |
+
"eval_rewards/margins": 28.50965118408203,
|
11 |
+
"eval_rewards/rejected": -28.603578567504883,
|
12 |
+
"eval_runtime": 539.0442,
|
13 |
+
"eval_samples": 9500,
|
14 |
+
"eval_samples_per_second": 17.624,
|
15 |
+
"eval_steps_per_second": 0.551,
|
16 |
+
"train_loss": 0.029375145110450352,
|
17 |
+
"train_runtime": 110770.2779,
|
18 |
+
"train_samples": 188284,
|
19 |
+
"train_samples_per_second": 5.099,
|
20 |
+
"train_steps_per_second": 0.08
|
21 |
+
}
|
config.json
ADDED
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/media/d1/huggingface.co/models/meta-llama/Llama-2-7b-chat-hf",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"bos_token_id": 1,
|
8 |
+
"eos_token_id": 2,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 4096,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 11008,
|
13 |
+
"max_position_embeddings": 4096,
|
14 |
+
"model_type": "llama",
|
15 |
+
"num_attention_heads": 32,
|
16 |
+
"num_hidden_layers": 32,
|
17 |
+
"num_key_value_heads": 32,
|
18 |
+
"pretraining_tp": 1,
|
19 |
+
"rms_norm_eps": 1e-05,
|
20 |
+
"rope_scaling": null,
|
21 |
+
"rope_theta": 10000.0,
|
22 |
+
"tie_word_embeddings": false,
|
23 |
+
"torch_dtype": "bfloat16",
|
24 |
+
"transformers_version": "4.35.0",
|
25 |
+
"use_cache": false,
|
26 |
+
"vocab_size": 32000
|
27 |
+
}
|
eval_results.json
ADDED
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 3.0,
|
3 |
+
"eval_logits/chosen": -0.6432551741600037,
|
4 |
+
"eval_logits/rejected": -0.7195827960968018,
|
5 |
+
"eval_logps/chosen": -219.0950927734375,
|
6 |
+
"eval_logps/rejected": -700.8997192382812,
|
7 |
+
"eval_loss": 0.0013098448980599642,
|
8 |
+
"eval_rewards/accuracies": 0.9991582632064819,
|
9 |
+
"eval_rewards/chosen": -0.09392862766981125,
|
10 |
+
"eval_rewards/margins": 28.50965118408203,
|
11 |
+
"eval_rewards/rejected": -28.603578567504883,
|
12 |
+
"eval_runtime": 539.0442,
|
13 |
+
"eval_samples": 9500,
|
14 |
+
"eval_samples_per_second": 17.624,
|
15 |
+
"eval_steps_per_second": 0.551
|
16 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token_id": 1,
|
3 |
+
"do_sample": true,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"max_length": 4096,
|
6 |
+
"pad_token_id": 0,
|
7 |
+
"temperature": 0.6,
|
8 |
+
"top_p": 0.9,
|
9 |
+
"transformers_version": "4.35.0"
|
10 |
+
}
|
model-00001-of-00003.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:652fc7bd7091184305e4dfd2e8e30a1178bf5ef7190e75fb91a3370e0e4f7603
|
3 |
+
size 4938985352
|
model-00002-of-00003.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6cd74a276ea84263a4753a44473bbbe5ed196bff538f69cf8016c6f64fe4712b
|
3 |
+
size 4947390880
|
model-00003-of-00003.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c03af98920dcd4aadfa44c03ade096c698b9377d41e2bc138e783501e76c8aef
|
3 |
+
size 3590488816
|
model.safetensors.index.json
ADDED
@@ -0,0 +1,298 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"metadata": {
|
3 |
+
"total_size": 13476831232
|
4 |
+
},
|
5 |
+
"weight_map": {
|
6 |
+
"lm_head.weight": "model-00003-of-00003.safetensors",
|
7 |
+
"model.embed_tokens.weight": "model-00001-of-00003.safetensors",
|
8 |
+
"model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
9 |
+
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
10 |
+
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
11 |
+
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
12 |
+
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
13 |
+
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
14 |
+
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
15 |
+
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
16 |
+
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
17 |
+
"model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
18 |
+
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
19 |
+
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
20 |
+
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
21 |
+
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
22 |
+
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
23 |
+
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
24 |
+
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
25 |
+
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
26 |
+
"model.layers.10.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
27 |
+
"model.layers.10.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
28 |
+
"model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
29 |
+
"model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
30 |
+
"model.layers.10.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
31 |
+
"model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
32 |
+
"model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
33 |
+
"model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
34 |
+
"model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
35 |
+
"model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
36 |
+
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
37 |
+
"model.layers.11.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
38 |
+
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
39 |
+
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
40 |
+
"model.layers.11.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
41 |
+
"model.layers.11.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
42 |
+
"model.layers.11.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
43 |
+
"model.layers.11.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
44 |
+
"model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
45 |
+
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
46 |
+
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
47 |
+
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
48 |
+
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
49 |
+
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
50 |
+
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
51 |
+
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
52 |
+
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
53 |
+
"model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
54 |
+
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
55 |
+
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
56 |
+
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
57 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
58 |
+
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
59 |
+
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
60 |
+
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
61 |
+
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
62 |
+
"model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
63 |
+
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
64 |
+
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
65 |
+
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
66 |
+
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
67 |
+
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
68 |
+
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
69 |
+
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
70 |
+
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
71 |
+
"model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
72 |
+
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
73 |
+
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
74 |
+
"model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
75 |
+
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
76 |
+
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
77 |
+
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
78 |
+
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
79 |
+
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
80 |
+
"model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
81 |
+
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
82 |
+
"model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
83 |
+
"model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
84 |
+
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
85 |
+
"model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
86 |
+
"model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
87 |
+
"model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
88 |
+
"model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
89 |
+
"model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
90 |
+
"model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
91 |
+
"model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
92 |
+
"model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
93 |
+
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
94 |
+
"model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
95 |
+
"model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
96 |
+
"model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
97 |
+
"model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
98 |
+
"model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
99 |
+
"model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
100 |
+
"model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
101 |
+
"model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
102 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
103 |
+
"model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
104 |
+
"model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
105 |
+
"model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
106 |
+
"model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
107 |
+
"model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
108 |
+
"model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
109 |
+
"model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
110 |
+
"model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
111 |
+
"model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
112 |
+
"model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
113 |
+
"model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
114 |
+
"model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
115 |
+
"model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
116 |
+
"model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
117 |
+
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
118 |
+
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
119 |
+
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
120 |
+
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
121 |
+
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
122 |
+
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
123 |
+
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
124 |
+
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
125 |
+
"model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
126 |
+
"model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
127 |
+
"model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
128 |
+
"model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
129 |
+
"model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
130 |
+
"model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
131 |
+
"model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
132 |
+
"model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
133 |
+
"model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
134 |
+
"model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
135 |
+
"model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
136 |
+
"model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
137 |
+
"model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
138 |
+
"model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
139 |
+
"model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
140 |
+
"model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
141 |
+
"model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
142 |
+
"model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
143 |
+
"model.layers.22.input_layernorm.weight": "model-00002-of-00003.safetensors",
|
144 |
+
"model.layers.22.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
|
145 |
+
"model.layers.22.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
146 |
+
"model.layers.22.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
147 |
+
"model.layers.22.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
|
148 |
+
"model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
149 |
+
"model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
150 |
+
"model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
151 |
+
"model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
152 |
+
"model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
153 |
+
"model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
154 |
+
"model.layers.23.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
|
155 |
+
"model.layers.23.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
|
156 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
157 |
+
"model.layers.23.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
|
158 |
+
"model.layers.23.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
|
159 |
+
"model.layers.23.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
160 |
+
"model.layers.23.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
|
161 |
+
"model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
162 |
+
"model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
163 |
+
"model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
164 |
+
"model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
165 |
+
"model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
166 |
+
"model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
167 |
+
"model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
168 |
+
"model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
169 |
+
"model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
170 |
+
"model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
171 |
+
"model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
172 |
+
"model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
173 |
+
"model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
174 |
+
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
175 |
+
"model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
176 |
+
"model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
177 |
+
"model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
178 |
+
"model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
179 |
+
"model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
180 |
+
"model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
181 |
+
"model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
182 |
+
"model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
183 |
+
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
184 |
+
"model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
185 |
+
"model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
186 |
+
"model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
187 |
+
"model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
188 |
+
"model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
189 |
+
"model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
190 |
+
"model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
191 |
+
"model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
192 |
+
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
193 |
+
"model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
194 |
+
"model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
195 |
+
"model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
196 |
+
"model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
197 |
+
"model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
198 |
+
"model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
199 |
+
"model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
200 |
+
"model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
201 |
+
"model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
202 |
+
"model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
203 |
+
"model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
204 |
+
"model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
205 |
+
"model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
206 |
+
"model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
207 |
+
"model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
208 |
+
"model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
209 |
+
"model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
210 |
+
"model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
211 |
+
"model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
212 |
+
"model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
213 |
+
"model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
214 |
+
"model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
215 |
+
"model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
216 |
+
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
217 |
+
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
218 |
+
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
219 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
220 |
+
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
221 |
+
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
222 |
+
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
223 |
+
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
224 |
+
"model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
225 |
+
"model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
226 |
+
"model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
227 |
+
"model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
228 |
+
"model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
229 |
+
"model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
230 |
+
"model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
231 |
+
"model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
232 |
+
"model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
233 |
+
"model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
|
234 |
+
"model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
|
235 |
+
"model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
|
236 |
+
"model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
|
237 |
+
"model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
|
238 |
+
"model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
239 |
+
"model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
|
240 |
+
"model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
|
241 |
+
"model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
|
242 |
+
"model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
243 |
+
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
244 |
+
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
245 |
+
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
246 |
+
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
247 |
+
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
248 |
+
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
249 |
+
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
250 |
+
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
251 |
+
"model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
252 |
+
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
253 |
+
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
254 |
+
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
255 |
+
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
256 |
+
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
257 |
+
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
258 |
+
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
259 |
+
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
260 |
+
"model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
261 |
+
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
262 |
+
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
263 |
+
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
264 |
+
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
265 |
+
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
266 |
+
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
267 |
+
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
268 |
+
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
269 |
+
"model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
270 |
+
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
271 |
+
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
272 |
+
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
273 |
+
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
274 |
+
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
275 |
+
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
276 |
+
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
277 |
+
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
278 |
+
"model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
279 |
+
"model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
280 |
+
"model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
281 |
+
"model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
282 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
283 |
+
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
284 |
+
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
285 |
+
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
286 |
+
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
287 |
+
"model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
|
288 |
+
"model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
|
289 |
+
"model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
|
290 |
+
"model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
|
291 |
+
"model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
|
292 |
+
"model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
293 |
+
"model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
|
294 |
+
"model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
|
295 |
+
"model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
|
296 |
+
"model.norm.weight": "model-00003-of-00003.safetensors"
|
297 |
+
}
|
298 |
+
}
|
runs/Jan14_05-17-19_amaterasu/events.out.tfevents.1705209484.amaterasu.1284454.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6fc1dfbfe6dc44b4da406e5f961123b274f4ab59bd640ccd87e3b0964df1b1e
|
3 |
+
size 629555
|
runs/Jan14_05-17-19_amaterasu/events.out.tfevents.1705320794.amaterasu.1284454.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2b99fffc0eeb42f5f722641295159ff60f1901cfac06d303d2d85c673dd00f53
|
3 |
+
size 828
|
special_tokens_map.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": "</s>",
|
17 |
+
"unk_token": {
|
18 |
+
"content": "<unk>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
}
|
24 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "<unk>",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "<s>",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "</s>",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
}
|
27 |
+
},
|
28 |
+
"bos_token": "<s>",
|
29 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<<SYS>>\\n' + system_message + '\\n<</SYS>>\\n\\n' + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content.strip() + ' ' + eos_token }}{% endif %}{% endfor %}",
|
30 |
+
"clean_up_tokenization_spaces": false,
|
31 |
+
"eos_token": "</s>",
|
32 |
+
"legacy": false,
|
33 |
+
"model_max_length": 2048,
|
34 |
+
"pad_token": "</s>",
|
35 |
+
"padding_side": "right",
|
36 |
+
"sp_model_kwargs": {},
|
37 |
+
"tokenizer_class": "LlamaTokenizer",
|
38 |
+
"unk_token": "<unk>",
|
39 |
+
"use_default_system_prompt": false
|
40 |
+
}
|
train_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 3.0,
|
3 |
+
"train_loss": 0.029375145110450352,
|
4 |
+
"train_runtime": 110770.2779,
|
5 |
+
"train_samples": 188284,
|
6 |
+
"train_samples_per_second": 5.099,
|
7 |
+
"train_steps_per_second": 0.08
|
8 |
+
}
|
trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53e4faf8feeb2652608b9454c8ed733658b81937eb366bd52ed00cdfc4d717d4
|
3 |
+
size 5752
|