xshubhamx commited on
Commit
b51c916
·
verified ·
1 Parent(s): 0550376

End of training

Browse files
Files changed (2) hide show
  1. README.md +104 -0
  2. adapter_model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-sa-4.0
3
+ base_model: nlpaueb/legal-bert-base-uncased
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - precision
9
+ - recall
10
+ model-index:
11
+ - name: legal-bert-lora-no-grad
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # legal-bert-lora-no-grad
19
+
20
+ This model is a fine-tuned version of [nlpaueb/legal-bert-base-uncased](https://huggingface.co/nlpaueb/legal-bert-base-uncased) on an unknown dataset.
21
+ It achieves the following results on the evaluation set:
22
+ - Loss: 1.5803
23
+ - Accuracy: 0.8234
24
+ - Precision: 0.8234
25
+ - Recall: 0.8234
26
+ - Precision Macro: 0.7941
27
+ - Recall Macro: 0.7774
28
+ - Macro Fpr: 0.0157
29
+ - Weighted Fpr: 0.0151
30
+ - Weighted Specificity: 0.9767
31
+ - Macro Specificity: 0.9867
32
+ - Weighted Sensitivity: 0.8234
33
+ - Macro Sensitivity: 0.7774
34
+ - F1 Micro: 0.8234
35
+ - F1 Macro: 0.7824
36
+ - F1 Weighted: 0.8225
37
+
38
+ ## Model description
39
+
40
+ More information needed
41
+
42
+ ## Intended uses & limitations
43
+
44
+ More information needed
45
+
46
+ ## Training and evaluation data
47
+
48
+ More information needed
49
+
50
+ ## Training procedure
51
+
52
+ ### Training hyperparameters
53
+
54
+ The following hyperparameters were used during training:
55
+ - learning_rate: 5e-05
56
+ - train_batch_size: 8
57
+ - eval_batch_size: 8
58
+ - seed: 42
59
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
+ - lr_scheduler_type: linear
61
+ - num_epochs: 30
62
+
63
+ ### Training results
64
+
65
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | Precision Macro | Recall Macro | Macro Fpr | Weighted Fpr | Weighted Specificity | Macro Specificity | Weighted Sensitivity | Macro Sensitivity | F1 Micro | F1 Macro | F1 Weighted |
66
+ |:-------------:|:-----:|:-----:|:---------------:|:--------:|:---------:|:------:|:---------------:|:------------:|:---------:|:------------:|:--------------------:|:-----------------:|:--------------------:|:-----------------:|:--------:|:--------:|:-----------:|
67
+ | 1.5473 | 1.0 | 643 | 0.8485 | 0.7173 | 0.6897 | 0.7173 | 0.3892 | 0.4422 | 0.0278 | 0.0274 | 0.9658 | 0.9789 | 0.7173 | 0.4422 | 0.7173 | 0.3920 | 0.6869 |
68
+ | 0.7816 | 2.0 | 1286 | 0.7113 | 0.7545 | 0.7492 | 0.7545 | 0.5282 | 0.5054 | 0.0231 | 0.0227 | 0.9711 | 0.9817 | 0.7545 | 0.5054 | 0.7545 | 0.4751 | 0.7304 |
69
+ | 0.6956 | 3.0 | 1929 | 0.6460 | 0.7986 | 0.7793 | 0.7986 | 0.5436 | 0.5701 | 0.0184 | 0.0177 | 0.9741 | 0.9848 | 0.7986 | 0.5701 | 0.7986 | 0.5439 | 0.7833 |
70
+ | 0.4942 | 4.0 | 2572 | 0.6430 | 0.8110 | 0.8014 | 0.8110 | 0.6315 | 0.6670 | 0.0169 | 0.0164 | 0.9763 | 0.9858 | 0.8110 | 0.6670 | 0.8110 | 0.6413 | 0.8037 |
71
+ | 0.4088 | 5.0 | 3215 | 0.7148 | 0.8319 | 0.8301 | 0.8319 | 0.7951 | 0.7444 | 0.0150 | 0.0142 | 0.9758 | 0.9872 | 0.8319 | 0.7444 | 0.8319 | 0.7459 | 0.8254 |
72
+ | 0.3722 | 6.0 | 3858 | 0.7203 | 0.8319 | 0.8282 | 0.8319 | 0.7594 | 0.7604 | 0.0149 | 0.0142 | 0.9779 | 0.9873 | 0.8319 | 0.7604 | 0.8319 | 0.7582 | 0.8285 |
73
+ | 0.3088 | 7.0 | 4501 | 0.7796 | 0.8218 | 0.8283 | 0.8218 | 0.7927 | 0.7436 | 0.0158 | 0.0152 | 0.9765 | 0.9866 | 0.8218 | 0.7436 | 0.8218 | 0.7486 | 0.8203 |
74
+ | 0.245 | 8.0 | 5144 | 0.8732 | 0.8187 | 0.8171 | 0.8187 | 0.7696 | 0.7403 | 0.0163 | 0.0156 | 0.9752 | 0.9863 | 0.8187 | 0.7403 | 0.8187 | 0.7409 | 0.8135 |
75
+ | 0.2331 | 9.0 | 5787 | 0.8710 | 0.8265 | 0.8280 | 0.8265 | 0.7595 | 0.7521 | 0.0152 | 0.0148 | 0.9776 | 0.9869 | 0.8265 | 0.7521 | 0.8265 | 0.7530 | 0.8261 |
76
+ | 0.1878 | 10.0 | 6430 | 0.9866 | 0.8257 | 0.8261 | 0.8257 | 0.7672 | 0.7604 | 0.0153 | 0.0149 | 0.9788 | 0.9870 | 0.8257 | 0.7604 | 0.8257 | 0.7588 | 0.8240 |
77
+ | 0.1627 | 11.0 | 7073 | 1.0530 | 0.8257 | 0.8269 | 0.8257 | 0.7706 | 0.7711 | 0.0154 | 0.0149 | 0.9787 | 0.9870 | 0.8257 | 0.7711 | 0.8257 | 0.7675 | 0.8234 |
78
+ | 0.1301 | 12.0 | 7716 | 1.1042 | 0.8265 | 0.8246 | 0.8265 | 0.7633 | 0.7587 | 0.0155 | 0.0148 | 0.9763 | 0.9869 | 0.8265 | 0.7587 | 0.8265 | 0.7576 | 0.8228 |
79
+ | 0.1291 | 13.0 | 8359 | 1.1461 | 0.8234 | 0.8215 | 0.8234 | 0.7582 | 0.7613 | 0.0157 | 0.0151 | 0.9768 | 0.9867 | 0.8234 | 0.7613 | 0.8234 | 0.7581 | 0.8211 |
80
+ | 0.11 | 14.0 | 9002 | 1.1837 | 0.8226 | 0.8182 | 0.8226 | 0.7716 | 0.7576 | 0.0159 | 0.0152 | 0.9756 | 0.9865 | 0.8226 | 0.7576 | 0.8226 | 0.7627 | 0.8195 |
81
+ | 0.0863 | 15.0 | 9645 | 1.2020 | 0.8218 | 0.8185 | 0.8218 | 0.7616 | 0.7458 | 0.0160 | 0.0152 | 0.9754 | 0.9865 | 0.8218 | 0.7458 | 0.8218 | 0.7498 | 0.8183 |
82
+ | 0.0735 | 16.0 | 10288 | 1.2491 | 0.8187 | 0.8160 | 0.8187 | 0.7620 | 0.7464 | 0.0162 | 0.0156 | 0.9755 | 0.9863 | 0.8187 | 0.7464 | 0.8187 | 0.7517 | 0.8168 |
83
+ | 0.0802 | 17.0 | 10931 | 1.3288 | 0.8164 | 0.8165 | 0.8164 | 0.7471 | 0.7531 | 0.0164 | 0.0158 | 0.9765 | 0.9862 | 0.8164 | 0.7531 | 0.8164 | 0.7483 | 0.8152 |
84
+ | 0.0525 | 18.0 | 11574 | 1.3620 | 0.8133 | 0.8127 | 0.8133 | 0.7557 | 0.7478 | 0.0168 | 0.0161 | 0.9754 | 0.9859 | 0.8133 | 0.7478 | 0.8133 | 0.7486 | 0.8106 |
85
+ | 0.0474 | 19.0 | 12217 | 1.3783 | 0.8187 | 0.8220 | 0.8187 | 0.8154 | 0.7972 | 0.0162 | 0.0156 | 0.9755 | 0.9863 | 0.8187 | 0.7972 | 0.8187 | 0.8024 | 0.8187 |
86
+ | 0.0315 | 20.0 | 12860 | 1.4004 | 0.8226 | 0.8266 | 0.8226 | 0.8011 | 0.7910 | 0.0157 | 0.0152 | 0.9778 | 0.9867 | 0.8226 | 0.7910 | 0.8226 | 0.7907 | 0.8231 |
87
+ | 0.0325 | 21.0 | 13503 | 1.4683 | 0.8187 | 0.8198 | 0.8187 | 0.8030 | 0.7876 | 0.0161 | 0.0156 | 0.9765 | 0.9863 | 0.8187 | 0.7876 | 0.8187 | 0.7921 | 0.8182 |
88
+ | 0.0192 | 22.0 | 14146 | 1.4677 | 0.8249 | 0.8224 | 0.8249 | 0.7598 | 0.7482 | 0.0155 | 0.0149 | 0.9768 | 0.9868 | 0.8249 | 0.7482 | 0.8249 | 0.7505 | 0.8225 |
89
+ | 0.0235 | 23.0 | 14789 | 1.4610 | 0.8211 | 0.8221 | 0.8211 | 0.8090 | 0.7941 | 0.0160 | 0.0153 | 0.9765 | 0.9865 | 0.8211 | 0.7941 | 0.8211 | 0.7982 | 0.8200 |
90
+ | 0.0142 | 24.0 | 15432 | 1.4787 | 0.8273 | 0.8270 | 0.8273 | 0.8179 | 0.7977 | 0.0153 | 0.0147 | 0.9774 | 0.9870 | 0.8273 | 0.7977 | 0.8273 | 0.8054 | 0.8260 |
91
+ | 0.0172 | 25.0 | 16075 | 1.5374 | 0.8211 | 0.8231 | 0.8211 | 0.7830 | 0.7711 | 0.0159 | 0.0153 | 0.9767 | 0.9865 | 0.8211 | 0.7711 | 0.8211 | 0.7737 | 0.8212 |
92
+ | 0.0097 | 26.0 | 16718 | 1.5153 | 0.8242 | 0.8262 | 0.8242 | 0.8105 | 0.7894 | 0.0156 | 0.0150 | 0.9772 | 0.9868 | 0.8242 | 0.7894 | 0.8242 | 0.7954 | 0.8240 |
93
+ | 0.0109 | 27.0 | 17361 | 1.5565 | 0.8218 | 0.8237 | 0.8218 | 0.7876 | 0.7848 | 0.0158 | 0.0152 | 0.9770 | 0.9866 | 0.8218 | 0.7848 | 0.8218 | 0.7832 | 0.8214 |
94
+ | 0.0076 | 28.0 | 18004 | 1.5574 | 0.8226 | 0.8247 | 0.8226 | 0.7882 | 0.7843 | 0.0157 | 0.0152 | 0.9771 | 0.9867 | 0.8226 | 0.7843 | 0.8226 | 0.7836 | 0.8229 |
95
+ | 0.006 | 29.0 | 18647 | 1.5721 | 0.8234 | 0.8235 | 0.8234 | 0.7868 | 0.7771 | 0.0157 | 0.0151 | 0.9769 | 0.9867 | 0.8234 | 0.7771 | 0.8234 | 0.7788 | 0.8224 |
96
+ | 0.0035 | 30.0 | 19290 | 1.5803 | 0.8234 | 0.8234 | 0.8234 | 0.7941 | 0.7774 | 0.0157 | 0.0151 | 0.9767 | 0.9867 | 0.8234 | 0.7774 | 0.8234 | 0.7824 | 0.8225 |
97
+
98
+
99
+ ### Framework versions
100
+
101
+ - Transformers 4.35.2
102
+ - Pytorch 2.1.0+cu121
103
+ - Datasets 2.18.0
104
+ - Tokenizers 0.15.1
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:488870c093dc4ac86d865be8a00f7c1869150b9ee39f8285d8cdcb1bb83f6d87
3
  size 104549068
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecdfdae9dca393b48b94071f2bd5190415e21c4f18d3acf459a1e52aff85df67
3
  size 104549068