MikkoLipsanen
commited on
Commit
·
9ef1847
1
Parent(s):
2ca67a2
Update README.md
Browse files
README.md
CHANGED
@@ -86,12 +86,13 @@ Test|2414|5577|179|2445|1097|183|2838|272|374|356
|
|
86 |
This model was trained using a NVIDIA RTX A6000 GPU with the following hyperparameters:
|
87 |
|
88 |
- learning rate: 2e-05
|
89 |
-
- train batch size:
|
90 |
- epochs: 10
|
91 |
- optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
|
92 |
- scheduler: linear scheduler with num_warmup_steps=round(len(train_dataloader)/5) and num_training_steps=len(train_dataloader)*epochs
|
93 |
- maximum length of data sequence: 512
|
94 |
- patience: 2 epochs
|
|
|
95 |
|
96 |
In the preprocessing stage, the input texts were split into chunks with a maximum length of 300 tokens,
|
97 |
in order to avoid the tokenized chunks exceeding the maximum length of 512. Tokenization was performed
|
@@ -106,15 +107,15 @@ Evaluation results using the test dataset are listed below:
|
|
106 |
|
107 |
||Precision|Recall|F1-score
|
108 |
-|-|-|-
|
109 |
-
PERSON|0.
|
110 |
-
ORG|0.
|
111 |
-
LOC|0.
|
112 |
-
GPE|0.
|
113 |
-
PRODUCT|0.
|
114 |
-
EVENT|0.
|
115 |
-
DATE|0.
|
116 |
-
JON|0.
|
117 |
-
FIBC|0.
|
118 |
NORP|0.91|0.95|0.93
|
119 |
|
120 |
The metrics were calculated using the [seqeval](https://github.com/chakki-works/seqeval) library.
|
|
|
86 |
This model was trained using a NVIDIA RTX A6000 GPU with the following hyperparameters:
|
87 |
|
88 |
- learning rate: 2e-05
|
89 |
+
- train batch size: 24
|
90 |
- epochs: 10
|
91 |
- optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
|
92 |
- scheduler: linear scheduler with num_warmup_steps=round(len(train_dataloader)/5) and num_training_steps=len(train_dataloader)*epochs
|
93 |
- maximum length of data sequence: 512
|
94 |
- patience: 2 epochs
|
95 |
+
- classifier dropout: 0.3
|
96 |
|
97 |
In the preprocessing stage, the input texts were split into chunks with a maximum length of 300 tokens,
|
98 |
in order to avoid the tokenized chunks exceeding the maximum length of 512. Tokenization was performed
|
|
|
107 |
|
108 |
||Precision|Recall|F1-score
|
109 |
-|-|-|-
|
110 |
+
PERSON|0.90|0.91|0.90
|
111 |
+
ORG|0.84|0.87|0.86
|
112 |
+
LOC|0.84|0.86|0.85
|
113 |
+
GPE|0.91|0.91|0.91
|
114 |
+
PRODUCT|0.73|0.77|0.75
|
115 |
+
EVENT|0.69|0.73|0.71
|
116 |
+
DATE|0.90|0.92|0.91
|
117 |
+
JON|0.83|0.95|0.89
|
118 |
+
FIBC|0.95|0.99|0.97
|
119 |
NORP|0.91|0.95|0.93
|
120 |
|
121 |
The metrics were calculated using the [seqeval](https://github.com/chakki-works/seqeval) library.
|