Update README.md
Browse files
README.md
CHANGED
@@ -5,4 +5,11 @@ datasets:
|
|
5 |
language:
|
6 |
- en
|
7 |
pipeline_tag: token-classification
|
8 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
language:
|
6 |
- en
|
7 |
pipeline_tag: token-classification
|
8 |
+
---
|
9 |
+
Another name for this model is sciDeBERta v2[1].
|
10 |
+
This model is trained from scratch using S2ORC dataset(260GB), which include abstract, body text of papers, on DeBERTa v2.
|
11 |
+
This model achieves the SOTA in NET of SciERC dataset.
|
12 |
+
From this model, MediBioDeBERTa, which continuously leaned from scidebert v2. to medibiodeberta using the data from the domain (bio, medical, chemistry domain data)
|
13 |
+
and additional intermediate fine-tuning for specific blurb benchmark tasks, achieve the 11 rank in the BLURB benchmark.
|
14 |
+
|
15 |
+
[1] Eunhui Kim, Yuna Jeong, Myung-seok Choi, "MediBioDeBERTa: BioMedical Language Model with Continous Learning and Intermediate Fine-Tuning, Dec. 2023, IEEE Access"
|