|
--- |
|
language: |
|
- "en" |
|
license: mit |
|
tags: |
|
- fill-mask |
|
--- |
|
|
|
|
|
# MedBERT Model |
|
|
|
MedBERT is a newly pre-trained transformer-based language model for biomedical named entity recognition: initialised with Bio_ClinicalBERT & pre-trained on N2C2, BioNLP and CRAFT community datasets. |
|
|
|
|
|
## How to use |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModel |
|
tokenizer = AutoTokenizer.from_pretrained("Charangan/MedBERT") |
|
model = AutoModel.from_pretrained("Charangan/MedBERT") |
|
``` |
|
|
|
|
|
## Citation |
|
``` |
|
@inproceedings{medbert, |
|
title = {{MedBERT: A Pre-Trained Language Model for Biomedical Named Entity Recognition}}, |
|
author = {Charangan Vasantharajan and Kyaw Zin Tun and Ho Thi-Nga and Sparsh Jain and Tong Rong and Chng Eng Siong}, |
|
booktitle = {Asia Pacific Signal and Information Processing Association Annual Summit and Conference 2022}, |
|
year = {2022}, |
|
month = {November} |
|
} |
|
``` |