--- license: gpl-3.0 language: - en metrics: - accuracy base_model: facebook/bart-large --- # Model Card for ANGEL_pretrained This model card provides detailed information about the ANGEL_pretrained model, designed for biomedical entity linking. # Model Details #### Model Description - **Developed by:** Chanhwi Kim, Hyunjae Kim, Sihyeon Park, Jiwoo Lee, Mujeen Sung, Jaewoo Kang - **Model type:** Generative Biomedical Entity Linking Model - **Language(s):** English - **License:** GPL-3.0 - **Finetuned from model:** BART-large (Base architecture) #### Model Sources - **Repository:** https://github.com/dmis-lab/ANGEL - **Paper:** https://arxiv.org/pdf/2408.16493 # Direct Use ANGEL_pretrained is pretrained with UMLS dataset. We recommand to finetune this model to downstream dataset rather directly use. If you still want to run the model on a single sample, no preprocessing is required. Simply execute the run_sample.sh script: ```bash bash script/inference/run_sample.sh pretrained ``` To modify the sample with your own example, refer to the [Direct Use](https://github.com/dmis-lab/ANGEL?tab=readme-ov-file#direct-use) section in our GitHub repository. If you're interested in training or evaluating the model, check out the [Fine-tuning](https://github.com/dmis-lab/ANGEL?tab=readme-ov-file#fine-tuning) section and [Evaluation](https://github.com/dmis-lab/ANGEL?tab=readme-ov-file#evaluation) section. # Training Details #### Training Data The model was pretrained on the UMLS-2020-AA dataset. #### Training Procedure Positive-only Pre-training: Initial training using only positive examples, following the standard approach. Negative-aware Training: Subsequent training incorporated negative examples to improve the model's discriminative capabilities. # Evaluation #### Testing Data The model was evaluated using multiple biomedical datasets, including NCBI-disease, BC5CDR, COMETA, AAP, and MedMentions. The fine-tuned scores have also been included. #### Metrics **Accuracy at Top-1 (Acc@1)**: Measures the percentage of times the model's top prediction matches the correct entity. ### Results
Model | NCBI-disease | BC5CDR | COMETA | AAP | MedMentions ST21pv |
Average |
---|---|---|---|---|---|---|
GenBioEL_pretrained | 58.2 | 33.1 | 42.4 | 50.6 | 10.6 | 39.0 |
ANGEL_pretrained (Ours) | 64.6 | 49.7 | 46.8 | 61.5 | 18.2 | 48.2 |
GenBioEL_pt_ft | 91.0 | 93.1 | 80.9 | 89.3 | 70.7 | 85.0 |
ANGEL_pt_ft (Ours) | 92.8 | 94.5 | 82.8 | 90.2 | 73.3 | 86.7 |