|
--- |
|
license: cc-by-nc-4.0 |
|
extra_gated_prompt: >- |
|
You agree not to use the model for healthcare decision-making or commercial |
|
use |
|
extra_gated_fields: |
|
I agree to use this model for non-commercial use ONLY: checkbox |
|
I agree not to use this model for healthcare decision-making: checkbox |
|
language: |
|
- en |
|
--- |
|
|
|
# Clinical Camel |
|
|
|
## Model Description |
|
|
|
Clinical Camel is an open large language model (LLM), fine-tuned on the LLaMA-2 70B architecture using [QLoRA](https://github.com/artidoro/qlora). It is tailored for the medical and clinical research, capable of processing and generating relevant content. |
|
|
|
Review our pre-print for more details: [Clinical Camel - Pre-print](https://arxiv.org/abs/2305.12031) |
|
|
|
## Performance |
|
|
|
Clinical Camel demonstrates competitive performance on medical benchmarks. |
|
|
|
**Table: Five-Shot Performance of Clinical Camel-70B (C70), GPT3.5, GPT4, and Med-PaLM 2 on Various Medical Datasets** |
|
|
|
| Dataset | ClinicalCamel-70B | GPT3.5 | GPT4 | Med-PaLM 2 | |
|
|-----------------------------|-------------|--------|-------|--------------| |
|
| MMLU Anatomy | 65.2 | 60.7 | 80.0 | 77.8 | |
|
| MMLU Clinical Knowledge | 72.8 | 68.7 | 86.4 | 88.3 | |
|
| MMLU College Biology | 81.2 | 72.9 | 93.8 | 94.4 | |
|
| MMLU College Medicine | 68.2 | 63.6 | 76.3 | 80.9 | |
|
| MMLU Medical Genetics | 69.0 | 68.0 | 92.0 | 90.0 | |
|
| MMLU Professional Medicine | 75.0 | 69.8 | 93.8 | 95.2 | |
|
| MedMCQA | 54.2 | 51.0 | 72.4 | 71.3 | |
|
| MedQA (USMLE) | 60.7 | 53.6 | 81.4 | 79.7 | |
|
| PubMedQA | 77.9 | 60.2 | 74.4 | 79.2 | |
|
| USMLE Sample Exam | 64.3 | 58.5 | 86.6 | - | |
|
|
|
## Evaluation Datasets: |
|
|
|
The performance of Clinical Camel was benchmarked across several datasets, including: |
|
- [USMLE Step 1](https://huggingface.co/datasets/augtoma/usmle_step_1) |
|
- [USMLE Step 2](https://huggingface.co/datasets/augtoma/usmle_step_2) |
|
- [USMLE Step 3](https://huggingface.co/datasets/augtoma/usmle_step_3) |
|
- [MedMCQA](https://huggingface.co/datasets/augtoma/medmcqa) |
|
- [MedQA USMLE](https://huggingface.co/datasets/augtoma/medqa_usmle) |
|
|
|
## Evaluation Reproduction: |
|
|
|
To reproduce the evaluations with [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) see the 'TaskFiles' folder |