metadata
datasets:
- BI55/MedText
metrics:
- bleu
- perplexity
pipeline_tag: text-generation
A further fine-tuned version of Locutusque/gpt2-large-conversational on MedText
Evaluation
This model was evaluated using GPT-3.5, and it was asked medical questions. It achieved an average accuracy of 75.8%.