mrm8488 commited on
Commit
13764c0
1 Parent(s): 3e1ccf7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -1
README.md CHANGED
@@ -15,7 +15,7 @@ tags:
15
 
16
  # BioGPT (Large) fine-tuned on ChatDoctor for QA
17
 
18
- Microsoft's BioGPT Large fine-tuned on ChatDoctor dataset for Question Answering.
19
 
20
 
21
  ## Intended Use
@@ -29,12 +29,21 @@ TBA
29
 
30
  ## Model
31
 
 
 
32
  Pre-trained language models have attracted increasing attention in the biomedical domain, inspired by their great success in the general natural language domain. Among the two main branches of pre-trained language models in the general language domain, i.e. BERT (and its variants) and GPT (and its variants), the first one has been extensively studied in the biomedical domain, such as BioBERT and PubMedBERT. While they have achieved great success on a variety of discriminative downstream biomedical tasks, the lack of generation ability constrains their application scope. In this paper, we propose BioGPT, a domain-specific generative Transformer language model pre-trained on large-scale biomedical literature. We evaluate BioGPT on six biomedical natural language processing tasks and demonstrate that our model outperforms previous models on most tasks. Especially, we get 44.98%, 38.42% and 40.76% F1 score on BC5CDR, KD-DTI and DDI end-to-end relation extraction tasks, respectively, and 78.2% accuracy on PubMedQA, creating a new record. Our case study on text generation further demonstrates the advantage of BioGPT on biomedical literature to generate fluent descriptions for biomedical terms.
33
 
34
 
35
  ## Dataset
36
 
37
  ChatDoctor-200K dataset is collected from this paper https://arxiv.org/pdf/2303.14070.pdf
 
 
 
 
 
 
 
38
 
39
 
40
  ## Usage
 
15
 
16
  # BioGPT (Large) fine-tuned on ChatDoctor for QA
17
 
18
+ [Microsoft's BioGPT Large](https://huggingface.co/microsoft/BioGPT-Large) fine-tuned on ChatDoctor dataset for Question Answering.
19
 
20
 
21
  ## Intended Use
 
29
 
30
  ## Model
31
 
32
+ [Microsoft's BioGPT Large](https://huggingface.co/microsoft/BioGPT-Large):
33
+
34
  Pre-trained language models have attracted increasing attention in the biomedical domain, inspired by their great success in the general natural language domain. Among the two main branches of pre-trained language models in the general language domain, i.e. BERT (and its variants) and GPT (and its variants), the first one has been extensively studied in the biomedical domain, such as BioBERT and PubMedBERT. While they have achieved great success on a variety of discriminative downstream biomedical tasks, the lack of generation ability constrains their application scope. In this paper, we propose BioGPT, a domain-specific generative Transformer language model pre-trained on large-scale biomedical literature. We evaluate BioGPT on six biomedical natural language processing tasks and demonstrate that our model outperforms previous models on most tasks. Especially, we get 44.98%, 38.42% and 40.76% F1 score on BC5CDR, KD-DTI and DDI end-to-end relation extraction tasks, respectively, and 78.2% accuracy on PubMedQA, creating a new record. Our case study on text generation further demonstrates the advantage of BioGPT on biomedical literature to generate fluent descriptions for biomedical terms.
35
 
36
 
37
  ## Dataset
38
 
39
  ChatDoctor-200K dataset is collected from this paper https://arxiv.org/pdf/2303.14070.pdf
40
+ The dataset is composed by:
41
+
42
+ - 100k real conversations between patients and doctors from HealthCareMagic.com [HealthCareMagic-100k](https://drive.google.com/file/d/1lyfqIwlLSClhgrCutWuEe_IACNq6XNUt/view?usp=sharing).
43
+
44
+ - 10k real conversations between patients and doctors from icliniq.com [icliniq-10k](https://drive.google.com/file/d/1ZKbqgYqWc7DJHs3N9TQYQVPdDQmZaClA/view?usp=sharing).
45
+
46
+ - 5k generated conversations between patients and physicians from ChatGPT [GenMedGPT-5k](https://drive.google.com/file/d/1nDTKZ3wZbZWTkFMBkxlamrzbNz0frugg/view?usp=sharing) and [disease database](https://github.com/Kent0n-Li/ChatDoctor/blob/main/format_dataset.csv)
47
 
48
 
49
  ## Usage