hung200504 commited on
Commit
4fed219
1 Parent(s): aab2dff

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -27
README.md CHANGED
@@ -4,45 +4,52 @@ base_model: microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
- - squad
8
  model-index:
9
- - name: bert-squadv2
10
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
- should probably proofread and complete it, then remove this comment. -->
15
 
16
- # bert-squadv2
17
 
18
- This model is a fine-tuned version of [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext) on the squad dataset.
19
- It achieves the following results on the evaluation set:
20
- - Loss: 1.1930
21
 
22
- ## Model description
23
 
24
- More information needed
25
 
26
- ## Intended uses & limitations
27
 
28
- More information needed
 
29
 
30
- ## Training and evaluation data
31
 
32
- More information needed
33
-
34
- ## Training procedure
35
-
36
- ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
- - learning_rate: 3e-05
40
- - train_batch_size: 16
41
- - eval_batch_size: 16
42
- - seed: 42
43
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
- - lr_scheduler_type: linear
45
- - num_epochs: 3
46
 
47
  ### Training results
48
 
@@ -205,4 +212,4 @@ The following hyperparameters were used during training:
205
  - Transformers 4.34.1
206
  - Pytorch 2.1.0+cu118
207
  - Datasets 2.14.5
208
- - Tokenizers 0.14.1
 
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
+ - squad_v2
8
  model-index:
9
+ - name: bert-squadv2-biomed
10
+ results:
11
+ - task:
12
+ type: question-answering
13
+ dataset:
14
+ type: reading-comprehension
15
+ name: SQuADv2
16
+ metrics:
17
+ - name: accuracy
18
+ type: accuracy
19
+ value: 0.77
20
+ verified: false
21
+ language:
22
+ - en
23
+ pipeline_tag: question-answering
24
  ---
25
 
26
+ # bert-squadv2-biomed
 
27
 
28
+ This model is a fine-tuned version of [microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext](https://huggingface.co/microsoft/BiomedNLP-PubMedBERT-base-uncased-abstract-fulltext) on the SQuADv2 dataset. It has been fine-tuned for question-answering tasks specifically related to biomedical texts, leveraging the SQuAD v2 dataset to enhance its ability to manage both answerable and unanswerable questions.
29
 
30
+ ## Model Description
 
 
31
 
32
+ The base model, **PubMedBERT**, was originally pre-trained on biomedical abstracts and full-text articles from PubMed. This fine-tuned version adapts PubMedBERT for biomedical question-answering by training it with **SQuADv2**, a dataset that includes over 100,000 questions with answerable and unanswerable queries.
33
 
34
+ - **Use Cases**: This model is particularly useful in applications where quick and accurate question-answering from biomedical literature is needed. It is designed to provide answers to specific questions, as well as to detect when no relevant answer exists.
35
 
36
+ ## Training and Evaluation Data
37
 
38
+ - **Dataset**: The model was fine-tuned on the **SQuADv2** dataset, which consists of reading comprehension tasks where some questions have no answer in the provided context.
39
+ - **Training Environment**: The model was trained in a Colab environment. A link to the training notebook can be found here: [Training Notebook](https://colab.research.google.com/drive/11je7-YnFQ-oISxC_7KS4QTfs3fgWOseU?usp=sharing).
40
 
41
+ ## Training Procedure
42
 
43
+ ### Hyperparameters
 
 
 
 
44
 
45
  The following hyperparameters were used during training:
46
+ - `learning_rate`: 3e-05
47
+ - `train_batch_size`: 16
48
+ - `eval_batch_size`: 16
49
+ - `seed`: 42
50
+ - `optimizer`: Adam (betas=(0.9, 0.999), epsilon=1e-08)
51
+ - `lr_scheduler_type`: linear
52
+ - `num_epochs`: 3
53
 
54
  ### Training results
55
 
 
212
  - Transformers 4.34.1
213
  - Pytorch 2.1.0+cu118
214
  - Datasets 2.14.5
215
+ - Tokenizers 0.14.1