venkatd commited on
Commit
005339e
·
verified ·
1 Parent(s): 24fbde2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +58 -3
README.md CHANGED
@@ -1,3 +1,58 @@
1
- ---
2
- license: unknown
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: unknown
3
+ datasets:
4
+ - ncbi/pubmed
5
+ language:
6
+ - en
7
+ metrics:
8
+ - f1
9
+ base_model:
10
+ - microsoft/deberta-v3-base
11
+ pipeline_tag: token-classification
12
+ tags:
13
+ - NER
14
+ - phenotypes
15
+ - diseases
16
+ - bio
17
+ - classification
18
+ ---
19
+
20
+
21
+ **Model Summary and Training Details**
22
+
23
+ ### Model Architecture
24
+ - **Base Model**: `microsoft/deberta-v3-base`
25
+ - **Task**: Token Classification for Named Entity Recognition (NER) with a focus on disease entities.
26
+ - **Number of Labels**: 3 (O, B-Disease, I-Disease)
27
+
28
+ ### Dataset
29
+ - **Dataset**: NCBI Disease Corpus
30
+ - **Description**: The NCBI Disease corpus is a specialized medical dataset that includes 793 PubMed abstracts. It is structured to help in identifying disease mentions within scientific literature, and each mention is annotated with disease concepts from the MeSH (Medical Subject Headings) or OMIM (Online Mendelian Inheritance in Man) databases.
31
+ - **Split**:
32
+ - Training Set: 593 abstracts
33
+ - Development (Validation) Set: 100 abstracts
34
+ - Test Set: 100 abstracts
35
+
36
+ ### Training Details
37
+ - **Training Steps**: The model was trained using a cross-entropy loss function for token classification tasks. To optimize performance, we used gradient accumulation to achieve a stable loss and improve resource efficiency.
38
+ - **Gradient Accumulation**: 2 steps
39
+ - **Batch Size**: 8
40
+ - **Device**: Trained on a GPU if available, using mixed-precision training for better performance.
41
+
42
+ ### Optimizer and Learning Rate Scheduler
43
+ - **Optimizer**: AdamW
44
+ - **Learning Rate**: 1e-5
45
+ - **Betas**: (0.9, 0.999)
46
+ - **Epsilon**: 1e-8
47
+
48
+ - **Learning Rate Scheduler**: Cosine Scheduler with Warmup
49
+ - **Warmup Steps**: 10% of total training steps
50
+ - **Total Training Steps**: Calculated as `len(train_loader) * num_epochs`
51
+
52
+ ### Epochs and Validation
53
+ - **Epochs**: 5
54
+ - **Training and Validation Loss**: The model achieved a stable loss over 5 epochs, with the best validation loss recorded. The best model based on validation loss was saved for evaluation.
55
+
56
+ ### Evaluation and Performance
57
+ - **Test Dataset F1 Score**: 0.9772
58
+ - **Evaluation Metric**: F1 score, which indicates the balance between precision and recall, was used as the primary metric to assess the model’s performance.