naimul011 commited on
Commit
f9f7b52
1 Parent(s): 12664ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -1
README.md CHANGED
@@ -3,11 +3,40 @@ library_name: transformers
3
  tags:
4
  - trl
5
  - sft
 
 
6
  ---
7
 
8
  # Model Card for Model ID
9
 
10
- <!-- Provide a quick summary of what the model is/does. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
 
13
 
 
3
  tags:
4
  - trl
5
  - sft
6
+ base_model:
7
+ - microsoft/Orca-2-7b
8
  ---
9
 
10
  # Model Card for Model ID
11
 
12
+ ### Description of GlueOrca Model
13
+
14
+ The GlueOrca model is fine-tuned on four distinct tasks from the GLUE benchmark: SST-2 (Sentiment Analysis), MRPC (Paraphrase Detection), CoLA (Linguistic Acceptability), and MNLI (Natural Language Inference). The base model used is "naimul011/GlueOrca," which is designed for various natural language understanding tasks, leveraging a strong pre-training phase followed by fine-tuning to improve task-specific performance.
15
+
16
+ The fine-tuning of GlueOrca focuses on enhancing the model's capabilities in retaining previously learned knowledge while adapting to new tasks, measuring aspects like catastrophic forgetting, learning capability, and training performance across different language tasks.
17
+
18
+ ### Benchmark Table for GlueOrca Fine-Tuning Performance
19
+
20
+ | **Model** | **Parameter Size (B)** | **Pretrained Performance** | **Forgetting** | **Learning** | **Training Performance** |
21
+ |-----------------|------------------------|----------------------------|----------------|--------------|--------------------------|
22
+ | Llama-3.2-1B | 1 | 0.50 | 0.24 | 0.33 | 0.54 |
23
+ | Llama-3.2-3B | 3 | 0.56 | 0.225 | 0.36 | 0.61 |
24
+ | Llama-3.1-8B | 8 | 0.56 | 0.59 | 0.84 | 0.67 |
25
+ | Llama-3-8B | 8 | 0.53 | 0.39 | 0.98 | 0.70 |
26
+ | Llama-2-7B | 7 | 0.67 | 0.23 | 0.12 | 0.63 |
27
+ | GPT-J-6B | 6 | 0.50 | 0.39 | 0.45 | 0.54 |
28
+ | Phi-2 | 2.7 | 0.59 | 0.10 | 0.15 | 0.61 |
29
+ | Phi-3.5-mini | 3.82 | 0.69 | **0.02** | 0.30 | 0.76 |
30
+ | Orca-2-7b | 7 | **0.76** | 0.185 | 0.33 | **0.81** |
31
+ | Qwen2.5-0.5B | 0.5 | 0.52 | 0.23 | 0.56 | 0.61 |
32
+ | Qwen2.5-7B | 7 | 0.56 | 0.51 | **1.12** | 0.77 |
33
+ | Qwen2.5-14B | 14 | 0.71 | 0.935 | 0.66 | 0.80 |
34
+ | **GlueOrca** | **7** | **0.61** | **0.35** | **1.05** | **0.75** |
35
+
36
+ ### Analysis
37
+
38
+ GlueOrca, fine-tuned on multiple tasks from the GLUE dataset, shows a pre-trained performance of 0.61. The model demonstrates moderate forgetting at 0.35, indicating some loss of prior knowledge. However, it excels in learning capability with a score of 1.05, reflecting its strong ability to adapt to new tasks. The overall training performance is 0.75, making GlueOrca a well-rounded model with balanced retention and learning improvements across various tasks.
39
+
40
 
41
 
42