gz987 commited on
Commit
6b1769b
·
verified ·
1 Parent(s): 5ef7ddd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -6
README.md CHANGED
@@ -5,18 +5,23 @@ base_model:
5
  tags:
6
  - merge
7
  ---
8
- This model is a merged model based on [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) using a novel model merging technique.
9
 
10
- ## **Performance (Self-Tested on A100)**
11
 
12
- The following results are obtained using **batch_size=6** on an A100 GPU. Official results are pending submission to `open_llm_leaderboard`.
 
 
 
 
13
 
14
  | IFEVAL | BBH | MATH | GPQA | MUSR | MMLU-PRO | AVG |
15
  |--------|------|------|------|------|----------|-------|
16
- | 75.06 | 35.87 | 49.62 | 7.27 | 14.24 | 37.73 | 36.63 |
 
 
17
 
18
- **Note**: These results will be updated once officially verified.
19
 
20
  ## **Recipe Coming Soon**
21
 
22
- We will release details on the merging technique and methodology soon. Stay tuned! 🚀
 
5
  tags:
6
  - merge
7
  ---
 
8
 
9
+ # **Qwen2.5-7B-CABS-v1: A Merged Model Using a Novel Technique**
10
 
11
+ This model is a merged version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) utilizing a novel model merging technique to optimize performance while maintaining robustness across tasks.
12
+
13
+ ## **Official Performance**
14
+
15
+ This model has been officially evaluated on `open_llm_leaderboard`, achieving the following results:
16
 
17
  | IFEVAL | BBH | MATH | GPQA | MUSR | MMLU-PRO | AVG |
18
  |--------|------|------|------|------|----------|-------|
19
+ | 75.06 | 35.84 | 47.96 | 8.50 | 14.17 | 37.84 | 36.56 |
20
+
21
+ As of **February 19, 2025**, this model ranks **4th among all 7B and smaller models** on `open_llm_leaderboard`.
22
 
23
+ For the latest and complete leaderboard rankings, please refer to the [open_llm_leaderboard](https://huggingface.co/spaces/open-llm-leaderboard).
24
 
25
  ## **Recipe Coming Soon**
26
 
27
+ We will release details on the merging technique and methodology soon, Stay tuned for updates! 🚀