File size: 12,727 Bytes
f625280
8857129
cfb8432
 
 
ded8724
bf541c0
4780f32
a0d791a
 
 
 
cfb8432
fa23fe6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
---
datasets:
- Locutusque/TM-DATA-V2
- LLM360/TxT360
- mlfoundations/dclm-baseline-1.0
- Skylion007/openwebtext
- JeanKaddour/minipile
- eminorhan/gutenberg_en
language:
- en
license: apache-2.0
---

still in training. Trained on about ~21 billion tokens so far.

|                 Tasks                  |Version|     Filter     |n-shot|  Metric   |   | Value |   |Stderr|
|----------------------------------------|-------|----------------|-----:|-----------|---|------:|---|-----:|
|Open LLM Leaderboard                    |    N/A|                |      |           |   |       |   |      |
| - arc_challenge                        |      1|none            |    25|acc        |↑  | 0.2005|±  |0.0117|
|                                        |       |none            |    25|acc_norm   |↑  | 0.2406|±  |0.0125|
| - gsm8k                                |      3|flexible-extract|     5|exact_match|↑  | 0.0083|±  |0.0025|
|                                        |       |strict-match    |     5|exact_match|↑  | 0.0000|±  |0.0000|
| - hellaswag                            |      1|none            |    10|acc        |↑  | 0.2724|±  |0.0044|
|                                        |       |none            |    10|acc_norm   |↑  | 0.2838|±  |0.0045|
| - mmlu                                 |      2|none            |      |acc        |↑  | 0.2290|±  |0.0035|
|  - humanities                          |      2|none            |      |acc        |↑  | 0.2380|±  |0.0062|
|   - formal_logic                       |      1|none            |     5|acc        |↑  | 0.2460|±  |0.0385|
|   - high_school_european_history       |      1|none            |     5|acc        |↑  | 0.1818|±  |0.0301|
|   - high_school_us_history             |      1|none            |     5|acc        |↑  | 0.2647|±  |0.0310|
|   - high_school_world_history          |      1|none            |     5|acc        |↑  | 0.2911|±  |0.0296|
|   - international_law                  |      1|none            |     5|acc        |↑  | 0.2149|±  |0.0375|
|   - jurisprudence                      |      1|none            |     5|acc        |↑  | 0.2685|±  |0.0428|
|   - logical_fallacies                  |      1|none            |     5|acc        |↑  | 0.2209|±  |0.0326|
|   - moral_disputes                     |      1|none            |     5|acc        |↑  | 0.2457|±  |0.0232|
|   - moral_scenarios                    |      1|none            |     5|acc        |↑  | 0.2369|±  |0.0142|
|   - philosophy                         |      1|none            |     5|acc        |↑  | 0.1865|±  |0.0221|
|   - prehistory                         |      1|none            |     5|acc        |↑  | 0.1975|±  |0.0222|
|   - professional_law                   |      1|none            |     5|acc        |↑  | 0.2432|±  |0.0110|
|   - world_religions                    |      1|none            |     5|acc        |↑  | 0.3099|±  |0.0355|
|  - other                               |      2|none            |      |acc        |↑  | 0.2375|±  |0.0076|
|   - business_ethics                    |      1|none            |     5|acc        |↑  | 0.3200|±  |0.0469|
|   - clinical_knowledge                 |      1|none            |     5|acc        |↑  | 0.2226|±  |0.0256|
|   - college_medicine                   |      1|none            |     5|acc        |↑  | 0.1965|±  |0.0303|
|   - global_facts                       |      1|none            |     5|acc        |↑  | 0.1800|±  |0.0386|
|   - human_aging                        |      1|none            |     5|acc        |↑  | 0.3004|±  |0.0308|
|   - management                         |      1|none            |     5|acc        |↑  | 0.1942|±  |0.0392|
|   - marketing                          |      1|none            |     5|acc        |↑  | 0.2735|±  |0.0292|
|   - medical_genetics                   |      1|none            |     5|acc        |↑  | 0.3000|±  |0.0461|
|   - miscellaneous                      |      1|none            |     5|acc        |↑  | 0.2478|±  |0.0154|
|   - nutrition                          |      1|none            |     5|acc        |↑  | 0.2222|±  |0.0238|
|   - professional_accounting            |      1|none            |     5|acc        |↑  | 0.2021|±  |0.0240|
|   - professional_medicine              |      1|none            |     5|acc        |↑  | 0.1912|±  |0.0239|
|   - virology                           |      1|none            |     5|acc        |↑  | 0.2590|±  |0.0341|
|  - social sciences                     |      2|none            |      |acc        |↑  | 0.2203|±  |0.0075|
|   - econometrics                       |      1|none            |     5|acc        |↑  | 0.2368|±  |0.0400|
|   - high_school_geography              |      1|none            |     5|acc        |↑  | 0.2020|±  |0.0286|
|   - high_school_government_and_politics|      1|none            |     5|acc        |↑  | 0.1865|±  |0.0281|
|   - high_school_macroeconomics         |      1|none            |     5|acc        |↑  | 0.2205|±  |0.0210|
|   - high_school_microeconomics         |      1|none            |     5|acc        |↑  | 0.2143|±  |0.0267|
|   - high_school_psychology             |      1|none            |     5|acc        |↑  | 0.1908|±  |0.0168|
|   - human_sexuality                    |      1|none            |     5|acc        |↑  | 0.2672|±  |0.0388|
|   - professional_psychology            |      1|none            |     5|acc        |↑  | 0.2386|±  |0.0172|
|   - public_relations                   |      1|none            |     5|acc        |↑  | 0.1727|±  |0.0362|
|   - security_studies                   |      1|none            |     5|acc        |↑  | 0.2367|±  |0.0272|
|   - sociology                          |      1|none            |     5|acc        |↑  | 0.2488|±  |0.0306|
|   - us_foreign_policy                  |      1|none            |     5|acc        |↑  | 0.2600|±  |0.0441|
|  - stem                                |      2|none            |      |acc        |↑  | 0.2157|±  |0.0073|
|   - abstract_algebra                   |      1|none            |     5|acc        |↑  | 0.2200|±  |0.0416|
|   - anatomy                            |      1|none            |     5|acc        |↑  | 0.1778|±  |0.0330|
|   - astronomy                          |      1|none            |     5|acc        |↑  | 0.1908|±  |0.0320|
|   - college_biology                    |      1|none            |     5|acc        |↑  | 0.2778|±  |0.0375|
|   - college_chemistry                  |      1|none            |     5|acc        |↑  | 0.2200|±  |0.0416|
|   - college_computer_science           |      1|none            |     5|acc        |↑  | 0.2100|±  |0.0409|
|   - college_mathematics                |      1|none            |     5|acc        |↑  | 0.2100|±  |0.0409|
|   - college_physics                    |      1|none            |     5|acc        |↑  | 0.2157|±  |0.0409|
|   - computer_security                  |      1|none            |     5|acc        |↑  | 0.2700|±  |0.0446|
|   - conceptual_physics                 |      1|none            |     5|acc        |↑  | 0.2638|±  |0.0288|
|   - electrical_engineering             |      1|none            |     5|acc        |↑  | 0.2483|±  |0.0360|
|   - elementary_mathematics             |      1|none            |     5|acc        |↑  | 0.2037|±  |0.0207|
|   - high_school_biology                |      1|none            |     5|acc        |↑  | 0.1774|±  |0.0217|
|   - high_school_chemistry              |      1|none            |     5|acc        |↑  | 0.2020|±  |0.0282|
|   - high_school_computer_science       |      1|none            |     5|acc        |↑  | 0.2500|±  |0.0435|
|   - high_school_mathematics            |      1|none            |     5|acc        |↑  | 0.2148|±  |0.0250|
|   - high_school_physics                |      1|none            |     5|acc        |↑  | 0.2053|±  |0.0330|
|   - high_school_statistics             |      1|none            |     5|acc        |↑  | 0.1481|±  |0.0242|
|   - machine_learning                   |      1|none            |     5|acc        |↑  | 0.3125|±  |0.0440|
| - truthfulqa_gen                       |      3|none            |     0|bleu_acc   |↑  | 0.2362|±  |0.0149|
|                                        |       |none            |     0|bleu_diff  |↑  |-1.0138|±  |0.2569|
|                                        |       |none            |     0|bleu_max   |↑  | 7.9522|±  |0.4088|
|                                        |       |none            |     0|rouge1_acc |↑  | 0.2595|±  |0.0153|
|                                        |       |none            |     0|rouge1_diff|↑  |-1.9129|±  |0.4349|
|                                        |       |none            |     0|rouge1_max |↑  |21.7885|±  |0.7307|
|                                        |       |none            |     0|rouge2_acc |↑  | 0.1200|±  |0.0114|
|                                        |       |none            |     0|rouge2_diff|↑  |-1.9771|±  |0.3475|
|                                        |       |none            |     0|rouge2_max |↑  | 9.0199|±  |0.5842|
|                                        |       |none            |     0|rougeL_acc |↑  | 0.2570|±  |0.0153|
|                                        |       |none            |     0|rougeL_diff|↑  |-1.8812|±  |0.4185|
|                                        |       |none            |     0|rougeL_max |↑  |19.6284|±  |0.6850|
| - truthfulqa_mc1                       |      2|none            |     0|acc        |↑  | 0.1983|±  |0.0140|
| - truthfulqa_mc2                       |      2|none            |     0|acc        |↑  | 0.3861|±  |0.0147|
| - winogrande                           |      1|none            |     5|acc        |↑  | 0.4972|±  |0.0141|

|      Groups       |Version|Filter|n-shot|Metric|   |Value |   |Stderr|
|-------------------|------:|------|------|------|---|-----:|---|-----:|
| - mmlu            |      2|none  |      |acc   |↑  |0.2290|±  |0.0035|
|  - humanities     |      2|none  |      |acc   |↑  |0.2380|±  |0.0062|
|  - other          |      2|none  |      |acc   |↑  |0.2375|±  |0.0076|
|  - social sciences|      2|none  |      |acc   |↑  |0.2203|±  |0.0075|
|  - stem           |      2|none  |      |acc   |↑  |0.2157|±  |0.0073|

|              Tasks              |Version|Filter|n-shot| Metric |   |Value |   |Stderr|
|---------------------------------|------:|------|-----:|--------|---|-----:|---|-----:|
|agieval_nous                     |      0|none  |      |acc_norm|↑  |0.2133|±  |0.0081|
| - agieval_aqua_rat              |      1|none  |     0|acc     |↑  |0.2047|±  |0.0254|
|                                 |       |none  |     0|acc_norm|↑  |0.1969|±  |0.0250|
| - agieval_logiqa_en             |      1|none  |     0|acc     |↑  |0.2043|±  |0.0158|
|                                 |       |none  |     0|acc_norm|↑  |0.2304|±  |0.0165|
| - agieval_lsat_ar               |      1|none  |     0|acc     |↑  |0.1739|±  |0.0250|
|                                 |       |none  |     0|acc_norm|↑  |0.1957|±  |0.0262|
| - agieval_lsat_lr               |      1|none  |     0|acc     |↑  |0.1549|±  |0.0160|
|                                 |       |none  |     0|acc_norm|↑  |0.1608|±  |0.0163|
| - agieval_lsat_rc               |      1|none  |     0|acc     |↑  |0.1636|±  |0.0226|
|                                 |       |none  |     0|acc_norm|↑  |0.2119|±  |0.0250|
| - agieval_sat_en                |      1|none  |     0|acc     |↑  |0.2670|±  |0.0309|
|                                 |       |none  |     0|acc_norm|↑  |0.2621|±  |0.0307|
| - agieval_sat_en_without_passage|      1|none  |     0|acc     |↑  |0.2670|±  |0.0309|
|                                 |       |none  |     0|acc_norm|↑  |0.2621|±  |0.0307|
| - agieval_sat_math              |      1|none  |     0|acc     |↑  |0.2182|±  |0.0279|
|                                 |       |none  |     0|acc_norm|↑  |0.2318|±  |0.0285|
|arc_challenge                    |      1|none  |     0|acc     |↑  |0.1945|±  |0.0116|
|                                 |       |none  |     0|acc_norm|↑  |0.2372|±  |0.0124|
|truthfulqa_mc2                   |      2|none  |     0|acc     |↑  |0.3861|±  |0.0147|

|   Groups   |Version|Filter|n-shot| Metric |   |Value |   |Stderr|
|------------|------:|------|------|--------|---|-----:|---|-----:|
|agieval_nous|      0|none  |      |acc_norm|↑  |0.2133|±  |0.0081|