KlaudiaTH
commited on
Commit
β’
aa451eb
1
Parent(s):
362f22e
Fixed README
Browse files
README.md
CHANGED
@@ -1,13 +1,3 @@
|
|
1 |
-
# New data model
|
2 |
-
|
3 |
-
The new model is constructed by taking individual json files in data/new_eval, combining them together into
|
4 |
-
a simple format, and from the combined df, we create individual files for each models.
|
5 |
-
|
6 |
-
For the new eval runs which has to be appended, we first analyze the model associated with the json file
|
7 |
-
produced from eval harness, select the corresponding model file to append, find the unique rows (unique configuration
|
8 |
-
of model name, language, task group and few shot) in the json file, append if unique rows are not 0.
|
9 |
-
|
10 |
-
|
11 |
---
|
12 |
title: Leaderboard
|
13 |
emoji: π
|
@@ -20,30 +10,8 @@ pinned: false
|
|
20 |
license: unknown
|
21 |
---
|
22 |
|
23 |
-
# Introduction
|
24 |
|
25 |
This is the OpenGPT-X mutlilingual leaderboard source code repository.
|
26 |
The leaderboard aims to provied an overview of LLM performance over various languages.
|
27 |
The basic task set consists of MMLU, ARC, HellaSwag, GSM8k, TruthfulQA and belebele.
|
28 |
To make the results comparable to the Open LLM leaderboard (https://huggingface.co/open-llm-leaderboard) we selected the former five tasks based on our internal machine translations of the English base tasks, in addition to the high-quality multilingual benchmark belebele by Meta.
|
29 |
-
|
30 |
-
# Usage
|
31 |
-
|
32 |
-
The actually hosted leaderboard can be found under https://huggingface.co/spaces/openGPT-X/leaderboard.
|
33 |
-
In order to extend its functionality please create a PR.
|
34 |
-
|
35 |
-
# Adding new tasks
|
36 |
-
|
37 |
-
In order to add new evaluation tasks proceed as follows:
|
38 |
-
|
39 |
-
1. Add task information to `TASK_INFO` in `src/data.py`. It should be a dict mapping the task display name to the metric to be shown, as well as a dict containing mappings from two-letter language codes to the corresponding lm-eval-harness task selection string. See existing task information for reference.
|
40 |
-
2. Add evaluation results as detailed below.
|
41 |
-
|
42 |
-
# Adding new models
|
43 |
-
|
44 |
-
It is possible to change the display name of a particular model.
|
45 |
-
Simply add an entry to `_MODEL_NAMES` in `src/data.py`.
|
46 |
-
|
47 |
-
# Adding evaluation results
|
48 |
-
|
49 |
-
Copy the `.json`-output generated by the lm-eval-harness into `data`.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
title: Leaderboard
|
3 |
emoji: π
|
|
|
10 |
license: unknown
|
11 |
---
|
12 |
|
|
|
13 |
|
14 |
This is the OpenGPT-X mutlilingual leaderboard source code repository.
|
15 |
The leaderboard aims to provied an overview of LLM performance over various languages.
|
16 |
The basic task set consists of MMLU, ARC, HellaSwag, GSM8k, TruthfulQA and belebele.
|
17 |
To make the results comparable to the Open LLM leaderboard (https://huggingface.co/open-llm-leaderboard) we selected the former five tasks based on our internal machine translations of the English base tasks, in addition to the high-quality multilingual benchmark belebele by Meta.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|