germeo-7b-laser / README.md
aari1995's picture
Update README.md
2385765 verified
|
raw
history blame
3.6 kB
---
license: apache-2.0
datasets:
- LeoLM/OpenSchnabeltier
language:
- de
pipeline_tag: conversational
tags:
- hermeo
- laser
---
(Evaluation WIP)
## Hermes + Leo + German Laser = Germeo
Germeo-7B-AWQ
A German-English understanding, but German-only speaking model merged from Hermeo-7B.
### Model details
**Merged from**: leo-mistral-hessianai-7b-chat and DPOpenHermes-7B-v2
**Model type**: Causal decoder-only transformer language model
**Languages**: German replies with English Understanding Capabilities
**Laser-Data**: LeoLM/OpenSchnabeltier
This is an early experiment on laser and its influence on language understanding. It generally improves the language understanding capabilities.
The hypothesis is that it degrades the probability of English replies and increasing those of German replies. The models internal German capabilities are boosted.
Will keep you updated..
### German benchmarks
| **German tasks:** | **MMLU-DE** | **Hellaswag-DE** | **ARC-DE** |**Average** |
|-------------------------------|-------------|---------------|--------------|--------------|
| **Models / Few-shots:** | _(5 shots)_ | _(10 shots)_ | _(24 shots)_ | |
| _7B parameters_ | | | | |
| llama-2-7b | 0.400 | 0.513 | 0.381 | 0.431 |
| leo-hessianai-7b | 0.400 | 0.609 | 0.429 | 0.479 |
| bloom-6b4-clp-german | 0.274 | 0.550 | 0.351 | 0.392 |
| mistral-7b | **0.524** | 0.588 | 0.473 | 0.528 |
| leo-mistral-hessianai-7b | 0.481 | 0.663 | 0.485 | 0.543 |
| leo-mistral-hessianai-7b-chat | 0.458 | 0.617 | 0.465 | 0.513 |
| DPOpenHermes-7B-v2 | 0.517 | 0.603 | 0.515 | 0.545 |
| hermeo-7b | 0.511 | **0.668** | **0.528** | **0.569** |
| **germeo-7b-laser (this model)**| ? | ? | ? | ? |
| _13B parameters_ | | | | |
| llama-2-13b | 0.469 | 0.581 | 0.468 | 0.506 |
| leo-hessianai-13b | **0.486** | **0.658** | **0.509** | **0.551** |
| _70B parameters_ | | | | |
| llama-2-70b | 0.597 | 0.674 | 0.561 | 0.611 |
| leo-hessianai-70b | **0.653** | **0.721** | **0.600** | **0.658** |
Even though the model does not generate English text without being explicitly asked, performance on English Benchmarks is still up:
### English benchmarks
| **English tasks:** | **MMLU** | **Hellaswag** | **ARC** | **Average** |
|------------------------------------|-------------|---------------|--------------|-------------|
| **Models / Few-shots:** | _(5 shots)_ | _(10 shots)_ | _(24 shots)_ | |
| llama-2-7b | 0.466 | 0.786 | 0.530 | 0.594 |
| leolm-hessianai-7b | 0.423 | 0.759 | 0.522 | 0.568 |
| bloom-6b4-clp-german | 0.264 | 0.525 | 0.328 | 0.372 |
| mistral-7b | **0.635** | **0.832** | 0.607 | **0.691** |
| leolm-mistral-hessianai-7b | 0.550 | 0.777 | 0.518 | 0.615 |
| hermeo-7b | 0.601 | 0.821 | **0.620** | 0.681 |
| germeo-7b-laser (this model) | 0.601 | 0.828 | 0.608 | 0.679 |