|
--- |
|
language: de |
|
license: mit |
|
tags: |
|
- german |
|
|
|
--- |
|
|
|
We released the German Question Answering model fine-tuned with our own German Question Answering dataset (**deQuAD**) containing **130k** training and **11k** test QA pairs. |
|
|
|
## Overview |
|
- **Language model:** [electra-base-german-uncased](https://huggingface.co/german-nlp-group/electra-base-german-uncased) |
|
- **Language:** German |
|
- **Training data:** deQuAD2.0 training set (~42MB) |
|
- **Evaluation data:** deQuAD2.0 test set (~4MB) |
|
- **Infrastructure:** 8xV100 GPU |
|
|
|
## Evaluation |
|
We benchmarked the question answering performance on our deQuAD test data with some German language models. The fine-tuned electra-base-german-uncased model gives the best performance (Exact Match/F1). |
|
|
|
|
|
| Model | All | HasAns | NoAns | |
|
|-------|--------|--------|--------| |
|
| electra-base-german-uncased | 70.97/76.18 | 67.73/78.02 | 74.29/74.29 | |
|
| bert-base-german-cased |58.98/64.77| 49.19/60.63| 69.03/69.03| |
|
|bert-base-german-dbmdz-uncased|63.70/68.00| 57.03/65.52| 70.51/70.51 | |
|
|dbmdz/bert-base-german-europeana-uncased| 58.79/63.38| 52.14/61.22| 65.59/65.59| |
|
|