|
--- |
|
language: |
|
- de |
|
- en |
|
- es |
|
- fr |
|
--- |
|
|
|
# Model Card for `answer-finder-v1-L-multilingual` |
|
|
|
This model is a question answering model developed by Sinequa. It produces two lists of logit scores corresponding to |
|
the start token and end token of an answer. |
|
|
|
Model name: `answer-finder-v1-L-multilingual` |
|
|
|
## Supported Languages |
|
|
|
The model was trained and tested in the following languages: |
|
|
|
- English |
|
- French |
|
- German |
|
- Spanish |
|
|
|
## Scores |
|
|
|
| Metric | Value | |
|
|:--------------------------------------------------------------|-------:| |
|
| F1 Score on SQuAD v2 EN with Hugging Face evaluation pipeline | 75 | |
|
| F1 Score on SQuAD v2 EN with Haystack evaluation pipeline | 75 | |
|
| F1 Score on SQuAD v2 FR with Haystack evaluation pipeline | 73.4 | |
|
| F1 Score on SQuAD v2 DE with Haystack evaluation pipeline | 90.8 | |
|
| F1 Score on SQuAD v2 ES with Haystack evaluation pipeline | 67.1 | |
|
|
|
## Inference Time |
|
|
|
| GPU Info | Batch size 1 | Batch size 32 | |
|
|:--------------------------------------------------------------|---------------:|---------------:| |
|
| NVIDIA A10 | 4 ms | 84 ms | |
|
| NVIDIA T4 | 15 ms | 362 ms | |
|
|
|
**Note that the Answer Finder models are only used at query time.** |
|
|
|
## Requirements |
|
|
|
- Minimal Sinequa version: 11.10.0 |
|
- GPU memory usage: 1060 MiB |
|
|
|
Note that GPU memory usage only includes how much GPU memory the actual model consumes on an NVIDIA T4 GPU with a batch |
|
size of 32. It does not include the fix amount of memory that is consumed by the ONNX Runtime upon initialization which |
|
can be around 0.5 to 1 GiB depending on the used GPU. |
|
|
|
## Model Details |
|
|
|
### Overview |
|
|
|
- Number of parameters: 110 million |
|
- Base language model: [bert-base-multilingual-cased](https://huggingface.co/bert-base-multilingual-cased) |
|
pre-trained by Sinequa in English, French, German and Spanish |
|
- Insensitive to casing and accents |
|
|
|
### Training Data |
|
|
|
- [SQuAD v2](https://rajpurkar.github.io/SQuAD-explorer/) |
|
- [French-SQuAD](https://github.com/Alikabbadj/French-SQuAD) + French translation of SQuAD v2 "impossible" query-passage pairs |
|
- [GermanQuAD](https://www.deepset.ai/germanquad) + German translation of SQuAD v2 "impossible" query-passage pairs |
|
- [SQuAD-es-v2](https://github.com/ccasimiro88/TranslateAlignRetrieve) |
|
|