tiedeman commited on
Commit
b3513c3
1 Parent(s): 6794f13

added a note about performance issues for low-resource languages

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -16675,7 +16675,7 @@ model-index:
16675
 
16676
  ## Model Details
16677
 
16678
- Neural machine translation model for translating from Multiple languages (mul) to unknown (deu+eng+fra+por+spa).
16679
 
16680
  This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
16681
  **Model Description:**
@@ -16708,6 +16708,8 @@ This model can be used for translation and text-to-text generation.
16708
 
16709
  Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
16710
 
 
 
16711
  ## How to Get Started With the Model
16712
 
16713
  A short example code:
 
16675
 
16676
  ## Model Details
16677
 
16678
+ Neural machine translation model for translating from Multiple languages (mul) to unknown (deu+eng+fra+por+spa). Note that many of the listed languages will not be well supported by the model as the training data is very limited for the majority of the languages. Translation performance varies a lot and for a large number of language pairs it will not work at all.
16679
 
16680
  This model is part of the [OPUS-MT project](https://github.com/Helsinki-NLP/Opus-MT), an effort to make neural machine translation models widely available and accessible for many languages in the world. All models are originally trained using the amazing framework of [Marian NMT](https://marian-nmt.github.io/), an efficient NMT implementation written in pure C++. The models have been converted to pyTorch using the transformers library by huggingface. Training data is taken from [OPUS](https://opus.nlpl.eu/) and training pipelines use the procedures of [OPUS-MT-train](https://github.com/Helsinki-NLP/Opus-MT-train).
16681
  **Model Description:**
 
16708
 
16709
  Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
16710
 
16711
+ Also note that many of the listed languages will not be well supported by the model as the training data is very limited for the majority of the languages. Translation performance varies a lot and for a large number of language pairs it will not work at all.
16712
+
16713
  ## How to Get Started With the Model
16714
 
16715
  A short example code: