language: rn
tags:
- translation
license: apache-2.0
run-fra
source group: Rundi
target group: French
OPUS readme: run-fra
model: transformer-align
source language(s): run
target language(s): fra
model: transformer-align
pre-processing: normalization + SentencePiece (spm4k,spm4k)
download original weights: opus-2020-06-16.zip
test set translations: opus-2020-06-16.test.txt
test set scores: opus-2020-06-16.eval.txt
Benchmarks
testset | BLEU | chr-F |
---|---|---|
Tatoeba-test.run.fra | 18.2 | 0.397 |
System Info:
hf_name: run-fra
source_languages: run
target_languages: fra
opus_readme_url: https://github.com/Helsinki-NLP/Tatoeba-Challenge/tree/master/models/run-fra/README.md
original_repo: Tatoeba-Challenge
tags: ['translation']
prepro: normalization + SentencePiece (spm4k,spm4k)
url_model: https://object.pouta.csc.fi/Tatoeba-MT-models/run-fra/opus-2020-06-16.zip
url_test_set: https://object.pouta.csc.fi/Tatoeba-MT-models/run-fra/opus-2020-06-16.test.txt
src_alpha3: run
tgt_alpha3: fra
short_pair: rn-fr
chrF2_score: 0.397
bleu: 18.2
brevity_penalty: 1.0
ref_len: 7496.0
src_name: Rundi
tgt_name: French
train_date: 2020-06-16
src_alpha2: rn
tgt_alpha2: fr
prefer_old: False
long_pair: run-fra
helsinki_git_sha: 480fcbe0ee1bf4774bcbe6226ad9f58e63f6c535
transformers_git_sha: 46e9f53347bbe9e989f0335f98465f30886d8173
port_machine: brutasse
port_time: 2020-08-18-01:48