imdbo commited on
Commit
8d0da75
verified
1 Parent(s): e8a8b61

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -3
README.md CHANGED
@@ -1,3 +1,33 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ ---
4
+ ---
5
+ license: mit
6
+ ---
7
+ **Model Description**
8
+
9
+ Model created with OpenNMT-py 3.2 for the Spanish-Aragonese pair using a transformer architecture. The model was converted to the ctranslate2 format.
10
+ This model was trained for the paper Training and fine-tuning NMT models for low-resource languages using Apertium-based synthetic corpora
11
+
12
+ **How to Translate with this Model**
13
+
14
+ + Install [Python 3.9](https://www.python.org/downloads/release/python-390/)
15
+ + Install [ctranslate 3.2](https://github.com/OpenNMT/CTranslate2)
16
+ + Translate an input_text using the NOS-MT-es-ast model with the following command:
17
+ ```bash
18
+ perl tokenizer.perl < input.txt > input.tok
19
+ ```
20
+ ```bash
21
+ subword_nmt.apply_bpe -c ./bpe/es.bpe < input.tok > input.bpe
22
+ ```
23
+ ```bash
24
+ python3 translate.py ./ct2-ast input.bpe > output.txt
25
+ ```
26
+ ```bash
27
+ sed -i 's/@@ //g' output.txt
28
+ ```
29
+
30
+ ## Citation
31
+
32
+ If you use this model in your research, please cite the following paper:
33
+ Sant, A., Bardanca Outeiri帽o, D., Pichel Campos, J. R., De Luca Fornaciari, F., Escolano, C., Garc铆a Gilabert, J., Gamallo Otero, P., Mash, A., Liao, X., & Melero, M. (2023). Training and fine-tuning NMT models for low-resource languages using Apertium-based synthetic corpora. arXiv.