alirezamsh commited on
Commit
70a4f18
1 Parent(s): 3e1147d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -119,11 +119,15 @@ SMaLL-100 is a compact and fast massively multilingual machine translation model
119
 
120
  The model architecture and config are the same as [M2M-100](https://huggingface.co/facebook/m2m100_418M/tree/main) implementation, but the tokenizer is modified to adjust language codes. So, you should load the tokenizer locally from [tokenization_small100.py](https://huggingface.co/alirezamsh/small100/blob/main/tokenization_small100.py) file for the moment.
121
 
122
- **Note**: SMALL100Tokenizer requires sentencepiece, so make sure to install it by ```pip install sentencepiece```
 
 
123
 
124
  - **Supervised Training**
125
 
126
- SMaLL-100 is a seq-to-seq model for the translation task. The input to the model is ```source:[tgt_lang_code] + src_tokens + [EOS]``` and ```target: tgt_tokens + [EOS]```. An example of supervised training is shown below:
 
 
127
 
128
  ```
129
  from transformers import M2M100ForConditionalGeneration
 
119
 
120
  The model architecture and config are the same as [M2M-100](https://huggingface.co/facebook/m2m100_418M/tree/main) implementation, but the tokenizer is modified to adjust language codes. So, you should load the tokenizer locally from [tokenization_small100.py](https://huggingface.co/alirezamsh/small100/blob/main/tokenization_small100.py) file for the moment.
121
 
122
+ **Note**: SMALL100Tokenizer requires sentencepiece, so make sure to install it by:
123
+
124
+ ```pip install sentencepiece```
125
 
126
  - **Supervised Training**
127
 
128
+ SMaLL-100 is a seq-to-seq model for the translation task. The input to the model is ```source:[tgt_lang_code] + src_tokens + [EOS]``` and ```target: tgt_tokens + [EOS]```.
129
+
130
+ An example of supervised training is shown below:
131
 
132
  ```
133
  from transformers import M2M100ForConditionalGeneration