alquist4121 commited on
Commit
b8ba25f
1 Parent(s): bb568bf

fix explanation

Browse files
Files changed (1) hide show
  1. examples/nllb200_3B_8bit.sh +5 -2
examples/nllb200_3B_8bit.sh CHANGED
@@ -1,5 +1,8 @@
1
- # Run NLLB200-3B model on sample text. We use FP16 precision, which requires a GPU with a lot of VRAM
2
- # For running this model in GPUs with less VRAM, use 8-bit quantization, see examples/nllb200_3B_8bit.sh
 
 
 
3
 
4
  python3 translate.py \
5
  --sentences_path sample_text/en.txt \
 
1
+ # Run NLLB200-3B on sample text. This model requires a GPU with a lot of VRAM, so we use
2
+ # 8-bit quantization to reduce the required VRAM so we can fit in customer grade GPUs. If you have a GPU
3
+ # with a lot of RAM, running the model in FP16 should be faster and produce sighly better results,
4
+ # see examples/nllb200-3B_fp16.sh
5
+
6
 
7
  python3 translate.py \
8
  --sentences_path sample_text/en.txt \