File size: 3,473 Bytes
3d5dc8c 129a8fc 3d5dc8c f4b6a4b 078e6cf 3d5dc8c ef22a52 3d5dc8c ef22a52 9ab6d30 078e6cf 3d5dc8c bb52312 3d5dc8c 3a44704 3d5dc8c 078e6cf 3d5dc8c 078e6cf 3d5dc8c 078e6cf ef22a52 078e6cf 9ab6d30 078e6cf ef22a52 078e6cf 31ce9f1 078e6cf 3d5dc8c 9ab6d30 129a8fc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 |
---
inference: false
language:
- ja
- en
license: llama2
tags:
- translation
---
# New Translation model released.
[C3TR-Adapter](https://huggingface.co/webbigdata/C3TR-Adapter) is the QLoRA adapter for google/gemma-7b.
Despite the 4-bit quantization, the memory GPU requirement has increased to 8.1 GB.
However, it is possible to run it with the free version of Colab and the performance is much improved!
# webbigdata/ALMA-7B-Ja-V2-GPTQ-Ja-En
ALMA-7B-Ja-V2は日本語から英語、英語から日本語への機械翻訳を行うモデルです。
ALMA-7B-Ja-V2 is a machine translation model that uses ALMA's learning method to translate Japanese to English.
ALMA-7B-Ja-V2-GPTQ-Ja-Enは量子化、つまり多少の性能は落ちますがサイズを小さくし、実行速度を早くし、使いやすくした版です。
ALMA-7B-Ja-V2-GPTQ-Ja-En is a quantized version, i.e., it is smaller in size, faster in execution, and easier to use, with some performance loss.
## サンプルコード
Googleアカウントをお持ちの方は以下のColabを使用して無料で動かす事が出来ます。
If you have a Google account, you can run it for free using the following Colab.
リンク先で「Open In Colab」ボタンを押してColabを起動してください
Click the "Open In Colab" button on the link to start Colab.
[Free Colab Sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_V2_GPTQ_Ja_En_Free_Colab_sample.ipynb)
テキストファイル全体を一気に翻訳したい方は、以下のColabをお試しください。
If you want to translate the entire file at once, try Colab below.
[ALMA_7B_Ja_V2_GPTQ_Ja_En_batch_translation_sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_V2_GPTQ_Ja_En_batch_translation_sample.ipynb)
以下のようなエラーが発生した場合は
if you enconter error below.
```RuntimeError: probability tensor contains either `inf`, `nan` or element < 0```
It's mean your memory is not enough. decrease your num_beams or token size or reduce target text length.
これはメモリ不足を意味します。num_beamsかtoken size、もしくは翻訳対象の文の長さを減らしてください。
## その他の版 Other Version
- 元のモデル [ALMA-7B-Ja-V2](https://huggingface.co/webbigdata/ALMA-7B-Ja-V2).
- original [ALMA-7B-Ja-V2](https://huggingface.co/webbigdata/ALMA-7B-Ja-V2).
## 本作業について about this work
- 本作業は[webbigdata](https://webbigdata.jp/)によって行われました
- **This work was done by :** [webbigdata](https://webbigdata.jp/).
**ALMA** (**A**dvanced **L**anguage **M**odel-based tr**A**nslator) is an LLM-based translation model, which adopts a new translation model paradigm: it begins with fine-tuning on monolingual data and is further optimized using high-quality parallel data. This two-step fine-tuning process ensures strong translation performance.
Please find more details in their [paper](https://arxiv.org/abs/2309.11674).
```
@misc{xu2023paradigm,
title={A Paradigm Shift in Machine Translation: Boosting Translation Performance of Large Language Models},
author={Haoran Xu and Young Jin Kim and Amr Sharaf and Hany Hassan Awadalla},
year={2023},
eprint={2309.11674},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
量子化設定 gptq 4bit/128G
Quantization settings gptq 4bit/128G |