Update README.md
Browse files
README.md
CHANGED
@@ -17,6 +17,33 @@ Qwen-7B-Chat-Cantonese is a fine-tuned version based on Qwen-7B-Chat, trained on
|
|
17 |
|
18 |
Qwen-7B-Chat-Cantonese係基於Qwen-7B-Chat嘅微調版本,基於大量粵語數據進行訓練。
|
19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
## Training Parameters
|
21 |
|
22 |
| Parameter | Description | Value |
|
|
|
17 |
|
18 |
Qwen-7B-Chat-Cantonese係基於Qwen-7B-Chat嘅微調版本,基於大量粵語數據進行訓練。
|
19 |
|
20 |
+
## Usage
|
21 |
+
|
22 |
+
### Requirements
|
23 |
+
|
24 |
+
* python 3.8 and above
|
25 |
+
* pytorch 1.12 and above, 2.0 and above are recommended
|
26 |
+
* CUDA 11.4 and above are recommended (this is for GPU users, flash-attention users, etc.)
|
27 |
+
|
28 |
+
### Dependency
|
29 |
+
|
30 |
+
To run Qwen-7B-Chat-Cantonese, please make sure you meet the above requirements, and then execute the following pip commands to install the dependent libraries.
|
31 |
+
|
32 |
+
```bash
|
33 |
+
pip install transformers==4.32.0 accelerate tiktoken einops scipy transformers_stream_generator==0.0.4 peft deepspeed
|
34 |
+
```
|
35 |
+
|
36 |
+
In addition, it is recommended to install the `flash-attention` library (**we support flash attention 2 now.**) for higher efficiency and lower memory usage.
|
37 |
+
|
38 |
+
```bash
|
39 |
+
git clone https://github.com/Dao-AILab/flash-attention
|
40 |
+
cd flash-attention && pip install .
|
41 |
+
```
|
42 |
+
|
43 |
+
### Quickstart
|
44 |
+
|
45 |
+
Pls turn to QwenLM/Qwen - [Quickstart](https://github.com/QwenLM/Qwen?tab=readme-ov-file#quickstart)
|
46 |
+
|
47 |
## Training Parameters
|
48 |
|
49 |
| Parameter | Description | Value |
|