stvlynn commited on
Commit
0b22261
1 Parent(s): 2ea90cb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -0
README.md CHANGED
@@ -17,6 +17,33 @@ Qwen-7B-Chat-Cantonese is a fine-tuned version based on Qwen-7B-Chat, trained on
17
 
18
  Qwen-7B-Chat-Cantonese係基於Qwen-7B-Chat嘅微調版本,基於大量粵語數據進行訓練。
19
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  ## Training Parameters
21
 
22
  | Parameter | Description | Value |
 
17
 
18
  Qwen-7B-Chat-Cantonese係基於Qwen-7B-Chat嘅微調版本,基於大量粵語數據進行訓練。
19
 
20
+ ## Usage
21
+
22
+ ### Requirements
23
+
24
+ * python 3.8 and above
25
+ * pytorch 1.12 and above, 2.0 and above are recommended
26
+ * CUDA 11.4 and above are recommended (this is for GPU users, flash-attention users, etc.)
27
+
28
+ ### Dependency
29
+
30
+ To run Qwen-7B-Chat-Cantonese, please make sure you meet the above requirements, and then execute the following pip commands to install the dependent libraries.
31
+
32
+ ```bash
33
+ pip install transformers==4.32.0 accelerate tiktoken einops scipy transformers_stream_generator==0.0.4 peft deepspeed
34
+ ```
35
+
36
+ In addition, it is recommended to install the `flash-attention` library (**we support flash attention 2 now.**) for higher efficiency and lower memory usage.
37
+
38
+ ```bash
39
+ git clone https://github.com/Dao-AILab/flash-attention
40
+ cd flash-attention && pip install .
41
+ ```
42
+
43
+ ### Quickstart
44
+
45
+ Pls turn to QwenLM/Qwen - [Quickstart](https://github.com/QwenLM/Qwen?tab=readme-ov-file#quickstart)
46
+
47
  ## Training Parameters
48
 
49
  | Parameter | Description | Value |