Upload README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ license: mit
|
|
5 |
model_creator: haoranxu
|
6 |
model_name: ALMA 13B Pretrain
|
7 |
model_type: llama
|
8 |
-
prompt_template: 'Translate this from Chinese to English
|
9 |
|
10 |
Chinese: {prompt}
|
11 |
|
@@ -57,7 +57,7 @@ Multiple GPTQ parameter permutations are provided; see Provided Files below for
|
|
57 |
## Prompt template: ALMA
|
58 |
|
59 |
```
|
60 |
-
Translate this from Chinese to English
|
61 |
Chinese: {prompt}
|
62 |
English:
|
63 |
|
@@ -231,7 +231,7 @@ model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
|
|
231 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
232 |
|
233 |
prompt = "Tell me about AI"
|
234 |
-
prompt_template=f'''Translate this from Chinese to English
|
235 |
Chinese: {prompt}
|
236 |
English:
|
237 |
'''
|
|
|
5 |
model_creator: haoranxu
|
6 |
model_name: ALMA 13B Pretrain
|
7 |
model_type: llama
|
8 |
+
prompt_template: 'Translate this from Chinese to English:
|
9 |
|
10 |
Chinese: {prompt}
|
11 |
|
|
|
57 |
## Prompt template: ALMA
|
58 |
|
59 |
```
|
60 |
+
Translate this from Chinese to English:
|
61 |
Chinese: {prompt}
|
62 |
English:
|
63 |
|
|
|
231 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
232 |
|
233 |
prompt = "Tell me about AI"
|
234 |
+
prompt_template=f'''Translate this from Chinese to English:
|
235 |
Chinese: {prompt}
|
236 |
English:
|
237 |
'''
|