TheBloke commited on
Commit
f95ea5a
1 Parent(s): c95455f

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -18
README.md CHANGED
@@ -5,16 +5,9 @@ license: llama2
5
  model_creator: Xwin-LM
6
  model_name: Xwin-LM 7B V0.1
7
  model_type: llama
8
- prompt_template: 'Below is an instruction that describes a task. Write a response
9
- that appropriately completes the request.
10
-
11
-
12
- ### Instruction:
13
-
14
- {prompt}
15
-
16
-
17
- ### Response:
18
 
19
  '
20
  quantized_by: TheBloke
@@ -75,15 +68,10 @@ Here is an incomplate list of clients and libraries that are known to support GG
75
  <!-- repositories-available end -->
76
 
77
  <!-- prompt-template start -->
78
- ## Prompt template: Alpaca
79
 
80
  ```
81
- Below is an instruction that describes a task. Write a response that appropriately completes the request.
82
-
83
- ### Instruction:
84
- {prompt}
85
-
86
- ### Response:
87
 
88
  ```
89
 
@@ -199,7 +187,7 @@ Windows Command Line users: You can set the environment variable by running `set
199
  Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
200
 
201
  ```shell
202
- ./main -ngl 32 -m xwin-lm-7b-v0.1.Q4_K_M.gguf --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{prompt}\n\n### Response:"
203
  ```
204
 
205
  Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
 
5
  model_creator: Xwin-LM
6
  model_name: Xwin-LM 7B V0.1
7
  model_type: llama
8
+ prompt_template: 'A chat between a curious user and an artificial intelligence assistant.
9
+ The assistant gives helpful, detailed, and polite answers to the user''s questions.
10
+ USER: {prompt} ASSISTANT:
 
 
 
 
 
 
 
11
 
12
  '
13
  quantized_by: TheBloke
 
68
  <!-- repositories-available end -->
69
 
70
  <!-- prompt-template start -->
71
+ ## Prompt template: Vicuna
72
 
73
  ```
74
+ A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {prompt} ASSISTANT:
 
 
 
 
 
75
 
76
  ```
77
 
 
187
  Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
188
 
189
  ```shell
190
+ ./main -ngl 32 -m xwin-lm-7b-v0.1.Q4_K_M.gguf --color -c 4096 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {prompt} ASSISTANT:"
191
  ```
192
 
193
  Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.