TheBloke commited on
Commit
7a1cec0
1 Parent(s): 9fed75e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -26
README.md CHANGED
@@ -26,9 +26,9 @@ pipeline_tag: text-generation
26
  </div>
27
  <!-- header end -->
28
 
29
- # FreeWilly 2 - GPTQ
30
  - Model creator: [Stability AI](https://huggingface.co/stabilityai)
31
- - Original model: [FreeWilly 2](https://huggingface.co/stabilityai/FreeWilly2)
32
 
33
  ## Description
34
 
@@ -40,8 +40,9 @@ None
40
 
41
  ## Repositories available
42
 
43
- * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/FreeWilly2-GPTQ)
44
- * [Stability AI's original unquantised fp32 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/stabilityai/FreeWilly2)
 
45
 
46
  ## Prompt template: Orca-Hashes
47
 
@@ -211,25 +212,25 @@ Thank you to all my generous patrons and donaters!
211
 
212
  <!-- footer end -->
213
 
214
- # Original model card: Stability AI's FreeWilly 2
215
 
216
- # FreeWilly
217
 
218
  ## Model Description
219
 
220
- `FreeWilly2` is a Llama2 70B model finetuned on an Orca style Dataset
221
 
222
  ## Usage
223
 
224
- Start chatting with `FreeWilly2` using the following code snippet:
225
 
226
  ```python
227
  import torch
228
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
229
 
230
- tokenizer = AutoTokenizer.from_pretrained("stabilityai/FreeWilly2", use_fast=False)
231
- model = AutoModelForCausalLM.from_pretrained("stabilityai/FreeWilly2", torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
232
- system_prompt = "### System:\nYou are Free Willy, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
233
 
234
  message = "Write me a poem please"
235
  prompt = f"{system_prompt}### User: {message}\n\n### Assistant:\n"
@@ -239,7 +240,7 @@ output = model.generate(**inputs, do_sample=True, top_p=0.95, top_k=0, max_new_t
239
  print(tokenizer.decode(output[0], skip_special_tokens=True))
240
  ```
241
 
242
- FreeWilly should be used with this prompt format:
243
  ```
244
  ### System:
245
  This is a system prompt, please behave and help the user.
@@ -247,22 +248,22 @@ This is a system prompt, please behave and help the user.
247
  ### User:
248
  Your prompt here
249
 
250
- ### Assistant
251
- The output of FreeWilly2
252
  ```
253
 
254
  ## Model Details
255
 
256
  * **Developed by**: [Stability AI](https://stability.ai/)
257
- * **Model type**: FreeWilly is an auto-regressive language model fine-tuned on Llama2 70B.
258
  * **Language(s)**: English
259
  * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
260
- * **License**: Fine-tuned checkpoints (`FreeWilly2`) is licensed under the Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
261
  * **Contact**: For questions and comments about the model, please email `lm@stability.ai`
262
 
263
  ### Training Dataset
264
 
265
- `FreeWilly2` is trained on our internal Orca-style dataset
266
 
267
  ### Training Procedure
268
 
@@ -273,15 +274,9 @@ Models are learned via supervised fine-tuning on the aforementioned datasets, tr
273
  | Orca pt1 packed | 256 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
274
  | Orca pt2 unpacked | 512 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
275
 
276
- ## Use and Limitations
277
-
278
- ### Intended Use
279
-
280
- These models are intended for research only, in adherence with the [CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/) license.
281
 
282
- ### Limitations and bias
283
-
284
- Although the aforementioned dataset helps to steer the base language models into "safer" distributions of text, not all biases and toxicity can be mitigated through fine-tuning. We ask that users be mindful of such potential issues that can arise in generated responses. Do not treat model outputs as substitutes for human judgment or as sources of truth. Please use it responsibly.
285
 
286
  ## Citations
287
 
@@ -305,4 +300,4 @@ Although the aforementioned dataset helps to steer the base language models into
305
  archivePrefix={arXiv},
306
  primaryClass={cs.CL}
307
  }
308
- ```
 
26
  </div>
27
  <!-- header end -->
28
 
29
+ # StableBeluga 2 - GGML
30
  - Model creator: [Stability AI](https://huggingface.co/stabilityai)
31
+ - Original model: [StableBeluga 2](https://huggingface.co/stabilityai/StableBeluga2)
32
 
33
  ## Description
34
 
 
40
 
41
  ## Repositories available
42
 
43
+ * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/StableBeluga2-GPTQ)
44
+ * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/StableBeluga2-GGML)
45
+ * [Stability AI's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/stabilityai/StableBeluga2)
46
 
47
  ## Prompt template: Orca-Hashes
48
 
 
212
 
213
  <!-- footer end -->
214
 
215
+ # Original model card: Stability AI's StableBeluga 2
216
 
217
+ # Stable Beluga 2
218
 
219
  ## Model Description
220
 
221
+ `Stable Beluga 2` is a Llama2 70B model finetuned on an Orca style Dataset
222
 
223
  ## Usage
224
 
225
+ Start chatting with `Stable Beluga 2` using the following code snippet:
226
 
227
  ```python
228
  import torch
229
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
230
 
231
+ tokenizer = AutoTokenizer.from_pretrained("stabilityai/StableBeluga2", use_fast=False)
232
+ model = AutoModelForCausalLM.from_pretrained("stabilityai/StableBeluga2", torch_dtype=torch.float16, low_cpu_mem_usage=True, device_map="auto")
233
+ system_prompt = "### System:\nYou are Stable Beluga, an AI that follows instructions extremely well. Help as much as you can. Remember, be safe, and don't do anything illegal.\n\n"
234
 
235
  message = "Write me a poem please"
236
  prompt = f"{system_prompt}### User: {message}\n\n### Assistant:\n"
 
240
  print(tokenizer.decode(output[0], skip_special_tokens=True))
241
  ```
242
 
243
+ Stable Beluga 2 should be used with this prompt format:
244
  ```
245
  ### System:
246
  This is a system prompt, please behave and help the user.
 
248
  ### User:
249
  Your prompt here
250
 
251
+ ### Assistant:
252
+ The output of Stable Beluga 2
253
  ```
254
 
255
  ## Model Details
256
 
257
  * **Developed by**: [Stability AI](https://stability.ai/)
258
+ * **Model type**: Stable Beluga 2 is an auto-regressive language model fine-tuned on Llama2 70B.
259
  * **Language(s)**: English
260
  * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
261
+ * **License**: Fine-tuned checkpoints (`Stable Beluga 2`) is licensed under the [STABLE BELUGA NON-COMMERCIAL COMMUNITY LICENSE AGREEMENT](https://huggingface.co/stabilityai/StableBeluga2/blob/main/LICENSE.txt)
262
  * **Contact**: For questions and comments about the model, please email `lm@stability.ai`
263
 
264
  ### Training Dataset
265
 
266
+ ` Stable Beluga 2` is trained on our internal Orca-style dataset
267
 
268
  ### Training Procedure
269
 
 
274
  | Orca pt1 packed | 256 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
275
  | Orca pt2 unpacked | 512 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
276
 
277
+ ## Ethical Considerations and Limitations
 
 
 
 
278
 
279
+ Beluga is a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Beluga's potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Beluga, developers should perform safety testing and tuning tailored to their specific applications of the model.
 
 
280
 
281
  ## Citations
282
 
 
300
  archivePrefix={arXiv},
301
  primaryClass={cs.CL}
302
  }
303
+ ```