andreaskoepf
commited on
Commit
•
dd8d1fd
1
Parent(s):
efd290b
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,59 @@
|
|
1 |
---
|
2 |
license: llama2
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: llama2
|
3 |
+
datasets:
|
4 |
+
- OpenAssistant/oasst1
|
5 |
+
- shahules786/orca-best
|
6 |
---
|
7 |
+
# Open-Assistant CodeLlama 13B SFT v10
|
8 |
+
|
9 |
+
This model is an Open-Assistant fine-tuning of Meta's CodeLlama 13B LLM.
|
10 |
+
|
11 |
+
## Model Details
|
12 |
+
|
13 |
+
- **Finetuned from:** [codellama](https://github.com/facebookresearch/codellama) via [epfLLM/Megatron-LLM](https://github.com/epfLLM/Megatron-LLM)
|
14 |
+
- **Model type:** Causal decoder-only transformer language model
|
15 |
+
- **Language:** English (and limited capabilities in German, Spanish, French, Italian, Portuguese, Polish, Dutch, Romanian, Czech, Swedish)
|
16 |
+
- **Weights & Biases training logs:** 6123 steps [run56_oa_llamacode](https://wandb.ai/open-assistant/public-sft/runs/run56_oa_llamacode)
|
17 |
+
- **Demo:** [Continuations for 250 random prompts (without system message)](https://open-assistant.github.io/oasst-model-eval/?f=https%3A%2F%2Fraw.githubusercontent.com%2FOpen-Assistant%2Foasst-model-eval%2Fmain%2Fsampling_reports%2Foasst-sft%2F2023-08-26_OpenAssistant_codellama-13b-oasst-sft-v10_sampling_noprefix2.json)
|
18 |
+
- **License:** [LLAMA 2 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta-llama/Llama-2-70b/raw/main/LICENSE.txt)
|
19 |
+
- **Contact:** [Open-Assistant Discord](https://ykilcher.com/open-assistant-discord)
|
20 |
+
|
21 |
+
## Prompting / Prompt Template
|
22 |
+
|
23 |
+
Due to public demand (see [survey](https://twitter.com/erhartford/status/1682403597525430272)) we changed the prompt-template for this model from custom prompter/assistant tokens to OpenAI's [chatml](https://github.com/openai/openai-python/blob/main/chatml.md) standard prompt format.
|
24 |
+
We hope that this leads to greater compatibility with chat inference/frontend applications.
|
25 |
+
|
26 |
+
Prompt dialogue template:
|
27 |
+
|
28 |
+
```
|
29 |
+
"""
|
30 |
+
<|im_start|>system
|
31 |
+
{system_message}<|im_end|>
|
32 |
+
<|im_start|>user
|
33 |
+
{prompt}<|im_end|>
|
34 |
+
<|im_start|>assistant
|
35 |
+
"""
|
36 |
+
```
|
37 |
+
|
38 |
+
The model input can contain multiple conversation turns between user and assistant, e.g.
|
39 |
+
```
|
40 |
+
<|im_start|>user
|
41 |
+
{prompt 1}<|im_end|>
|
42 |
+
<|im_start|>assistant
|
43 |
+
{reply 1}<|im_end|>
|
44 |
+
<|im_start|>user
|
45 |
+
{prompt 2}<|im_end|>
|
46 |
+
<|im_start|>assistant
|
47 |
+
(...)
|
48 |
+
```
|
49 |
+
|
50 |
+
The model was partly trained with orca system messages.
|
51 |
+
For inference we recommend to use the official [Llama2 system message](https://github.com/facebookresearch/llama/blob/ea9f33d6d3ea8ed7d560d270986407fd6c2e52b7/example_chat_completion.py#L57-L61):
|
52 |
+
```
|
53 |
+
<|im_start|>system
|
54 |
+
You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.
|
55 |
+
|
56 |
+
If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.
|
57 |
+
<|im_end|>
|
58 |
+
```
|
59 |
+
|