andreaskoepf commited on
Commit
dd8d1fd
1 Parent(s): efd290b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -0
README.md CHANGED
@@ -1,3 +1,59 @@
1
  ---
2
  license: llama2
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: llama2
3
+ datasets:
4
+ - OpenAssistant/oasst1
5
+ - shahules786/orca-best
6
  ---
7
+ # Open-Assistant CodeLlama 13B SFT v10
8
+
9
+ This model is an Open-Assistant fine-tuning of Meta's CodeLlama 13B LLM.
10
+
11
+ ## Model Details
12
+
13
+ - **Finetuned from:** [codellama](https://github.com/facebookresearch/codellama) via [epfLLM/Megatron-LLM](https://github.com/epfLLM/Megatron-LLM)
14
+ - **Model type:** Causal decoder-only transformer language model
15
+ - **Language:** English (and limited capabilities in German, Spanish, French, Italian, Portuguese, Polish, Dutch, Romanian, Czech, Swedish)
16
+ - **Weights & Biases training logs:** 6123 steps [run56_oa_llamacode](https://wandb.ai/open-assistant/public-sft/runs/run56_oa_llamacode)
17
+ - **Demo:** [Continuations for 250 random prompts (without system message)](https://open-assistant.github.io/oasst-model-eval/?f=https%3A%2F%2Fraw.githubusercontent.com%2FOpen-Assistant%2Foasst-model-eval%2Fmain%2Fsampling_reports%2Foasst-sft%2F2023-08-26_OpenAssistant_codellama-13b-oasst-sft-v10_sampling_noprefix2.json)
18
+ - **License:** [LLAMA 2 COMMUNITY LICENSE AGREEMENT](https://huggingface.co/meta-llama/Llama-2-70b/raw/main/LICENSE.txt)
19
+ - **Contact:** [Open-Assistant Discord](https://ykilcher.com/open-assistant-discord)
20
+
21
+ ## Prompting / Prompt Template
22
+
23
+ Due to public demand (see [survey](https://twitter.com/erhartford/status/1682403597525430272)) we changed the prompt-template for this model from custom prompter/assistant tokens to OpenAI's [chatml](https://github.com/openai/openai-python/blob/main/chatml.md) standard prompt format.
24
+ We hope that this leads to greater compatibility with chat inference/frontend applications.
25
+
26
+ Prompt dialogue template:
27
+
28
+ ```
29
+ """
30
+ <|im_start|>system
31
+ {system_message}<|im_end|>
32
+ <|im_start|>user
33
+ {prompt}<|im_end|>
34
+ <|im_start|>assistant
35
+ """
36
+ ```
37
+
38
+ The model input can contain multiple conversation turns between user and assistant, e.g.
39
+ ```
40
+ <|im_start|>user
41
+ {prompt 1}<|im_end|>
42
+ <|im_start|>assistant
43
+ {reply 1}<|im_end|>
44
+ <|im_start|>user
45
+ {prompt 2}<|im_end|>
46
+ <|im_start|>assistant
47
+ (...)
48
+ ```
49
+
50
+ The model was partly trained with orca system messages.
51
+ For inference we recommend to use the official [Llama2 system message](https://github.com/facebookresearch/llama/blob/ea9f33d6d3ea8ed7d560d270986407fd6c2e52b7/example_chat_completion.py#L57-L61):
52
+ ```
53
+ <|im_start|>system
54
+ You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.
55
+
56
+ If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.
57
+ <|im_end|>
58
+ ```
59
+