--- name: "llama3-instruct-grammar" config_file: | mmap: true function: disable_no_action: true grammar: no_mixed_free_string: true mixed_mode: true schema_type: llama3.1 # or JSON is supported too (json) response_regex: - \w+)>(?P.*) template: chat_message: | <|start_header_id|>{{if eq .RoleName "assistant"}}assistant{{else if eq .RoleName "system"}}system{{else if eq .RoleName "tool"}}tool{{else if eq .RoleName "user"}}user{{end}}<|end_header_id|> {{ if .FunctionCall -}} Function call: {{ else if eq .RoleName "tool" -}} Function response: {{ end -}} {{ if .Content -}} {{.Content -}} {{ else if .FunctionCall -}} {{ toJson .FunctionCall -}} {{ end -}} <|eot_id|> function: | <|start_header_id|>system<|end_header_id|> You have access to the following functions: {{range .Functions}} Use the function '{{.Name}}' to '{{.Description}}' {{toJson .Parameters}} {{end}} Think very carefully before calling functions. If a you choose to call a function ONLY reply in the following format with no prefix or suffix: {{`{{"example_name": "example_value"}}`}} Reminder: - If looking for real time information use relevant functions before falling back to searching on internet - Function calls MUST follow the specified format, start with - Required parameters MUST be specified - Only call one function at a time - Put the entire function call reply on one line <|eot_id|> {{.Input }} <|start_header_id|>assistant<|end_header_id|> chat: | {{.Input }} <|start_header_id|>assistant<|end_header_id|> completion: | {{.Input}} context_size: 8192 f16: true stopwords: - <|im_end|> - - "<|eot_id|>" - <|end_of_text|>