dranger003 commited on
Commit
f3d504a
1 Parent(s): 4cdd623

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +72 -29
README.md CHANGED
@@ -15,45 +15,88 @@ GGUF version of [c4ai-command-r7b-12-2024](https://huggingface.co/CohereForAI/c4
15
 
16
  https://github.com/ggerganov/llama.cpp/issues/10816#issuecomment-2548574766
17
 
 
18
  ```
19
- llama_new_context_with_model: n_seq_max = 1
20
- llama_new_context_with_model: n_ctx = 8192
21
- llama_new_context_with_model: n_ctx_per_seq = 8192
22
- llama_new_context_with_model: n_batch = 2048
23
- llama_new_context_with_model: n_ubatch = 512
24
- llama_new_context_with_model: flash_attn = 1
25
- llama_new_context_with_model: freq_base = 50000.0
26
- llama_new_context_with_model: freq_scale = 1
27
- llama_kv_cache_init: CPU KV buffer size = 1024.00 MiB
28
- llama_new_context_with_model: KV self size = 1024.00 MiB, K (f16): 512.00 MiB, V (f16): 512.00 MiB
29
- llama_new_context_with_model: CPU output buffer size = 0.98 MiB
30
- llama_new_context_with_model: CUDA0 compute buffer size = 1328.31 MiB
31
- llama_new_context_with_model: CUDA_Host compute buffer size = 24.01 MiB
32
- llama_new_context_with_model: graph nodes = 841
33
- llama_new_context_with_model: graph splits = 324 (with bs=512), 1 (with bs=1)
34
- common_init_from_params: setting dry_penalty_last_n to ctx_size = 8192
35
- common_init_from_params: warming up the model with an empty run - please wait ... (--no-warmup to disable)
36
  main: llama threadpool init, n_threads = 16
37
 
38
- system_info: n_threads = 16 (n_threads_batch = 16) / 32 | CUDA : ARCHS = 890 | USE_GRAPHS = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX_VNNI = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | AVX512_BF16 = 1 | AMX_INT8 = 1 | LLAMAFILE = 1 | OPENMP = 1 | AARCH64_REPACK = 1 |
39
 
40
- sampler seed: 2760461191
41
  sampler params:
42
  repeat_last_n = 64, repeat_penalty = 1.000, frequency_penalty = 0.000, presence_penalty = 0.000
43
- dry_multiplier = 0.000, dry_base = 1.750, dry_allowed_length = 2, dry_penalty_last_n = 8192
44
- top_k = 40, top_p = 0.950, min_p = 0.050, xtc_probability = 0.000, xtc_threshold = 0.100, typical_p = 1.000, temp = 0.800
45
  mirostat = 0, mirostat_lr = 0.100, mirostat_ent = 5.000
46
  sampler chain: logits -> logit-bias -> penalties -> dry -> top-k -> typical -> top-p -> min-p -> xtc -> temp-ext -> dist
47
- generate: n_ctx = 8192, n_batch = 2048, n_predict = -1, n_keep = 1
48
 
49
- I am Command, a sophisticated large language model built by the company Cohere. I assist users by providing thorough responses to a wide range of queries, offering information, and performing various tasks. My capabilities include answering questions, generating text, summarizing content, extracting data, and performing various other tasks based on the user's requirements.
 
50
 
51
- I strive to provide accurate and helpful information while ensuring a positive and informative user experience. Feel free to ask me about any topic, and I'll do my best to assist you! [end of text]
52
 
 
53
 
54
- llama_perf_sampler_print: sampling time = 15.07 ms / 128 runs ( 0.12 ms per token, 8491.44 tokens per second)
55
- llama_perf_context_print: load time = 1076.84 ms
56
- llama_perf_context_print: prompt eval time = 181.62 ms / 22 tokens ( 8.26 ms per token, 121.13 tokens per second)
57
- llama_perf_context_print: eval time = 4938.01 ms / 105 runs ( 47.03 ms per token, 21.26 tokens per second)
58
- llama_perf_context_print: total time = 5163.42 ms / 127 tokens
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
59
  ```
 
15
 
16
  https://github.com/ggerganov/llama.cpp/issues/10816#issuecomment-2548574766
17
 
18
+ Example [Tool Use](https://docs.cohere.com/docs/tool-use)
19
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
  main: llama threadpool init, n_threads = 16
21
 
22
+ system_info: n_threads = 16 (n_threads_batch = 16) / 32 | CUDA : ARCHS = 890 | USE_GRAPHS = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | LLAMAFILE = 1 | OPENMP = 1 | AARCH64_REPACK = 1 |
23
 
24
+ sampler seed: 2269839035
25
  sampler params:
26
  repeat_last_n = 64, repeat_penalty = 1.000, frequency_penalty = 0.000, presence_penalty = 0.000
27
+ dry_multiplier = 0.000, dry_base = 1.750, dry_allowed_length = 2, dry_penalty_last_n = 16384
28
+ top_k = 40, top_p = 0.950, min_p = 0.050, xtc_probability = 0.000, xtc_threshold = 0.100, typical_p = 1.000, temp = 0.000
29
  mirostat = 0, mirostat_lr = 0.100, mirostat_ent = 5.000
30
  sampler chain: logits -> logit-bias -> penalties -> dry -> top-k -> typical -> top-p -> min-p -> xtc -> temp-ext -> dist
31
+ generate: n_ctx = 16384, n_batch = 2048, n_predict = -1, n_keep = 1
32
 
33
+ <BOS_TOKEN><|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|># System Preamble
34
+ You are in contextual safety mode. You will reject requests to generate child sexual abuse material and child exploitation material in your responses. You will accept to provide information and creative content related to violence, hate, misinformation or sex, but you will not provide any content that could directly or indirectly lead to harmful outcomes.
35
 
36
+ Your information cutoff date is June 2024.
37
 
38
+ You have been trained on data in English, French, Spanish, Italian, German, Portuguese, Japanese, Korean, Modern Standard Arabic, Mandarin, Russian, Indonesian, Turkish, Dutch, Polish, Persian, Vietnamese, Czech, Hindi, Ukrainian, Romanian, Greek and Hebrew but have the ability to speak many more languages.
39
 
40
+ You have been trained to have advanced reasoning and tool-use capabilities and you should make best use of these skills to serve user's requests.
41
+
42
+ ## Tool Use
43
+ Think about how you can make best use of the provided tools to help with the task and come up with a high level plan that you will execute first.
44
+
45
+ 0. Start by writing <|START_THINKING|> followed by a detailed step by step plan of how you will solve the problem. For each step explain your thinking fully and give details of required tool calls (if needed). Unless specified otherwise, you write your plan in natural language. When you finish, close it out with <|END_THINKING|>.
46
+ You can optionally choose to skip this step when the user request is so straightforward to address that only a trivial plan would be needed.
47
+ NOTE: You MUST skip this step when you are directly responding to the user's request without using any tools.
48
+
49
+ Then carry out your plan by repeatedly executing the following steps.
50
+ 1. Action: write <|START_ACTION|> followed by a list of JSON-formatted tool calls, with each one containing "tool_name" and "parameters" fields.
51
+ When there are multiple tool calls which are completely independent of each other (i.e. they can be executed in parallel), you should list them out all together in one step. When you finish, close it out with <|END_ACTION|>.
52
+ 2. Observation: you will then receive results of those tool calls in JSON format in the very next turn, wrapped around by <|START_TOOL_RESULT|> and <|END_TOOL_RESULT|>. Carefully observe those results and think about what to do next. Note that these results will be provided to you in a separate turn. NEVER hallucinate results.
53
+ Every tool call produces a list of results (when a tool call produces no result or a single result, it'll still get wrapped inside a list). Each result is clearly linked to its originating tool call via its "tool_call_id".
54
+ 3. Reflection: start the next turn by writing <|START_THINKING|> followed by what you've figured out so far, any changes you need to make to your plan, and what you will do next. When you finish, close it out with <|END_THINKING|>.
55
+ You can optionally choose to skip this step when everything is going according to plan and no special pieces of information or reasoning chains need to be recorded.
56
+ NOTE: You MUST skip this step when you are done with tool-use actions and are ready to respond to the user.
57
+
58
+ You can repeat the above 3 steps multiple times (could be 0 times too if no suitable tool calls are available or needed), until you decide it's time to finally respond to the user.
59
+
60
+ 4. Response: then break out of the loop and write <|START_RESPONSE|> followed by a piece of text which serves as a response to the user's last request. Use all previous tool calls and results to help you when formulating your response. When you finish, close it out with <|END_RESPONSE|>.
61
+
62
+ ## Available Tools
63
+ Here is the list of tools that you have available to you.
64
+ You can ONLY use the tools listed here. When a tool is not listed below, it is NOT available and you should NEVER attempt to use it.
65
+ Each tool is represented as a JSON object with fields like "name", "description", "parameters" (per JSON Schema), and optionally, "responses" (per JSON Schema).
66
+
67
+ ```json
68
+ [
69
+ {"name": "get_current_time", "description": "Function to retrieve the current date and time as a formatted string", "parameters": {}, "responses": null}
70
+ ]
71
+ ``
72
+
73
+
74
+ # Default Preamble
75
+ The following instructions are your defaults unless specified elsewhere in developer preamble or user prompt.
76
+ - Your name is Command.
77
+ - You are a large language model built by Cohere.
78
+ - You reply conversationally with a friendly and informative tone and often include introductory statements and follow-up questions.
79
+ - If the input is ambiguous, ask clarifying follow-up questions.
80
+ - Use Markdown-specific formatting in your response (for example to highlight phrases in bold or italics, create tables, or format code blocks).
81
+ - Use LaTeX to generate mathematical notation for complex equations.
82
+ - When responding in English, use American English unless context indicates otherwise.
83
+ - When outputting responses of more than seven sentences, split the response into paragraphs.
84
+ - Prefer the active voice.
85
+ - Adhere to the APA style guidelines for punctuation, spelling, hyphenation, capitalization, numbers, lists, and quotation marks. Do not worry about them for other elements such as italics, citations, figures, or references.
86
+ - Use gender-neutral pronouns for unspecified persons.
87
+ - Limit lists to no more than 10 items unless the list is a set of finite instructions, in which case complete the list.
88
+ - Use the third person when asked to write a summary.
89
+ - When asked to extract values from source material, use the exact form, separated by commas.
90
+ - When generating code output, please provide an explanation after the code.
91
+ - When generating code output without specifying the programming language, please generate Python code.
92
+ - If you are asked a question that requires reasoning, first think through your answer, slowly and step by step, then answer.<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|USER_TOKEN|>What time is it?<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|><|START_THINKING|>I will use the get_current_time tool to find out what time it is.<|END_THINKING|><|START_ACTION|>[
93
+ {"tool_call_id": "0", "tool_name": "get_current_time", "parameters": {}}
94
+ ]<|END_ACTION|><|END_OF_TURN_TOKEN|> [end of text]
95
+
96
+
97
+ llama_perf_sampler_print: sampling time = 46.08 ms / 1287 runs ( 0.04 ms per token, 27929.08 tokens per second)
98
+ llama_perf_context_print: load time = 9836.34 ms
99
+ llama_perf_context_print: prompt eval time = 304.06 ms / 1233 tokens ( 0.25 ms per token, 4055.11 tokens per second)
100
+ llama_perf_context_print: eval time = 721.01 ms / 53 runs ( 13.60 ms per token, 73.51 tokens per second)
101
+ llama_perf_context_print: total time = 1408.60 ms / 1286 tokens
102
  ```