flyingfishinwater commited on
Commit
bafc0de
·
verified ·
1 Parent(s): 4a4b250

Upload models.json

Browse files
Files changed (1) hide show
  1. models.json +100 -14
models.json CHANGED
@@ -1,12 +1,12 @@
1
  [
2
  {
3
  "id": "Internal",
4
- "model_title": "Llama2 Lite",
5
  "model_file": "ggml-model-Q8_0.gguf",
6
  "model_url": "https://",
7
  "model_info_url": "https://huggingface.co/princeton-nlp/Sheared-LLaMA-1.3B",
8
  "model_avatar": "ava0",
9
- "model_description": "The standard Llama2 based 1.3B LLM.",
10
  "developer": "Meta",
11
  "developer_url": "https://ai.meta.com/llama/",
12
  "file_size": 1430,
@@ -19,29 +19,45 @@
19
  "n_batch" : 10,
20
  "template_name" : "HumanBot",
21
  "is_ready": true,
22
- "is_internal": true
 
 
 
 
 
 
 
 
23
  },
24
  {
25
  "id": "LiteLlama-460M-1T-Q8",
26
- "model_title": "LiteLlama-460M-1T",
27
  "model_file": "LiteLlama-460M-1T-Q8_0.gguf",
28
  "model_url": "https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/LiteLlama-460M-1T-Q8_0.gguf?download=true",
29
  "model_info_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
30
  "model_avatar": "logo_litellama",
31
- "model_description": "We present an open-source reproduction of Meta AI's LLaMa 2. However, with significantly reduced model sizes, LiteLlama-460M-1T has 460M parameters trained with 1T tokens.",
32
  "developer": "Xiaotian Han from Texas A&M University",
33
  "developer_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
34
  "file_size": 493,
35
- "context" : 4096,
36
  "temp" : 0.6,
37
- "prompt_format" : "<|system|>You are a friendly chatbot who always responds in the style of a pirate.</s><|user|>{{prompt}}</s><|assistant|>",
38
  "top_k" : 5,
39
  "top_p" : 0.9,
40
  "model_inference" : "llama",
41
  "n_batch" : 10,
42
  "template_name" : "TinyLlama",
43
  "is_ready": true,
44
- "is_internal": false
 
 
 
 
 
 
 
 
45
  },
46
  {
47
  "id": "tinyllama-1.1B-chat-Q8",
@@ -63,7 +79,15 @@
63
  "n_batch" : 10,
64
  "template_name" : "TinyLlama",
65
  "is_ready": true,
66
- "is_internal": false
 
 
 
 
 
 
 
 
67
  },
68
  {
69
  "id": "mistral-7b-instruct-v0.2-Q8",
@@ -85,7 +109,15 @@
85
  "n_batch" : 10,
86
  "template_name" : "Mistral",
87
  "is_ready": true,
88
- "is_internal": false
 
 
 
 
 
 
 
 
89
  },
90
  {
91
  "id": "openchat-3.5-1210-Q8",
@@ -107,7 +139,15 @@
107
  "n_batch" : 10,
108
  "template_name" : "Mistral",
109
  "is_ready": true,
110
- "is_internal": false
 
 
 
 
 
 
 
 
111
  },
112
  {
113
  "id": "phi-2",
@@ -129,7 +169,15 @@
129
  "n_batch" : 10,
130
  "template_name" : "PHI",
131
  "is_ready": true,
132
- "is_internal": false
 
 
 
 
 
 
 
 
133
  },
134
  {
135
  "id": "yi-6b",
@@ -151,6 +199,44 @@
151
  "n_batch" : 10,
152
  "template_name" : "yi",
153
  "is_ready": true,
154
- "is_internal": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
155
  }
156
- ]
 
1
  [
2
  {
3
  "id": "Internal",
4
+ "model_title": "AI Assistant",
5
  "model_file": "ggml-model-Q8_0.gguf",
6
  "model_url": "https://",
7
  "model_info_url": "https://huggingface.co/princeton-nlp/Sheared-LLaMA-1.3B",
8
  "model_avatar": "ava0",
9
+ "model_description": "It is an AI assistant who can talk with you and help solve simple problems. It's based on a lite LLAMA2 model developed by Meta Inc.",
10
  "developer": "Meta",
11
  "developer_url": "https://ai.meta.com/llama/",
12
  "file_size": 1430,
 
19
  "n_batch" : 10,
20
  "template_name" : "HumanBot",
21
  "is_ready": true,
22
+ "is_internal": true,
23
+ "use_metal": true,
24
+ "mlock": false,
25
+ "mmap": true,
26
+ "repeat_last_n": 64,
27
+ "repeat_penalty": 1.2,
28
+ "add_bos_token": true,
29
+ "add_eos_token": false,
30
+ "parse_special_tokens": true
31
  },
32
  {
33
  "id": "LiteLlama-460M-1T-Q8",
34
+ "model_title": "LiteLlama",
35
  "model_file": "LiteLlama-460M-1T-Q8_0.gguf",
36
  "model_url": "https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/LiteLlama-460M-1T-Q8_0.gguf?download=true",
37
  "model_info_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
38
  "model_avatar": "logo_litellama",
39
+ "model_description": "It's a very small LLAMA2 model with only 460M parameters trained with 1T tokens. It's best for testing.",
40
  "developer": "Xiaotian Han from Texas A&M University",
41
  "developer_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
42
  "file_size": 493,
43
+ "context" : 1024,
44
  "temp" : 0.6,
45
+ "prompt_format" : "<human>: {{prompt}}\n<bot>:",
46
  "top_k" : 5,
47
  "top_p" : 0.9,
48
  "model_inference" : "llama",
49
  "n_batch" : 10,
50
  "template_name" : "TinyLlama",
51
  "is_ready": true,
52
+ "is_internal": false,
53
+ "use_metal": true,
54
+ "mlock": false,
55
+ "mmap": true,
56
+ "repeat_last_n": 64,
57
+ "repeat_penalty": 1.2,
58
+ "add_bos_token": true,
59
+ "add_eos_token": false,
60
+ "parse_special_tokens": true
61
  },
62
  {
63
  "id": "tinyllama-1.1B-chat-Q8",
 
79
  "n_batch" : 10,
80
  "template_name" : "TinyLlama",
81
  "is_ready": true,
82
+ "is_internal": false,
83
+ "use_metal": true,
84
+ "mlock": false,
85
+ "mmap": true,
86
+ "repeat_last_n": 64,
87
+ "repeat_penalty": 1.2,
88
+ "add_bos_token": true,
89
+ "add_eos_token": false,
90
+ "parse_special_tokens": true
91
  },
92
  {
93
  "id": "mistral-7b-instruct-v0.2-Q8",
 
109
  "n_batch" : 10,
110
  "template_name" : "Mistral",
111
  "is_ready": true,
112
+ "is_internal": false,
113
+ "use_metal": true,
114
+ "mlock": false,
115
+ "mmap": true,
116
+ "repeat_last_n": 64,
117
+ "repeat_penalty": 1.2,
118
+ "add_bos_token": true,
119
+ "add_eos_token": false,
120
+ "parse_special_tokens": true
121
  },
122
  {
123
  "id": "openchat-3.5-1210-Q8",
 
139
  "n_batch" : 10,
140
  "template_name" : "Mistral",
141
  "is_ready": true,
142
+ "is_internal": false,
143
+ "use_metal": true,
144
+ "mlock": false,
145
+ "mmap": true,
146
+ "repeat_last_n": 64,
147
+ "repeat_penalty": 1.2,
148
+ "add_bos_token": true,
149
+ "add_eos_token": false,
150
+ "parse_special_tokens": true
151
  },
152
  {
153
  "id": "phi-2",
 
169
  "n_batch" : 10,
170
  "template_name" : "PHI",
171
  "is_ready": true,
172
+ "is_internal": false,
173
+ "use_metal": true,
174
+ "mlock": false,
175
+ "mmap": true,
176
+ "repeat_last_n": 64,
177
+ "repeat_penalty": 1.2,
178
+ "add_bos_token": true,
179
+ "add_eos_token": false,
180
+ "parse_special_tokens": true
181
  },
182
  {
183
  "id": "yi-6b",
 
199
  "n_batch" : 10,
200
  "template_name" : "yi",
201
  "is_ready": true,
202
+ "is_internal": false,
203
+ "use_metal": true,
204
+ "mlock": false,
205
+ "mmap": true,
206
+ "repeat_last_n": 64,
207
+ "repeat_penalty": 1.2,
208
+ "add_bos_token": true,
209
+ "add_eos_token": false,
210
+ "parse_special_tokens": true
211
+ },
212
+ {
213
+ "id": "gemma-2b",
214
+ "model_title": "Google Gemma 2B",
215
+ "model_file": "gemma-2b-it-q8_0.gguf",
216
+ "model_url": "https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/gemma-2b-it-q8_0.gguf?download=true",
217
+ "model_info_url": "https://huggingface.co/google/gemma-2b",
218
+ "model_avatar": "logo_google",
219
+ "model_description": "Gemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini models. Developed by Google DeepMind and other teams across Google, Gemma is named after the Latin gemma, meaning 'precious stone.' The Gemma model weights are supported by developer tools that promote innovation, collaboration, and the responsible use of artificial intelligence (AI).",
220
+ "developer": "Google",
221
+ "developer_url": "https://huggingface.co/google",
222
+ "file_size": 2669,
223
+ "context" : 8192,
224
+ "temp" : 0.6,
225
+ "prompt_format" : "<bos><start_of_turn>user\n{{prompt}}<end_of_turn>\n<start_of_turn>model\n",
226
+ "top_k" : 5,
227
+ "top_p" : 0.9,
228
+ "model_inference" : "llama",
229
+ "n_batch" : 10,
230
+ "template_name" : "gemma",
231
+ "is_ready": true,
232
+ "is_internal": false,
233
+ "use_metal": true,
234
+ "mlock": false,
235
+ "mmap": true,
236
+ "repeat_last_n": 64,
237
+ "repeat_penalty": 1.2,
238
+ "add_bos_token": true,
239
+ "add_eos_token": false,
240
+ "parse_special_tokens": true
241
  }
242
+ ]