flyingfishinwater
commited on
Upload models.json
Browse files- models.json +100 -14
models.json
CHANGED
@@ -1,12 +1,12 @@
|
|
1 |
[
|
2 |
{
|
3 |
"id": "Internal",
|
4 |
-
"model_title": "
|
5 |
"model_file": "ggml-model-Q8_0.gguf",
|
6 |
"model_url": "https://",
|
7 |
"model_info_url": "https://huggingface.co/princeton-nlp/Sheared-LLaMA-1.3B",
|
8 |
"model_avatar": "ava0",
|
9 |
-
"model_description": "
|
10 |
"developer": "Meta",
|
11 |
"developer_url": "https://ai.meta.com/llama/",
|
12 |
"file_size": 1430,
|
@@ -19,29 +19,45 @@
|
|
19 |
"n_batch" : 10,
|
20 |
"template_name" : "HumanBot",
|
21 |
"is_ready": true,
|
22 |
-
"is_internal": true
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
},
|
24 |
{
|
25 |
"id": "LiteLlama-460M-1T-Q8",
|
26 |
-
"model_title": "LiteLlama
|
27 |
"model_file": "LiteLlama-460M-1T-Q8_0.gguf",
|
28 |
"model_url": "https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/LiteLlama-460M-1T-Q8_0.gguf?download=true",
|
29 |
"model_info_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
|
30 |
"model_avatar": "logo_litellama",
|
31 |
-
"model_description": "
|
32 |
"developer": "Xiaotian Han from Texas A&M University",
|
33 |
"developer_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
|
34 |
"file_size": 493,
|
35 |
-
"context" :
|
36 |
"temp" : 0.6,
|
37 |
-
"prompt_format" : "
|
38 |
"top_k" : 5,
|
39 |
"top_p" : 0.9,
|
40 |
"model_inference" : "llama",
|
41 |
"n_batch" : 10,
|
42 |
"template_name" : "TinyLlama",
|
43 |
"is_ready": true,
|
44 |
-
"is_internal": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
},
|
46 |
{
|
47 |
"id": "tinyllama-1.1B-chat-Q8",
|
@@ -63,7 +79,15 @@
|
|
63 |
"n_batch" : 10,
|
64 |
"template_name" : "TinyLlama",
|
65 |
"is_ready": true,
|
66 |
-
"is_internal": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
67 |
},
|
68 |
{
|
69 |
"id": "mistral-7b-instruct-v0.2-Q8",
|
@@ -85,7 +109,15 @@
|
|
85 |
"n_batch" : 10,
|
86 |
"template_name" : "Mistral",
|
87 |
"is_ready": true,
|
88 |
-
"is_internal": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
89 |
},
|
90 |
{
|
91 |
"id": "openchat-3.5-1210-Q8",
|
@@ -107,7 +139,15 @@
|
|
107 |
"n_batch" : 10,
|
108 |
"template_name" : "Mistral",
|
109 |
"is_ready": true,
|
110 |
-
"is_internal": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
111 |
},
|
112 |
{
|
113 |
"id": "phi-2",
|
@@ -129,7 +169,15 @@
|
|
129 |
"n_batch" : 10,
|
130 |
"template_name" : "PHI",
|
131 |
"is_ready": true,
|
132 |
-
"is_internal": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
133 |
},
|
134 |
{
|
135 |
"id": "yi-6b",
|
@@ -151,6 +199,44 @@
|
|
151 |
"n_batch" : 10,
|
152 |
"template_name" : "yi",
|
153 |
"is_ready": true,
|
154 |
-
"is_internal": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
155 |
}
|
156 |
-
]
|
|
|
1 |
[
|
2 |
{
|
3 |
"id": "Internal",
|
4 |
+
"model_title": "AI Assistant",
|
5 |
"model_file": "ggml-model-Q8_0.gguf",
|
6 |
"model_url": "https://",
|
7 |
"model_info_url": "https://huggingface.co/princeton-nlp/Sheared-LLaMA-1.3B",
|
8 |
"model_avatar": "ava0",
|
9 |
+
"model_description": "It is an AI assistant who can talk with you and help solve simple problems. It's based on a lite LLAMA2 model developed by Meta Inc.",
|
10 |
"developer": "Meta",
|
11 |
"developer_url": "https://ai.meta.com/llama/",
|
12 |
"file_size": 1430,
|
|
|
19 |
"n_batch" : 10,
|
20 |
"template_name" : "HumanBot",
|
21 |
"is_ready": true,
|
22 |
+
"is_internal": true,
|
23 |
+
"use_metal": true,
|
24 |
+
"mlock": false,
|
25 |
+
"mmap": true,
|
26 |
+
"repeat_last_n": 64,
|
27 |
+
"repeat_penalty": 1.2,
|
28 |
+
"add_bos_token": true,
|
29 |
+
"add_eos_token": false,
|
30 |
+
"parse_special_tokens": true
|
31 |
},
|
32 |
{
|
33 |
"id": "LiteLlama-460M-1T-Q8",
|
34 |
+
"model_title": "LiteLlama",
|
35 |
"model_file": "LiteLlama-460M-1T-Q8_0.gguf",
|
36 |
"model_url": "https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/LiteLlama-460M-1T-Q8_0.gguf?download=true",
|
37 |
"model_info_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
|
38 |
"model_avatar": "logo_litellama",
|
39 |
+
"model_description": "It's a very small LLAMA2 model with only 460M parameters trained with 1T tokens. It's best for testing.",
|
40 |
"developer": "Xiaotian Han from Texas A&M University",
|
41 |
"developer_url": "https://huggingface.co/ahxt/LiteLlama-460M-1T",
|
42 |
"file_size": 493,
|
43 |
+
"context" : 1024,
|
44 |
"temp" : 0.6,
|
45 |
+
"prompt_format" : "<human>: {{prompt}}\n<bot>:",
|
46 |
"top_k" : 5,
|
47 |
"top_p" : 0.9,
|
48 |
"model_inference" : "llama",
|
49 |
"n_batch" : 10,
|
50 |
"template_name" : "TinyLlama",
|
51 |
"is_ready": true,
|
52 |
+
"is_internal": false,
|
53 |
+
"use_metal": true,
|
54 |
+
"mlock": false,
|
55 |
+
"mmap": true,
|
56 |
+
"repeat_last_n": 64,
|
57 |
+
"repeat_penalty": 1.2,
|
58 |
+
"add_bos_token": true,
|
59 |
+
"add_eos_token": false,
|
60 |
+
"parse_special_tokens": true
|
61 |
},
|
62 |
{
|
63 |
"id": "tinyllama-1.1B-chat-Q8",
|
|
|
79 |
"n_batch" : 10,
|
80 |
"template_name" : "TinyLlama",
|
81 |
"is_ready": true,
|
82 |
+
"is_internal": false,
|
83 |
+
"use_metal": true,
|
84 |
+
"mlock": false,
|
85 |
+
"mmap": true,
|
86 |
+
"repeat_last_n": 64,
|
87 |
+
"repeat_penalty": 1.2,
|
88 |
+
"add_bos_token": true,
|
89 |
+
"add_eos_token": false,
|
90 |
+
"parse_special_tokens": true
|
91 |
},
|
92 |
{
|
93 |
"id": "mistral-7b-instruct-v0.2-Q8",
|
|
|
109 |
"n_batch" : 10,
|
110 |
"template_name" : "Mistral",
|
111 |
"is_ready": true,
|
112 |
+
"is_internal": false,
|
113 |
+
"use_metal": true,
|
114 |
+
"mlock": false,
|
115 |
+
"mmap": true,
|
116 |
+
"repeat_last_n": 64,
|
117 |
+
"repeat_penalty": 1.2,
|
118 |
+
"add_bos_token": true,
|
119 |
+
"add_eos_token": false,
|
120 |
+
"parse_special_tokens": true
|
121 |
},
|
122 |
{
|
123 |
"id": "openchat-3.5-1210-Q8",
|
|
|
139 |
"n_batch" : 10,
|
140 |
"template_name" : "Mistral",
|
141 |
"is_ready": true,
|
142 |
+
"is_internal": false,
|
143 |
+
"use_metal": true,
|
144 |
+
"mlock": false,
|
145 |
+
"mmap": true,
|
146 |
+
"repeat_last_n": 64,
|
147 |
+
"repeat_penalty": 1.2,
|
148 |
+
"add_bos_token": true,
|
149 |
+
"add_eos_token": false,
|
150 |
+
"parse_special_tokens": true
|
151 |
},
|
152 |
{
|
153 |
"id": "phi-2",
|
|
|
169 |
"n_batch" : 10,
|
170 |
"template_name" : "PHI",
|
171 |
"is_ready": true,
|
172 |
+
"is_internal": false,
|
173 |
+
"use_metal": true,
|
174 |
+
"mlock": false,
|
175 |
+
"mmap": true,
|
176 |
+
"repeat_last_n": 64,
|
177 |
+
"repeat_penalty": 1.2,
|
178 |
+
"add_bos_token": true,
|
179 |
+
"add_eos_token": false,
|
180 |
+
"parse_special_tokens": true
|
181 |
},
|
182 |
{
|
183 |
"id": "yi-6b",
|
|
|
199 |
"n_batch" : 10,
|
200 |
"template_name" : "yi",
|
201 |
"is_ready": true,
|
202 |
+
"is_internal": false,
|
203 |
+
"use_metal": true,
|
204 |
+
"mlock": false,
|
205 |
+
"mmap": true,
|
206 |
+
"repeat_last_n": 64,
|
207 |
+
"repeat_penalty": 1.2,
|
208 |
+
"add_bos_token": true,
|
209 |
+
"add_eos_token": false,
|
210 |
+
"parse_special_tokens": true
|
211 |
+
},
|
212 |
+
{
|
213 |
+
"id": "gemma-2b",
|
214 |
+
"model_title": "Google Gemma 2B",
|
215 |
+
"model_file": "gemma-2b-it-q8_0.gguf",
|
216 |
+
"model_url": "https://huggingface.co/flyingfishinwater/goodmodels/resolve/main/gemma-2b-it-q8_0.gguf?download=true",
|
217 |
+
"model_info_url": "https://huggingface.co/google/gemma-2b",
|
218 |
+
"model_avatar": "logo_google",
|
219 |
+
"model_description": "Gemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini models. Developed by Google DeepMind and other teams across Google, Gemma is named after the Latin gemma, meaning 'precious stone.' The Gemma model weights are supported by developer tools that promote innovation, collaboration, and the responsible use of artificial intelligence (AI).",
|
220 |
+
"developer": "Google",
|
221 |
+
"developer_url": "https://huggingface.co/google",
|
222 |
+
"file_size": 2669,
|
223 |
+
"context" : 8192,
|
224 |
+
"temp" : 0.6,
|
225 |
+
"prompt_format" : "<bos><start_of_turn>user\n{{prompt}}<end_of_turn>\n<start_of_turn>model\n",
|
226 |
+
"top_k" : 5,
|
227 |
+
"top_p" : 0.9,
|
228 |
+
"model_inference" : "llama",
|
229 |
+
"n_batch" : 10,
|
230 |
+
"template_name" : "gemma",
|
231 |
+
"is_ready": true,
|
232 |
+
"is_internal": false,
|
233 |
+
"use_metal": true,
|
234 |
+
"mlock": false,
|
235 |
+
"mmap": true,
|
236 |
+
"repeat_last_n": 64,
|
237 |
+
"repeat_penalty": 1.2,
|
238 |
+
"add_bos_token": true,
|
239 |
+
"add_eos_token": false,
|
240 |
+
"parse_special_tokens": true
|
241 |
}
|
242 |
+
]
|