flyingfishinwater
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,3 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
---
|
4 |
# Llama3 8B
|
5 |
|
6 |
Llama 3 is the latest and most advanced LLM trained over 15T tokens, which improves its comprehension and handling of complex language nuances. It features an extended context window of 8k tokens allowing the model to access more information from lengthy passages for more informed decision-making.
|
@@ -178,6 +175,7 @@ OpenChat is an innovative library of open-source language models, fine-tuned wit
|
|
178 |
**Prompt Format:**
|
179 |
|
180 |
```
|
|
|
181 |
GPT4 Correct User: {{prompt}}<|end_of_turn|>GPT4 Correct Assistant:
|
182 |
```
|
183 |
|
@@ -395,15 +393,15 @@ Chinese Tiny LLM 2B 是首个以中文为中心的大型语言模型,主要在
|
|
395 |
|
396 |
---
|
397 |
|
398 |
-
#
|
399 |
|
400 |
Qwen is the large language model and large multimodal model series of the Qwen Team, Alibaba Group. It supports both Chinese and English. 通义千问是阿里巴巴公司开发的大大预言模型,支持中英文双语。
|
401 |
|
402 |
-
**Model Intention:**
|
403 |
|
404 |
-
**Model URL:** [https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/
|
405 |
|
406 |
-
**Model Info URL:** [https://huggingface.co/Qwen/
|
407 |
|
408 |
**Model License:** [License Info](https://huggingface.co/Qwen/Qwen1.5-4B-Chat/raw/main/LICENSE)
|
409 |
|
@@ -411,16 +409,59 @@ Qwen is the large language model and large multimodal model series of the Qwen T
|
|
411 |
|
412 |
**Developer:** [https://qwenlm.github.io/](https://qwenlm.github.io/)
|
413 |
|
414 |
-
**File Size:**
|
415 |
|
416 |
-
**Context Length:**
|
417 |
|
418 |
**Prompt Format:**
|
419 |
|
420 |
```
|
421 |
-
<|im_start|>
|
422 |
-
{{
|
423 |
-
<|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
424 |
<|im_start|>assistant
|
425 |
|
426 |
```
|
@@ -436,15 +477,15 @@ Qwen is the large language model and large multimodal model series of the Qwen T
|
|
436 |
|
437 |
---
|
438 |
|
439 |
-
# Dophin 2.
|
440 |
|
441 |
This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensored model and supports a variety of instruction, conversational, and coding skills.
|
442 |
|
443 |
**Model Intention:** It's a uncensored and good skilled English modal best for high performance iPhone, iPad & Mac
|
444 |
|
445 |
-
**Model URL:** [https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/dolphin-2.
|
446 |
|
447 |
-
**Model Info URL:** [https://huggingface.co/cognitivecomputations/dolphin-2.
|
448 |
|
449 |
**Model License:** [License Info](https://www.apache.org/licenses/LICENSE-2.0)
|
450 |
|
@@ -452,16 +493,17 @@ This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensor
|
|
452 |
|
453 |
**Developer:** [https://erichartford.com/](https://erichartford.com/)
|
454 |
|
455 |
-
**File Size:**
|
456 |
|
457 |
-
**Context Length:**
|
458 |
|
459 |
**Prompt Format:**
|
460 |
|
461 |
```
|
462 |
-
|
463 |
-
{{
|
464 |
-
<|
|
|
|
465 |
<|im_start|>assistant
|
466 |
|
467 |
```
|
@@ -511,4 +553,4 @@ ASSISTANT:
|
|
511 |
|
512 |
**Add EOS Token:** No
|
513 |
|
514 |
-
**Parse Special Tokens:** Yes
|
|
|
|
|
|
|
|
|
1 |
# Llama3 8B
|
2 |
|
3 |
Llama 3 is the latest and most advanced LLM trained over 15T tokens, which improves its comprehension and handling of complex language nuances. It features an extended context window of 8k tokens allowing the model to access more information from lengthy passages for more informed decision-making.
|
|
|
175 |
**Prompt Format:**
|
176 |
|
177 |
```
|
178 |
+
{{system}}
|
179 |
GPT4 Correct User: {{prompt}}<|end_of_turn|>GPT4 Correct Assistant:
|
180 |
```
|
181 |
|
|
|
393 |
|
394 |
---
|
395 |
|
396 |
+
# Qwen2 7B Chat
|
397 |
|
398 |
Qwen is the large language model and large multimodal model series of the Qwen Team, Alibaba Group. It supports both Chinese and English. 通义千问是阿里巴巴公司开发的大大预言模型,支持中英文双语。
|
399 |
|
400 |
+
**Model Intention:** Qwen2 is the new series that generally surpassed most opensource models and demonstrated competitiveness against proprietary models across a series of benchmarks targeting for language understanding, language generation, multilingual capability, coding, mathematics, reasoning, etc
|
401 |
|
402 |
+
**Model URL:** [https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/Qwen2-7B-Instruct-Q3_K_S.gguf?download=true](https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/Qwen2-7B-Instruct-Q3_K_S.gguf?download=true)
|
403 |
|
404 |
+
**Model Info URL:** [https://huggingface.co/Qwen/Qwen2-7B-Instruct-GGUF](https://huggingface.co/Qwen/Qwen2-7B-Instruct-GGUF)
|
405 |
|
406 |
**Model License:** [License Info](https://huggingface.co/Qwen/Qwen1.5-4B-Chat/raw/main/LICENSE)
|
407 |
|
|
|
409 |
|
410 |
**Developer:** [https://qwenlm.github.io/](https://qwenlm.github.io/)
|
411 |
|
412 |
+
**File Size:** 3490 MB
|
413 |
|
414 |
+
**Context Length:** 2048 tokens
|
415 |
|
416 |
**Prompt Format:**
|
417 |
|
418 |
```
|
419 |
+
<|im_start|>system
|
420 |
+
{{system_prompt}}<|im_end|>
|
421 |
+
<|im_start|>
|
422 |
+
{{prompt}}<|im_end|>
|
423 |
+
<|im_start|>assistant
|
424 |
+
|
425 |
+
```
|
426 |
+
|
427 |
+
**Template Name:** chatml
|
428 |
+
|
429 |
+
**Add BOS Token:** Yes
|
430 |
+
|
431 |
+
**Add EOS Token:** No
|
432 |
+
|
433 |
+
**Parse Special Tokens:** Yes
|
434 |
+
|
435 |
+
|
436 |
+
---
|
437 |
+
|
438 |
+
# Qwen2 1.5B Chat
|
439 |
+
|
440 |
+
Qwen is the large language model and large multimodal model series of the Qwen Team, Alibaba Group. It supports both Chinese and English. 通义千问是阿里巴巴公司开发的大大预言模型,支持中英文双语。
|
441 |
+
|
442 |
+
**Model Intention:** Qwen2 is the new series that generally surpassed most opensource models and demonstrated competitiveness against proprietary models across a series of benchmarks targeting for language understanding, language generation, multilingual capability, coding, mathematics, reasoning, etc
|
443 |
+
|
444 |
+
**Model URL:** [https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/Qwen2-1.5B-Instruct.Q4_K_M.gguf?download=true](https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/Qwen2-1.5B-Instruct.Q4_K_M.gguf?download=true)
|
445 |
+
|
446 |
+
**Model Info URL:** [https://huggingface.co/Qwen/Qwen2-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct)
|
447 |
+
|
448 |
+
**Model License:** [License Info](https://huggingface.co/Qwen/Qwen1.5-4B-Chat/raw/main/LICENSE)
|
449 |
+
|
450 |
+
**Model Description:** Qwen is the large language model and large multimodal model series of the Qwen Team, Alibaba Group. It supports both Chinese and English. 通义千问是阿里巴巴公司开发的大大预言模型,支持中英文双语。
|
451 |
+
|
452 |
+
**Developer:** [https://qwenlm.github.io/](https://qwenlm.github.io/)
|
453 |
+
|
454 |
+
**File Size:** 986 MB
|
455 |
+
|
456 |
+
**Context Length:** 2048 tokens
|
457 |
+
|
458 |
+
**Prompt Format:**
|
459 |
+
|
460 |
+
```
|
461 |
+
<|im_start|>system
|
462 |
+
{{system_prompt}}<|im_end|>
|
463 |
+
<|im_start|>
|
464 |
+
{{prompt}}<|im_end|>
|
465 |
<|im_start|>assistant
|
466 |
|
467 |
```
|
|
|
477 |
|
478 |
---
|
479 |
|
480 |
+
# Dophin 2.9.2 Qwen2 7B
|
481 |
|
482 |
This model is based on Mistral-7b-v0.2 with 16k context lengths. It's a uncensored model and supports a variety of instruction, conversational, and coding skills.
|
483 |
|
484 |
**Model Intention:** It's a uncensored and good skilled English modal best for high performance iPhone, iPad & Mac
|
485 |
|
486 |
+
**Model URL:** [https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/dolphin-2.9.2-qwen2-7b-Q3_K_S.gguf?download=true](https://huggingface.co/flyingfishinwater/good_and_small_models/resolve/main/dolphin-2.9.2-qwen2-7b-Q3_K_S.gguf?download=true)
|
487 |
|
488 |
+
**Model Info URL:** [https://huggingface.co/cognitivecomputations/dolphin-2.9.2-qwen2-7b-gguf](https://huggingface.co/cognitivecomputations/dolphin-2.9.2-qwen2-7b-gguf)
|
489 |
|
490 |
**Model License:** [License Info](https://www.apache.org/licenses/LICENSE-2.0)
|
491 |
|
|
|
493 |
|
494 |
**Developer:** [https://erichartford.com/](https://erichartford.com/)
|
495 |
|
496 |
+
**File Size:** 3490 MB
|
497 |
|
498 |
+
**Context Length:** 2048 tokens
|
499 |
|
500 |
**Prompt Format:**
|
501 |
|
502 |
```
|
503 |
+
<|im_start|>system
|
504 |
+
{{system_prompt}}<|im_end|>
|
505 |
+
<|im_start|>user
|
506 |
+
{{prompt}}<|im_end|>
|
507 |
<|im_start|>assistant
|
508 |
|
509 |
```
|
|
|
553 |
|
554 |
**Add EOS Token:** No
|
555 |
|
556 |
+
**Parse Special Tokens:** Yes
|