Updating model files
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
license: other
|
3 |
library_name: transformers
|
4 |
pipeline_tag: text-generation
|
5 |
-
datasets:
|
6 |
- RyokoAI/ShareGPT52K
|
7 |
- Hello-SimpleAI/HC3
|
8 |
tags:
|
@@ -12,6 +12,17 @@ tags:
|
|
12 |
- gptq
|
13 |
inference: false
|
14 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
# Koala: A Dialogue Model for Academic Research
|
16 |
This repo contains the weights of the Koala 7B model produced at Berkeley. It is the result of combining the diffs from https://huggingface.co/young-geng/koala with the original Llama 7B model.
|
17 |
|
@@ -24,7 +35,7 @@ I have the following Koala model repositories available:
|
|
24 |
* [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
|
25 |
* [GPTQ quantized 4bit 13B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
|
26 |
* [4-bit, 5-bit and 8-bit GGML models for `llama.cpp`](https://huggingface.co/TheBloke/koala-13B-GGML)
|
27 |
-
|
28 |
**7B models:**
|
29 |
* [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
|
30 |
* [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
|
@@ -57,7 +68,7 @@ Details of the files provided:
|
|
57 |
* The older GPTQ code does not support all the latest features, so the quality may be fractionally lower.
|
58 |
* Command to create:
|
59 |
* `python3 llama.py koala-7B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save koala-7B-4bit-128g.no-act-order.ooba.pt`
|
60 |
-
|
61 |
## How to run in `text-generation-webui`
|
62 |
|
63 |
File `koala-7B-4bit-128g.no-act-order.ooba.pt` can be loaded the same as any other GPTQ file, without requiring any updates to [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
@@ -122,6 +133,17 @@ PYTHON_PATH="${PWD}:$PYTHONPATH" python \
|
|
122 |
--tokenizer_path=/content/LLaMA-7B/tokenizer.model
|
123 |
```
|
124 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
125 |
## Further info
|
126 |
|
127 |
Check out the following links to learn more about the Berkeley Koala model.
|
|
|
2 |
license: other
|
3 |
library_name: transformers
|
4 |
pipeline_tag: text-generation
|
5 |
+
datasets:
|
6 |
- RyokoAI/ShareGPT52K
|
7 |
- Hello-SimpleAI/HC3
|
8 |
tags:
|
|
|
12 |
- gptq
|
13 |
inference: false
|
14 |
---
|
15 |
+
<div style="width: 100%;">
|
16 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
17 |
+
</div>
|
18 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
19 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
20 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
21 |
+
</div>
|
22 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
23 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
24 |
+
</div>
|
25 |
+
</div>
|
26 |
# Koala: A Dialogue Model for Academic Research
|
27 |
This repo contains the weights of the Koala 7B model produced at Berkeley. It is the result of combining the diffs from https://huggingface.co/young-geng/koala with the original Llama 7B model.
|
28 |
|
|
|
35 |
* [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
|
36 |
* [GPTQ quantized 4bit 13B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
|
37 |
* [4-bit, 5-bit and 8-bit GGML models for `llama.cpp`](https://huggingface.co/TheBloke/koala-13B-GGML)
|
38 |
+
|
39 |
**7B models:**
|
40 |
* [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
|
41 |
* [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
|
|
|
68 |
* The older GPTQ code does not support all the latest features, so the quality may be fractionally lower.
|
69 |
* Command to create:
|
70 |
* `python3 llama.py koala-7B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save koala-7B-4bit-128g.no-act-order.ooba.pt`
|
71 |
+
|
72 |
## How to run in `text-generation-webui`
|
73 |
|
74 |
File `koala-7B-4bit-128g.no-act-order.ooba.pt` can be loaded the same as any other GPTQ file, without requiring any updates to [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
|
|
|
133 |
--tokenizer_path=/content/LLaMA-7B/tokenizer.model
|
134 |
```
|
135 |
|
136 |
+
## Want to support my work?
|
137 |
+
|
138 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
139 |
+
|
140 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
141 |
+
|
142 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
143 |
+
|
144 |
+
* Patreon: coming soon! (just awaiting approval)
|
145 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
146 |
+
* Discord: https://discord.gg/UBgz4VXf
|
147 |
## Further info
|
148 |
|
149 |
Check out the following links to learn more about the Berkeley Koala model.
|