Update README.md
Browse files
README.md
CHANGED
@@ -19,14 +19,14 @@ license: other
|
|
19 |
|
20 |
# Allen AI's Tulu 30B merged with Kaio Ken's SuperHOT 8K fp16
|
21 |
|
22 |
-
These files are
|
|
|
|
|
23 |
|
24 |
-
It is the result of merging and/or converting the source repository to float16.
|
25 |
|
26 |
## Repositories available
|
27 |
|
28 |
-
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/Tulu-30B-SuperHOT-8K-
|
29 |
-
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/none)
|
30 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/Tulu-30B-SuperHOT-8K-fp16)
|
31 |
|
32 |
<!-- footer start -->
|
|
|
19 |
|
20 |
# Allen AI's Tulu 30B merged with Kaio Ken's SuperHOT 8K fp16
|
21 |
|
22 |
+
These files are fp16 model files for [Allen AI's Tulu 30B](https://huggingface.co/allenai/tulu-30b) merged with [Kaio Ken's SuperHOT 30B 8K LoRA](https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test) to produce a model capable of 8K context.
|
23 |
+
|
24 |
+
[Kaio Ken's SuperHOT 30B LoRA](https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test) is merged on to the base model to produce a model capable of 8K context, via a modified version of the Llama modelling code.
|
25 |
|
|
|
26 |
|
27 |
## Repositories available
|
28 |
|
29 |
+
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/Tulu-30B-SuperHOT-8K-GPTQ)
|
|
|
30 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/Tulu-30B-SuperHOT-8K-fp16)
|
31 |
|
32 |
<!-- footer start -->
|