metadata
license: cc-by-nc-4.0
language:
- en
WinterGoddess-1.4x-70B-L2 IQ2-GGUF
Description
IQ2-GGUF quants of Sao10K/WinterGoddess-1.4x-70B-L2
Unlike regular GGUF quants this uses important matrix similar to Quip# to keep the quant from degrading too much even at 2bpw allowing you to run larger models on less powerful machines.
NOTE: Currently you will need experimental branches of Koboldcpp or Ooba for this to work.
- Nexesenex have compiled Windows binaries HERE
- llamacpp_0.2.29 branch of Ooba also works
Models
Regular GGUF Quants: Here
Prompt Format
Alpaca:
### Instruction:
<Prompt>
### Response:
OR
### Instruction:
<Prompt>
### Input:
<Insert Context Here>
### Response:
Contact
Kooten on discord