DaringLotus-10.7B 8bpw EXL2
Description
EXL2 quant of BlueNipples/DaringLotus-10.7B
- 6bpw should be comfortable on 12 gb with 8k context
- 4bpw might just fit on 8gb of vram at 4k context
- if you have more ram get the 8bpw
Other quants:
Prompt Format
Alpaca:
I am not entirely certain of this but i think alpaca is correct for this model
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Input:
{input}
### Response:
Contact
Kooten on discord
- Downloads last month
- 6
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.