--- tags: - not-for-all-audiences --- # daybreak-kunoichi-2dpo-7b - EXL2 8bpw This is a 8bpw EXL2 quant of [crestf411/daybreak-kunoichi-2dpo-7b](https://huggingface.co/crestf411/daybreak-kunoichi-2dpo-7b) This quant was made using exllamav2-0.0.21 with default dataset. Context limit seems to be 8k (in webui it shows 16k by default but it gets incoherent past 8k, use alpha_value in webui to scale to 16k). I tested this quant shortly in some random RPs (including one over 8k with alpha_value in webui) and it seems to work fine. ## Prompt Templates This model seems to use Alpaca format. ### Original readme below --- Experimental model doing a DPO training on top of [Kunoichi-DPO-v2-7b](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B), i.e. double-DPO. Not suitable for any audience.