Edit model card

Tyr1

fp16: https://huggingface.co/Sao10K/Typhon-Mixtral-v1

gguf: https://huggingface.co/Sao10K/Typhon-Mixtral-v1-GGUF

3.5 bpw

4.5 bpw

6 bpw

I used the default calibration set to quant these.


Typhon - A Custom Experimental Mixtral Merge

An experimental Merge I tried for fun. Honestly did not expect it to work for Mixtral at all considering how its an MoE and the gates and all would be fucked by this custom merge.

From my testing it was able to handle SFW <--> NSFW scenarios fine, handle 1st and 3rd person roleplays fine, and seemed fairly smart.

It did pretty well for non NSFW tasks so that's a win.


Recipe Below:

base_model: mistralai/Mixtral-8x7B-v0.1
models:
  - model: mistralai/Mixtral-8x7B-v0.1
    # no parameters necessary for base model
  - model: smelborp/MixtralOrochi8x7B
    parameters:
      weight: 0.30
      density: 0.47
  - model:  notstoic/Nous-Hermes-2-Mixtruct-v0.1-8x7B-DPO-DARE_TIES
    parameters:
      weight: 0.31
      density: 0.56
  - model: Sao10K/Solstice-Mixtral-v1
    parameters: 
      weight: 0.36
      density: 0.64
  - model: Sao10K/Frostwind-Mixtral-v1
    parameters:
      weight: 0.22
      density: 0.44
  - model: KoboldAI/Mixtral-8x7B-Holodeck-v1
    parameters:
      weight: 0.21
      density: 0.36
merge_method: dare_ties
dtype: bfloat16
Downloads last month

-

Downloads are not tracked for this model. How to track
Unable to determine this model's library. Check the docs .

Finetuned from