File size: 643 Bytes
37aaa0a 81cdab0 37aaa0a dfa5c58 37aaa0a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
---
license: cc-by-nc-4.0
language:
- en
---
GGUF: https://huggingface.co/Sao10K/Franziska-Maxtral-8x22B-v1-GGUF
An experiment. Maxtral model lora-finetune of my own + merged with i think it was Tess at a low weight, slerp or something together.
the raw adapter was slopped despite trained on base so thats why I added Tess, it helped somewhat.
So it is a merge yes, at the same time part of the model is mine.
kinda meh with it it, but leaving it out here.
loves to yap, slight positivity bias and gpt-isms. kinda expected. its not special or unique, just another model out there.
use alpaca or vicuna or \[INST] blocks or whatever idc. |