File size: 643 Bytes
37aaa0a
 
 
 
 
 
 
81cdab0
37aaa0a
dfa5c58
 
37aaa0a
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
---
license: cc-by-nc-4.0
language:
- en
---
GGUF: https://huggingface.co/Sao10K/Franziska-Maxtral-8x22B-v1-GGUF

An experiment. Maxtral model lora-finetune of my own + merged with i think it was Tess at a low weight, slerp or something together.

the raw adapter was slopped despite trained on base so thats why I added Tess, it helped somewhat.

So it is a merge yes, at the same time part of the model is mine.

kinda meh with it it, but leaving it out here.

loves to yap, slight positivity bias and gpt-isms. kinda expected. its not special or unique, just another model out there.

use alpaca or vicuna or \[INST] blocks or whatever idc.