Just a merging step for https://huggingface.co/Nohobby/MS-Schisandra-22B-v0.3

Not sure if it can be used as a standalone model. Posting it here for archival purposes.

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: Mistral-Small-22B-ArliAI-RPMax-v1.1
    parameters:
      weight: [0.2, 0.3, 0.2, 0.3, 0.2]
      density: [0.45, 0.55, 0.45, 0.55, 0.45]
  - model: Mistral-Small-NovusKyver
    parameters:
      weight: [0.01768, -0.01675, 0.01285, -0.01696, 0.01421]
      density: [0.6, 0.4, 0.5, 0.4, 0.6]
  - model: MiS-Firefly-v0.2-22B
    parameters:
      weight: [0.208, 0.139, 0.139, 0.139, 0.208]
      density: [0.7]
  - model: magnum-v4-22b
    parameters:
      weight: [0.33]
      density: [0.45, 0.55, 0.45, 0.55, 0.45]
merge_method: della_linear
base_model: Mistral-Small-22B-ArliAI-RPMax-v1.1
parameters:
  epsilon: 0.05
  lambda: 1.05
  int8_mask: true
  rescale: true
  normalize: false
dtype: bfloat16
tokenizer_source: base
Downloads last month
6
Safetensors
Model size
22.2B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Nohobby/Karasik-22B-v0.3

Quantizations
1 model