merge

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using suayptalha/Lamarckvergence-14B as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

name: OptimalMerge-14B
merge_method: dare_ties
base_model: suayptalha/Lamarckvergence-14B
tokenizer_source: base
dtype: bfloat16
out_dtype: bfloat16

parameters:
  normalize: true
  int8_mask: true

models:
  - model: suayptalha/Lamarckvergence-14B
    parameters:
      weight: 0.35
      density: 0.75
  - model: sthenno/tempesthenno-ppo-ckpt40
    parameters:
      weight: 0.25
      density: 0.7
  - model: tanliboy/lambda-qwen2.5-14b-dpo-test
    parameters:
      weight: 0.2
      density: 0.65
  - model: djuna/Q2.5-Veltha-14B
    parameters:
      weight: 0.1
      density: 0.6
  - model: Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4
    parameters:
      weight: 0.1
      density: 0.6

adaptive_merge_parameters:
  task_weights:
    IFEval: 1.8
    BBH: 1.5
    MATH: 2.0
    GPQA: 1.4
    MUSR: 1.3
    MMLU-PRO: 1.5
  smoothing_factor: 0.1

gradient_clipping:
  suayptalha/Lamarckvergence-14B: 0.85
  sthenno/tempesthenno-ppo-ckpt40: 0.88
  tanliboy/lambda-qwen2.5-14b-dpo-test: 0.87
  djuna/Q2.5-Veltha-14B: 0.89
  Goekdeniz-Guelmez/Josiefied-Qwen2.5-14B-Instruct-abliterated-v4: 0.86
Downloads last month
9
Safetensors
Model size
14.8B params
Tensor type
BF16
ยท
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for CultriX/Qwen2.5-14B-DeepResearch

Space using CultriX/Qwen2.5-14B-DeepResearch 1