sometimesanotion's picture
Update README.md
37ab457 verified
metadata
language:
  - en
license: apache-2.0
library_name: transformers
tags:
  - mergekit
  - merge
base_model:
  - arcee-ai/Virtuoso-Small
  - Qwen2.5-14B-Qwenvergence-model_stock
  - sometimesanotion/Qwen2.5-14B-Qwenvergence-model_stock
metrics:
  - accuracy
pipeline_tag: text-generation

Lamarck.webp

Lamarck 14B v0.4 Qwenvergence: it's a big step up for Lamarck in terms of quality. All the same ingredients are involved as in previous releases of Lamarck; they are more effectively combined. This model features slightly improved reasoning from 0.3, but the multi-language and prose are greatly improved.

Merge Details

This model was initialized from model_stock, and refined from there. No fine-tuning, or use of models apart from those listed as the contents of Qwen2.5-14B-Qwenvergence-model_stock except for a very mild application of huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2.

Models Merged

Top influences: These ancestors are in the Qwenvergence model_stock, reinforced in later steps:

Prose added:

The prose quality has taken a leap, no doubt also to the way EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2, sthenno-com/miscii-14b-1028, oxyapi/oxy-1-small, and underwoods/medius-erebus-magnum-14b were applied.

Configuration

The following YAML configurations were used to initialize and finalize this model:

name:                Qwenvergence-model_stock
merge_method:        model_stock
base_model:          Qwen/Qwen2.5-14B
tokenizer_source:    base
parameters:
  int8_mask:         true
  normalize:         true
  rescale:           false
models:
  - model:           allura-org/TQ2.5-14B-Sugarquill-v1
  - model:           oxyapi/oxy-1-small
  - model:           sthenno-com/miscii-14b-1028
  - model:           underwoods/medius-erebus-magnum-14b
  - model:           EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
  - model:           CultriX/SeQwence-14B-EvolMerge
  - model:           arcee-ai/Virtuoso-Small
  - model:           VAGOsolutions/SauerkrautLM-v2-14b-DPO
  - model:           v000000/Qwen2.5-Lumen-14B
dtype:               bfloat16
out_dtype:           bfloat16
---
# Experimental merge methods involving above models
---
name:                Lamarck-14B-v0.4-Qwenvergence
merge_method:        ties
base_model:          sometimesanotion/lamarck-14b-base
tokenizer_source:    base
parameters:         
  density:           1.00
  weight:            1.00
  int8_mask:         true
  normalize:         true
  rescale:           false
models:
  - model:           merges/Qwen2.5-14B-Qwenvergence-slerp
    parameters:
      weight:        1.00
      density:       1.00
  - model:           arcee-ai/Virtuoso-Small
    parameters:
      weight:        1.00
      density:       1.00