license: apache-2.0
base_model:
- Qwen/Qwen2.5-7B
pipeline_tag: text-generation
tags:
- not-for-all-audiences
language:
- en
library_name: transformers
Model Description
Model created by analyzing and selecting the optimal layers from other Qwen2.5-7B models based on their dimensional utilization efficiency, measured by the Normalized Effective Rank (NER). Computed like:
Singular Value Decomposition:
- Input: Weight matrix A ∈ R^(m×n) # m = number of output features, n = number of input features
- Compute singular values σᵢ where σᵢ ≥ 0 # σᵢ represents the importance of each dimension
- Filter values above numerical threshold (>1e-12) # removes numerical noise from computation
Distribution Normalization:
- Sum all singular values: S = Σσᵢ # S acts as normalization factor
- Create probability distribution: pᵢ = σᵢ/S # converts singular values to probabilities summing to 1
Entropy Calculation:
- Compute Shannon entropy: H = -Σ(pᵢ * log₂(pᵢ)) # measures information content of distribution
- Calculate maximum possible entropy: H_max = log₂(n) # n = number of singular values where n is the number of singular values # maximum entropy occurs when all dimensions contribute equally
Normalization:
- Final NER score = H/H_max # normalizes score to [0,1] range
- Results in value between 0 and 1 # 0 = single dimension dominance, 1 = perfect dimensional utilization
- Higher scores indicate more uniform dimensional utilization
Creating Composite Model
Code here: https://huggingface.co/jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0/blob/main/ner_merge.py
Layer Analysis:
- Download base and fine-tuned models from Hugging Face Hub
- Calculate Normalized Effective Rank (NER) for each layer within each model
Layer Selection:
- Identify common layer structures across models
- Define model and layer name pairs that have highest NER for each layer based on their NER scores
Model Composition:
- Incrementally build a composite model using layer with highest NER from model pool.
Output Generation:
- Save merge reports documenting layer sources
- Copy config and tokenizer files from base model
- Save the composite model with complete weights # model ready to use
Configfile:
base_model: "Qwen/Qwen2.5-7B"
fine_tuned_models: # uncomment the models you want to merge
#- "Qwen/Qwen2.5-7B"
#- "Qwen/Qwen2.5-7B-Instruct"
#- "FourOhFour/Vapor_v2_7B"
#- "Goekdeniz-Guelmez/Josiefied-Qwen2.5-7B-Instruct-abliterated-v2"
#- "happzy2633/qwen2.5-7b-ins-v3"
#- "huihui-ai/Qwen2.5-7B-Instruct-abliterated-v2"
#- "HumanLLMs/Humanish-Qwen2.5-7B-Instruct"
#- "Orion-zhen/Qwen2.5-7B-Instruct-Uncensored"
#- "Orion-zhen/Meissa-Qwen2.5-7B-Instruct"
#- "jeffmeloy/Qwen2.5-7B-nerd-uncensored-v1.0"
#- "rombodawg/Rombos-LLM-V2.5-Qwen-7b"
#- "Cran-May/T.E-8.1"
#- "thomas-yanxin/XinYuan-Qwen2.5-7B-0917"
#- "beomi/Qwen2.5-7B-Instruct-kowiki-qa"
#- "Orion-zhen/Qwen2.5-7B-Gutenberg-KTO"
#- 'fblgit/cybertron-v4-qw7B-MGS'
#- 'nguyentd/FinancialAdvice-Qwen2.5-7B'
#- "Qwen/Qwen2.5-Coder-7B-Instruct"
#- "Qwen/Qwen2.5-Math-7B-Instruct"
#- "Qwen/Qwen2.5-Coder-7B"
#- "Qwen/Qwen2.5-Math-7B"
#- "WhiteRabbitNeo/WhiteRabbitNeo-2.5-Qwen-2.5-Coder-7B"
#- "edgerunner-ai/EdgeRunner-Command-Nested"
#- "katanemo/Arch-Function-7B"
models_dir: "./input_models/"
output_dir: "./merged_model/"
metric_dir: "./metrics/"