KoDolph-2x8b-Instruct / mergekit_config.yml
asiansoul's picture
Upload folder using huggingface_hub
e75ded6 verified
raw
history blame
701 Bytes
models:
- model: beomi/Llama-3-Open-Ko-8B-Instruct-preview
parameters:
weight: 0.5 # Equal weight to maintain balance between foundational language processing and advanced technical tasks
layer_range: [0, 20] # Use foundational and intermediate language processing layers in Korean
- model: cognitivecomputations/dolphin-2.9-llama3-8b
parameters:
weight: 0.5 # Equal weight to complement and balance the capabilities of the Llama model
layer_range: [15, 24] # Utilize advanced coding and domain-specific layers
merge_method: linear # Balanced combination of layers using weighted average
dtype: float16 # Efficient resource usage for computational performance