models: | |
- model: beomi/Llama-3-Open-Ko-8B-Instruct-preview | |
parameters: | |
weight: 0.5 # Equal weight to maintain balance between foundational language processing and advanced technical tasks | |
layer_range: [0, 20] # Use foundational and intermediate language processing layers in Korean | |
- model: cognitivecomputations/dolphin-2.9-llama3-8b | |
parameters: | |
weight: 0.5 # Equal weight to complement and balance the capabilities of the Llama model | |
layer_range: [15, 24] # Utilize advanced coding and domain-specific layers | |
merge_method: linear # Balanced combination of layers using weighted average | |
dtype: float16 # Efficient resource usage for computational performance | |