models: - model: beomi/Llama-3-Open-Ko-8B-Instruct-preview parameters: weight: 0.5 # Equal weight to maintain balance between foundational language processing and advanced technical tasks layer_range: [0, 20] # Use foundational and intermediate language processing layers in Korean - model: cognitivecomputations/dolphin-2.9-llama3-8b parameters: weight: 0.5 # Equal weight to complement and balance the capabilities of the Llama model layer_range: [15, 24] # Utilize advanced coding and domain-specific layers merge_method: linear # Balanced combination of layers using weighted average dtype: float16 # Efficient resource usage for computational performance