Llama-3-Elyza-Youko-moe-2x8B / mergekit_moe_config.yml
aixsatoshi's picture
Upload 12 files
b6f184b verified
raw
history blame
No virus
336 Bytes
base_model: rinna/llama-3-youko-8b
gate_mode: random # "hidden", "cheap_embed", "random" の中から選択
dtype: bfloat16
experts_per_token: 2 # 各トークンに対してアクティブなエキスパートの数
experts:
- source_model: aixsatoshi/Llama-3-youko-8b-instruct-chatvector
- source_model: elyza/Llama-3-ELYZA-JP-8B