arch: type: TransformerLMHeadModel args: transformer_config: type: TransformerDecoderOnlyModel args: embed_config: type: TransformerEmbeddingBlock args: token_embed_config: type: TokenEmbedding args: n_embed: 2048 n_vocab: 32256 pos_embed_config: null type_embed_config: null ln_config: null p_drop_embed: 0.0 concat_strategy: id_first decoder_config: type: TransformerDecoderBlock args: attn_config: type: LlamaAttention args: n_embed: 2048 n_pos: 16384 n_head: 16 n_key_value_head: 16 head_size: 128 p_drop_attn: 0.0 p_drop_resid: 0.0 bias_attn: false bias_proj: false cross_attn: false scale_dot_product: true scale_layer_wise: false layer_idx: null rope_config: type: RotaryPositionEmbedding args: head_size: 128 n_pos: 16384 base: 100000 scaling_type: linear scaling_factor: 4.0 mlp_config: type: LlamaMLP args: n_embed: 2048 n_inner: 5504 act_fn_config: type: SiLUActivation args: {} ln_config: type: LlamaRMSNorm args: n_embed: 2048 ln_eps: 1.0e-06 n_embed: 2048 post_norm: false add_cross_attn: false n_embed: 2048 n_layer: 24 n_head: 16 ln_config: type: LlamaRMSNorm args: n_embed: 2048 ln_eps: 1.0e-06 perform_linear_bias: false attn_window_size_loop_unit: null lm_head_config: type: TransformerLMHead args: n_vocab: 32256 n_embed: 2048 perform_transform: false act_fn_config: null ln_config: null