|
quant_stage: |
|
quant_modifiers: |
|
SmoothQuantModifier: |
|
smoothing_strength: 0.7 |
|
mappings: |
|
- - ['re:.*q_proj', 're:.*k_proj', 're:.*v_proj'] |
|
- re:.*input_layernorm |
|
- - ['re:.*gate_proj', 're:.*up_proj'] |
|
- re:.*post_attention_layernorm |
|
- - ['re:.*down_proj'] |
|
- re:.*up_proj |
|
GPTQModifier: |
|
sequential_update: true |
|
dampening_frac: 0.01 |
|
ignore: [lm_head] |
|
config_groups: |
|
group_0: |
|
targets: [Linear] |
|
weights: {num_bits: 8, type: int, symmetric: true, strategy: channel, observer: mse} |
|
input_activations: {num_bits: 8, type: int, symmetric: true, strategy: token, dynamic: true, |
|
observer: memoryless} |
|
|