| default_stage: | |
| default_modifiers: | |
| AWQModifier: | |
| config_groups: | |
| group_0: | |
| targets: ['re:.*gate_proj.*', 're:.*up_proj.*', 're:.*down_proj.*', 're:.*k_proj.*', | |
| 're:.*q_proj.*', 're:.*v_proj.*', 're:.*o_proj.*'] | |
| weights: | |
| num_bits: 4 | |
| type: int | |
| symmetric: true | |
| group_size: 128 | |
| strategy: group | |
| block_structure: null | |
| dynamic: false | |
| actorder: null | |
| observer: minmax | |
| observer_kwargs: {} | |
| input_activations: null | |
| output_activations: null | |
| format: null | |
| targets: ['re:.*gate_proj.*', 're:.*up_proj.*', 're:.*down_proj.*', 're:.*k_proj.*', | |
| 're:.*q_proj.*', 're:.*v_proj.*', 're:.*o_proj.*'] | |
| ignore: [lm_head, model.embed_tokens, 're:.*input_layernorm$', 're:.*post_attention_layernorm$', | |
| model.norm, 're:.*q_norm$', 're:.*k_norm$', 're:.*shared_experts.*', 're:.*mlp\.gate\.weight$', | |
| 're:.*mlp\.gate\..*bias$', 're:model.layers.[0-2]\.'] | |
| mappings: | |
| - smooth_layer: re:.*input_layernorm$ | |
| balance_layers: ['re:.*q_proj$', 're:.*k_proj$', 're:.*v_proj$'] | |
| - smooth_layer: re:.*v_proj$ | |
| balance_layers: ['re:.*o_proj$'] | |
| - smooth_layer: re:.*post_attention_layernorm$ | |
| balance_layers: ['re:.*gate_proj$', 're:.*up_proj$'] | |
| - smooth_layer: re:.*up_proj$ | |
| balance_layers: ['re:.*down_proj$'] | |
| duo_scaling: true | |