quant_stage: quant_modifiers: QuantizationModifier: ignore: [lm_head] kv_cache_scheme: {num_bits: 8, type: float, strategy: tensor, dynamic: false, symmetric: true} GPTQModifier: sequential_update: false ignore: [lm_head] config_groups: group_0: weights: {num_bits: 4, type: int, symmetric: true, strategy: channel, actorder: false} targets: [Linear]