test_stage_0: | |
obcq_modifiers: | |
LogarithmicEqualizationModifier: | |
mappings: | |
- - - re:.*q_proj | |
- re:.*k_proj | |
- re:.*v_proj | |
- re:.*input_layernorm | |
- - - re:.*gate_proj | |
- re:.*up_proj | |
- re:.*post_attention_layernorm | |
QuantizationModifier: | |
ignore: | |
- LlamaRotaryEmbedding | |
- LlamaRMSNorm | |
- SiLUActivation | |
- MatMulOutput_QK | |
- MatMulOutput_PV | |
- model.layers.21.mlp.down_proj | |
- model.layers.7.mlp.down_proj | |
- model.layers.2.mlp.down_proj | |
- model.layers.8.self_attn.q_proj | |
- model.layers.8.self_attn.k_proj | |
post_oneshot_calibration: true | |
scheme_overrides: | |
Linear: | |
weights: | |
num_bits: 8 | |
symmetric: true | |
strategy: channel | |
MatMulLeftInput_QK: | |
input_activations: | |
num_bits: 8 | |
symmetric: true | |
MatMulLeftInput_PV: | |
input_activations: | |
num_bits: 8 | |
symmetric: true | |
Embedding: | |
input_activations: null | |
weights: | |
num_bits: 8 | |
symmetric: false | |
SparseGPTModifier: | |
sparsity: 0.5 | |
block_size: 128 | |
sequential_update: true | |
quantize: true | |
percdamp: 0.01 | |
mask_structure: 0:0 | |
targets: | |
- re:model.layers.\d*$ | |
test_stage_1: | |
obcq_modifiers: | |
LogarithmicEqualizationModifier: | |
mappings: | |
- - - re:.*q_proj | |
- re:.*k_proj | |
- re:.*v_proj | |
- re:.*input_layernorm | |
- - - re:.*gate_proj | |
- re:.*up_proj | |
- re:.*post_attention_layernorm | |
QuantizationModifier: | |
ignore: | |
- LlamaRotaryEmbedding | |
- LlamaRMSNorm | |
- SiLUActivation | |
- MatMulOutput_QK | |
- MatMulOutput_PV | |
- model.layers.21.mlp.down_proj | |
- model.layers.7.mlp.down_proj | |
- model.layers.2.mlp.down_proj | |
- model.layers.8.self_attn.q_proj | |
- model.layers.8.self_attn.k_proj | |
post_oneshot_calibration: true | |
scheme_overrides: | |
Linear: | |
weights: | |
num_bits: 8 | |
symmetric: true | |
strategy: channel | |
MatMulLeftInput_QK: | |
input_activations: | |
num_bits: 8 | |
symmetric: true | |
MatMulLeftInput_PV: | |
input_activations: | |
num_bits: 8 | |
symmetric: true | |
Embedding: | |
input_activations: null | |
weights: | |
num_bits: 8 | |
symmetric: false | |
SparseGPTModifier: | |
sparsity: 0.5 | |
block_size: 128 | |
sequential_update: true | |
quantize: true | |
percdamp: 0.01 | |
mask_structure: 0:0 | |
targets: | |
- re:model.layers.\d*$ | |