mwitiderrick commited on
Commit
bd14183
1 Parent(s): d9ab5da

Create LogarithmicEqualizationModifier.yaml

Browse files
LogarithmicEqualizationModifier.yaml ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ test_stage:
2
+ obcq_modifiers:
3
+ LogarithmicEqualizationModifier:
4
+ mappings: [
5
+ [["re:.*q_proj", "re:.*k_proj", "re:.*v_proj"], "re:.*input_layernorm"],
6
+ [["re:.*gate_proj", "re:.*up_proj"], "re:.*post_attention_layernorm"]
7
+ ]
8
+ QuantizationModifier:
9
+ ignore:
10
+ # These operations don't make sense to quantize
11
+ - LlamaRotaryEmbedding
12
+ - LlamaRMSNorm
13
+ - SiLUActivation
14
+ # Skip quantizing the BMMs
15
+ - QuantizableMatMul
16
+ # Skip quantizing the layers with the most sensitive activations
17
+ - model.layers.21.mlp.down_proj
18
+ - model.layers.7.mlp.down_proj
19
+ - model.layers.2.mlp.down_proj
20
+ - model.layers.20.mlp.down_proj
21
+ - model.layers.19.mlp.down_proj
22
+ post_oneshot_calibration: true
23
+ scheme_overrides:
24
+ Embedding:
25
+ input_activations: null
26
+ weights:
27
+ num_bits: 8
28
+ symmetric: false
29
+ SparseGPTModifier:
30
+ sparsity: 0.5
31
+ block_size: 128
32
+ sequential_update: true
33
+ quantize: true
34
+ percdamp: 0.01
35
+ mask_structure: "0:0"
36
+ targets: ["re:model.layers.\\d*$"]