SreyanG-NVIDIA's picture
Upload 225 files
174ae06 verified
# Copyright (c) 2025 NVIDIA CORPORATION.
# Licensed under the MIT license.
# Adapted from https://github.com/NVlabs/VILA/tree/main under the Apache 2.0 license.
# LICENSE is in incl_licenses directory.
class QuantizationConfig:
def __init__(self):
self.qlinear_config = {
"mlp_gate": {"all", "linear"},
"mlp_up": {"all", "linear"},
"mlp_down": {"all", "linear"},
"attn_proj": {"all", "linear"},
"attn_q": {"all", "linear"},
"attn_k": {"all", "linear"},
"attn_v": {"all", "linear"},
}
self.qact_config = {
"mul_act_in1": {"all", "gelu"},
"mul_act_in2": {"all", "gelu", "te_like"},
"mul_act_out": {"all", "gelu", "te_like"},
"mlp_act_sum": {"all", "mlp", "te_like"},
"mlp_act_gate": {"all", "mlp", "te_like"},
"mlp_act_up": {"all", "mlp", "te_like"},
"mlp_act_in": {"all", "mlp", "te_like"},
"mlp_act_out": {"all", "mlp"},
"ln_attn_in": {"all", "layernorm"},
"ln_mlp_in": {"all", "layernorm"},
"ln_attn_out": {"all", "layernorm", "te_like"},
"ln_mlp_out": {"all", "layernorm", "te_like"},
"add_attn_in_re": {"all", "residual"},
"add_attn_in_fx": {"all", "residual", "te_like"},
"add_mlp_in_re": {"all", "residual"},
"add_mlp_in_fx": {"all", "residual", "te_like"},
"re_attn_out_re": {"all", "residual"},
"re_attn_out_fx": {"all", "residual"},
"re_mlp_out_re": {"all", "residual"},
"re_mlp_out_fx": {"all", "residual"},
"attn_qkv_sum": {"all", "attn", "te_like"},
"attn_q_in": {"all", "attn", "te_like"},
"attn_k_in": {"all", "attn", "te_like"},
"attn_v_in": {"all", "attn", "te_like"},
"attn_q_out": {"all", "attn", "te_like"},
"attn_k_out": {"all", "attn", "te_like"},
"attn_v_out": {"all", "attn", "te_like"},
"attn_proj_in": {"all", "attn", "te_like"},
}
self.qgelu_config = {"mlp_gelu": {"all", "gelu"}}
self.qlayernorm_config = {"ln_attn": {"all", "layernorm"}, "ln_mlp": {"all", "layernorm"}}
self.qadd_config = {"add_attn": {"all", "residual"}, "add_mlp": {"all", "residual"}}
self.qmul_config = {
"mul_act": {"all", "gelu"},
}
qconfig = QuantizationConfig()