Spaces:
Running
on
A100
Running
on
A100
# Copyright (c) 2025 NVIDIA CORPORATION. | |
# Licensed under the MIT license. | |
# Adapted from https://github.com/NVlabs/VILA/tree/main under the Apache 2.0 license. | |
# LICENSE is in incl_licenses directory. | |
# Activation | |
# Utils | |
from ._dequantize import fp8_dequantize | |
from ._division import fp8_division | |
from ._division_transpose import fp8_division_transpose | |
from ._quantize import fp8_quantize | |
from ._quantize_pertensor import fp8_quantize_pertensor | |
from ._quantize_pertensor_transpose import fp8_quantize_pertensor_transpose | |
from ._transpose import fp8_transpose | |
from .add_bwd import fp8_add_Ifp_Ifp_Ofp_Opt | |
from .add_fwd import fp8_add_Ifp_Ifp_Ofp_Og16 | |
# Normalization | |
from .func_layernorm_noparam import fp8_layernorm_noparam_backward, fp8_layernorm_noparam_forward | |
from .func_quantize import Coat_quantize_bgn, Coat_quantize_end | |
from .func_rmsnorm import fp8_rmsnorm_backward, fp8_rmsnorm_forward | |
from .gelu_bwd import fp8_gelu_backward | |
from .gelu_fwd import fp8_gelu_forward | |
# linear and add | |
from .linear import fp8_linear_backward, fp8_linear_forward | |
from .mul_bwd import fp8_mul_backward | |
from .mul_fwd import fp8_mul_forward | |
from .silu_bwd import fp8_silu_backward | |
from .silu_fwd import fp8_silu_forward | |