peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/deepspeed
/linear
/config.py
| # Copyright (c) Microsoft Corporation. | |
| # SPDX-License-Identifier: Apache-2.0 | |
| # DeepSpeed Team | |
| from dataclasses import dataclass | |
| class LoRAConfig: | |
| """ | |
| Configuration settings for LoRAOptimizedLinear. | |
| Attributes: | |
| lora_r (int): LoRA attention dimension, also know as the rank. Defaults is 64. | |
| lora_alpha (float): LoRA scaling factor, default is 16. | |
| base_weight_sharding (int): The degree to which the base weights are sharded, | |
| should typically be set to the data-parallel world size to maximize the memory | |
| reduction benefits. Defaults to 1, which means this feature is disabled. | |
| """ | |
| lora_r: int = 64 | |
| lora_alpha: float = 16. | |
| base_weight_sharding: int = 1 | |
| class QuantizationConfig: | |
| """ | |
| Configuration settings for quantization for LoRAOptimizedLinear, QuantizedLinear, | |
| and QuantizedParameter | |
| Attributes: | |
| q_bits (int): The number of bits used for quantization. Default is 8. | |
| mantissa_bits (int): The number of bits reserved for the mantissa in fixed-point quantization. Default is 3. | |
| group_size (int): The size of the group used for quantization. Default is 512. | |
| """ | |
| q_bits: int = 8 | |
| mantissa_bits: int = 3 | |
| group_size: int = 512 | |