python_code
stringlengths
0
992k
repo_name
stringlengths
8
46
file_path
stringlengths
5
162
import os import re from typing import Union import requests from hivemind.utils.logging import TextStyle, get_logger from packaging.version import parse import grid logger = get_logger(__name__) def validate_version() -> None: logger.info(f"Running {TextStyle.BOLD}Grid {grid.__version__}{TextStyle.RESET}") try: r = requests.get("https://pypi.python.org/pypi/grid/json") r.raise_for_status() response = r.json() versions = [parse(ver) for ver in response.get("releases")] latest = max(ver for ver in versions if not ver.is_prerelease) if parse(grid.__version__) < latest: logger.info( f"A newer version {latest} is available. Please upgrade with: " f"{TextStyle.BOLD}pip install --upgrade grid{TextStyle.RESET}" ) except Exception: logger.warning("Failed to fetch the latest Grid version from PyPI:", exc_info=True) def get_compatible_model_repo(model_name_or_path: Union[str, os.PathLike, None]) -> Union[str, os.PathLike, None]: if model_name_or_path is None: return None match = re.fullmatch(r"(Agora/.+)-grid", str(model_name_or_path)) if match is None: return model_name_or_path logger.info( f"Loading model from {match.group(1)}, since Grid 1.2.0+ uses original repos instead of converted ones" ) return match.group(1)
TheGrid-main
grid/utils/version.py
import fcntl import os import shutil from contextlib import contextmanager from pathlib import Path from typing import Optional import huggingface_hub from hivemind.utils.logging import get_logger logger = get_logger(__name__) DEFAULT_CACHE_DIR = os.getenv("GRID_CACHE", Path(Path.home(), ".cache", "grid")) BLOCKS_LOCK_FILE = "blocks.lock" @contextmanager def _blocks_lock(cache_dir: Optional[str], mode: int): if cache_dir is None: cache_dir = DEFAULT_CACHE_DIR lock_path = Path(cache_dir, BLOCKS_LOCK_FILE) os.makedirs(lock_path.parent, exist_ok=True) with open(lock_path, "wb") as lock_fd: fcntl.flock(lock_fd.fileno(), mode) # The OS will release the lock when lock_fd is closed or the process is killed yield def allow_cache_reads(cache_dir: Optional[str]): """Allows simultaneous reads, guarantees that blocks won't be removed along the way (shared lock)""" return _blocks_lock(cache_dir, fcntl.LOCK_SH) def allow_cache_writes(cache_dir: Optional[str]): """Allows saving new blocks and removing the old ones (exclusive lock)""" return _blocks_lock(cache_dir, fcntl.LOCK_EX) def free_disk_space_for( size: int, *, cache_dir: Optional[str], max_disk_space: Optional[int], os_quota: int = 1024**3, # Minimal space we should leave to keep OS function normally ): if cache_dir is None: cache_dir = DEFAULT_CACHE_DIR cache_info = huggingface_hub.scan_cache_dir(cache_dir) available_space = shutil.disk_usage(cache_dir).free - os_quota if max_disk_space is not None: available_space = min(available_space, max_disk_space - cache_info.size_on_disk) gib = 1024**3 logger.debug(f"Disk space: required {size / gib:.1f} GiB, available {available_space / gib:.1f} GiB") if size <= available_space: return cached_files = [file for repo in cache_info.repos for revision in repo.revisions for file in revision.files] # Remove as few least recently used files as possible removed_files = [] freed_space = 0 extra_space_needed = size - available_space for file in sorted(cached_files, key=lambda file: file.blob_last_accessed): os.remove(file.file_path) # Remove symlink os.remove(file.blob_path) # Remove contents removed_files.append(file) freed_space += file.size_on_disk if freed_space >= extra_space_needed: break if removed_files: logger.info(f"Removed {len(removed_files)} files to free {freed_space / gib:.1f} GiB of disk space") logger.debug(f"Removed paths: {[str(file.file_path) for file in removed_files]}") if freed_space < extra_space_needed: raise RuntimeError( f"Insufficient disk space to load a block. Please free {(extra_space_needed - freed_space) / gib:.1f} GiB " f"on the volume for {cache_dir} or increase --max_disk_space if you set it manually" )
TheGrid-main
grid/utils/disk_cache.py
import os from dataclasses import dataclass from typing import Optional, Type, Union from transformers import AutoConfig, PretrainedConfig, PreTrainedModel from grid.utils.hf_auth import always_needs_auth @dataclass class _ModelClasses: config: Type[PretrainedConfig] model: Optional[Type[PreTrainedModel]] = None model_for_causal_lm: Optional[Type[PreTrainedModel]] = None model_for_sequence_classification: Optional[Type[PreTrainedModel]] = None _CLASS_MAPPING = {} # Populated by grid.models.* subpackages with register_model_classes() def register_model_classes(*, config: Type[PretrainedConfig], **kwargs): assert issubclass(config, PretrainedConfig) assert config.model_type not in _CLASS_MAPPING, f"Model type {config.model_type} is already registered" _CLASS_MAPPING[config.model_type] = _ModelClasses(config=config, **kwargs) class _AutoDistributedBase: _mapping_field = None # Should be defined in child classes @classmethod def from_pretrained(cls, model_name_or_path: Union[str, os.PathLike, None], *args, **kwargs) -> PretrainedConfig: if ( always_needs_auth(model_name_or_path) and kwargs.get("token") is None and kwargs.get("use_auth_token") is None ): kwargs["use_auth_token"] = True config = AutoConfig.from_pretrained(model_name_or_path, *args, **kwargs) if config.model_type not in _CLASS_MAPPING: raise ValueError(f"Grid does not support model type {config.model_type}") proper_cls = getattr(_CLASS_MAPPING[config.model_type], cls._mapping_field) if proper_cls is None: raise ValueError(f"Grid does not have {cls.__name__} for model type {config.model_type}") return proper_cls.from_pretrained(model_name_or_path, *args, **kwargs) class AutoDistributedConfig(_AutoDistributedBase): _mapping_field = "config" class AutoDistributedModel(_AutoDistributedBase): _mapping_field = "model" class AutoDistributedModelForCausalLM(_AutoDistributedBase): _mapping_field = "model_for_causal_lm" class AutoDistributedModelForSequenceClassification(_AutoDistributedBase): _mapping_field = "model_for_sequence_classification"
TheGrid-main
grid/utils/auto_config.py
import contextlib import re import time from typing import Optional, Sequence, Union import bitsandbytes as bnb import torch import torch.nn as nn import transformers from accelerate import init_empty_weights from hivemind.utils.logging import get_logger from huggingface_hub import HfFileSystem, get_hf_file_metadata, hf_hub_url from peft.tuners import lora from peft.utils import COMMON_LAYERS_PATTERN, CONFIG_NAME, SAFETENSORS_WEIGHTS_NAME, PeftConfig from safetensors import safe_open from safetensors.torch import load_file from transformers.utils import get_file_from_repo from grid.server.block_utils import resolve_block_dtype from grid.utils.convert_block import QuantType from grid.utils.disk_cache import allow_cache_reads, allow_cache_writes, free_disk_space_for logger = get_logger(__name__) def check_peft_repository(repo_id: str) -> bool: fs = HfFileSystem() list_of_files = fs.glob(f"{repo_id}/{SAFETENSORS_WEIGHTS_NAME}", detail=False) return len(list_of_files) > 0 def load_specific_module(block_idx: int, filepath: str, framework: str = "pt", device: Optional[int] = None): tensors = dict() is_tensors_found = dict() common_layer_patter_re = ( ".+\." + "".join(f"({common_name})?" for common_name in COMMON_LAYERS_PATTERN) + f"\.({block_idx})?\..+" ) with safe_open(filepath, framework=framework, device=device) as f: for k in f.keys(): if re.match(common_layer_patter_re, k): is_tensors_found[block_idx] = True tensors[k] = f.get_tensor(k) if not is_tensors_found.get(block_idx, False): logger.warning(f"There is no peft weights for block {block_idx}") return tensors def get_adapter_from_repo( repo_id: str, block_idx: Optional[int] = None, device: Optional[int] = None, *, token: Optional[Union[str, bool]] = None, **kwargs, ): config_path = get_file_from_repo(repo_id, CONFIG_NAME, use_auth_token=token, **kwargs) if config_path is None: raise RuntimeError(f"File {CONFIG_NAME} does not exist in repo {repo_id}") config = PeftConfig.from_json_file(config_path) weight_path = get_file_from_repo(repo_id, SAFETENSORS_WEIGHTS_NAME, use_auth_token=token, **kwargs) if weight_path is None: raise RuntimeError(f"File {SAFETENSORS_WEIGHTS_NAME} does not exist in repo {repo_id}") if block_idx is None: return config, load_file(weight_path) return config, load_specific_module(block_idx, weight_path, device=device) def load_peft( repo_id: str, block_idx: Optional[int] = None, device: Optional[int] = None, *, revision: Optional[str] = None, token: Optional[Union[str, bool]] = None, cache_dir: str, max_disk_space: Optional[int] = None, delay: float = 30, ): # TODO: Check is it possible to add safetensors loading inside grid/server/from_pretrained.py and reuse it here if not check_peft_repository(repo_id): raise ValueError(f"Repo: {repo_id} doesn't have safetensors inside for a safe loading.") try: with allow_cache_reads(cache_dir): return get_adapter_from_repo( repo_id, block_idx, device, revision=revision, token=token, cache_dir=cache_dir, local_files_only=False, ) except Exception: logger.warning(f"Cache for peft weights {repo_id} is corrupted, it will be downloaded again", exc_info=True) while True: try: with allow_cache_writes(cache_dir): config_url = hf_hub_url(repo_id, CONFIG_NAME, revision=revision) config_file_size = get_hf_file_metadata(config_url, token=token).size weight_url = hf_hub_url(repo_id, SAFETENSORS_WEIGHTS_NAME, revision=revision) weight_file_size = get_hf_file_metadata(weight_url, token=token).size file_size = config_file_size + weight_file_size if file_size is not None: free_disk_space_for(file_size, cache_dir=cache_dir, max_disk_space=max_disk_space) else: logger.warning(f"Failed to fetch size from peft repo {repo_id}") return get_adapter_from_repo( repo_id, block_idx, device, revision=revision, token=token, cache_dir=cache_dir, local_files_only=False, ) except Exception: logger.warning( f"Failed to load peft weights {repo_id} from HF Hub (retry in {delay:.0f} sec)", exc_info=True ) time.sleep(delay) class AdapterContextMixin: """A mixin that makes LoRA-wrapped linear layers obey an adapter set from context""" ADAPTER_NOT_SET = "__ADAPTER_NOT_SET" _context_active_adapter = ADAPTER_NOT_SET @staticmethod @contextlib.contextmanager def using_adapter(active_adapter: Optional[str]): prev, AdapterContextMixin._context_active_adapter = AdapterContextMixin._context_active_adapter, active_adapter try: yield finally: AdapterContextMixin._context_active_adapter = prev @property def active_adapter(self): if self._context_active_adapter == self.ADAPTER_NOT_SET: logger.warning(f"Layer {self} was called without using_adapter. This should only be used for debug") return self._context_active_adapter @active_adapter.setter def active_adapter(self, value: Optional[str]): assert value == self.ADAPTER_NOT_SET, "active adapter can only be changed via .using_adapter" "" using_adapter = AdapterContextMixin.using_adapter class LoraLinear(lora.Linear, AdapterContextMixin): """LoRA linear layer that uses adapter selected via using_adapter""" class LoraLinear8bitLt(lora.Linear8bitLt, AdapterContextMixin): """LoRA linear 8-bit with outliers that uses adapter selected via using_adapter""" class LoraLinear4bit(lora.Linear4bit, AdapterContextMixin): """LoRA linear 4-bit that uses adapter selected via using_adapter""" def create_lora_adapter(block, quant_type: QuantType): for _, module in block.named_modules(): for child_name, child in module.named_children(): lora_wrapped_child = None if not isinstance(child, (nn.Linear, bnb.nn.Linear8bitLt, bnb.nn.Linear4bit)): continue if quant_type == QuantType.INT8: kwargs = { "has_fp16_weights": False, "threshold": 6.0, "bias": hasattr(child, "bias") and child.bias is not None, } lora_wrapped_child = LoraLinear8bitLt( AdapterContextMixin.ADAPTER_NOT_SET, child.in_features, child.out_features, **kwargs, ) elif quant_type == QuantType.NF4: kwargs = { "compress_statistics": True, "quant_type": "nf4", "blocksize": 64, "bias": hasattr(child, "bias") and child.bias is not None, } lora_wrapped_child = LoraLinear4bit( AdapterContextMixin.ADAPTER_NOT_SET, child.in_features, child.out_features, **kwargs, ) lora_wrapped_child.compute_dtype = child.compute_dtype else: bias = hasattr(child, "bias") and child.bias is not None lora_wrapped_child = LoraLinear( AdapterContextMixin.ADAPTER_NOT_SET, child.in_features, child.out_features, bias=bias, ) if lora_wrapped_child: lora_wrapped_child.weight = child.weight lora_wrapped_child.bias = child.bias for p in lora_wrapped_child.parameters(): p.requires_grad = False setattr(module, child_name, lora_wrapped_child) def add_adapter_to_block(block, block_index, adapter_name, peft_config, peft_state_dict): assert peft_config["peft_type"] == "LORA", "Grid works only with LORA adapters" if peft_config["lora_dropout"] > 0: logger.info(f"Adapter {adapter_name} has dropout enabled, this server will disable dropout") for _, module in block.named_modules(): for child_name, child in module.named_children(): if not isinstance(child, (lora.Linear, lora.Linear8bitLt, lora.Linear4bit)): continue if child_name in peft_config["target_modules"] or ( isinstance(peft_config["target_modules"], str) and re.fullmatch(peft_config["target_modules"], child_name) ): is_lora_a_loaded = False is_lora_b_loaded = False for peft_key in peft_state_dict: if child_name not in peft_key: continue if adapter_name not in child.lora_A: child.update_layer( adapter_name, peft_config["r"], peft_config["lora_alpha"], lora_dropout=peft_config["lora_dropout"], init_lora_weights=peft_config["init_lora_weights"], ) child.train(False) for p in child.parameters(): p.requires_grad = False if peft_key.endswith(".lora_A.weight"): child.lora_A[adapter_name].weight[...] = peft_state_dict[peft_key] is_lora_a_loaded = True elif peft_key.endswith(".lora_A.bias"): raise NotImplementedError(f"LoRA adapters with bias not supported: {peft_key}") elif peft_key.endswith(".lora_B.weight"): child.lora_B[adapter_name].weight[...] = peft_state_dict[peft_key] is_lora_b_loaded = True elif peft_key.endswith(".lora_B.bias"): raise NotImplementedError(f"LoRA adapters with bias not supported: {peft_key}") if is_lora_a_loaded and is_lora_b_loaded: logger.debug(f"Loaded adapter {adapter_name} for block {block_index}.{child_name}") elif is_lora_a_loaded or is_lora_b_loaded: raise ValueError(f"Invalid adapter {adapter_name} for block {block_index}.{child_name}") logger.info(f"Loaded adapter {adapter_name} for block {block_index}") def estimate_adapter_memory_per_block( block_config: transformers.PretrainedConfig, torch_dtype: Optional[torch.dtype], adapters: Sequence[str], **load_peft_kwargs, ) -> int: """Get the number of extra bytes used to store a set of adapters per given block""" with init_empty_weights(include_buffers=True): block = block_config.block_class(block_config) base_block_parameters = sum(p.numel() for p in block.parameters()) create_lora_adapter(block, quant_type=QuantType.NONE) for adapter in adapters: peft_config, peft_state_dict = load_peft(adapter, block_idx=0, **load_peft_kwargs) assert peft_config["peft_type"].upper() == "LORA", "only LoRA adapters are supported for now" add_adapter_to_block( block, block_index=0, adapter_name=adapter, peft_config=peft_config, peft_state_dict=peft_state_dict ) adapter_parameters = sum(p.numel() for p in block.parameters()) - base_block_parameters bytes_per_parameter = torch.finfo(resolve_block_dtype(block_config, torch_dtype)).bits / 8 return adapter_parameters * bytes_per_parameter
TheGrid-main
grid/utils/peft.py
import asyncio async def shield_and_wait(task): """ Works like asyncio.shield(), but waits for the task to finish before raising CancelledError to the caller. """ if not isinstance(task, asyncio.Task): task = asyncio.create_task(task) cancel_exc = None while True: try: result = await asyncio.shield(task) break except asyncio.CancelledError as e: cancel_exc = e if cancel_exc is not None: raise cancel_exc return result
TheGrid-main
grid/utils/asyncio.py
TheGrid-main
grid/utils/__init__.py
from abc import ABC import torch class ABCBloomConstraint(ABC): """ Base class of all kind of decoding constraints. It can be used to implement a new constraint. """ def __init__(self) -> None: pass def __call__(self, tokens_id: torch.Tensor, logits: torch.Tensor, hypo_ids: torch.Tensor) -> torch.Tensor: """ This method is called by the decoding algorithm to apply the constraint. It changes and returns new logits. :param tokens_id: The token id of the last chosen token. :param logits: The logits from the Bloom model. :param hypo_ids: The hypothesis ids of the last tokens. """ pass class EosConstraint(ABCBloomConstraint): """ This constrained repeats EOS token if it was generated on the previous step. Args: prefix: The prefix of the sequence. eos_token_id: The id of the end of sentence token. pad_token_id: The id of the padding token. min_logits: The minimum logits that can be generated. Default: -1e6. """ def __init__(self, prefix: torch.Tensor, eos_token_id: int, pad_token_id: int, min_logits: float = -1e8) -> None: self.eos_token_id = eos_token_id self.min_logits = min_logits self.past_tokens = None self.wait_until_starting = (prefix == pad_token_id).sum(1).unsqueeze(1) def __call__(self, tokens_id: torch.Tensor, logits: torch.Tensor, hypo_ids: torch.Tensor) -> torch.Tensor: if self.past_tokens is not None: mask = (self.wait_until_starting < 0) & (self.past_tokens == self.eos_token_id) logits += self.min_logits * mask logits[mask[:, 0], self.eos_token_id] = 0 if tokens_id is not None: self.past_tokens = tokens_id self.wait_until_starting -= 1 return logits
TheGrid-main
grid/utils/generation_constraints.py
import random from typing import Collection, TypeVar T = TypeVar("T") def sample_up_to(population: Collection[T], k: int) -> T: if not isinstance(population, list): population = list(population) if len(population) > k: population = random.sample(population, k) return population
TheGrid-main
grid/utils/random.py
""" Tools for converting transformer blocks, applying quantization and/or tensor parallelism """ import re from enum import Enum from typing import Optional, Sequence import tensor_parallel as tp import torch import torch.nn as nn from hivemind.utils.logging import get_logger, use_hivemind_log_handler from tensor_parallel.slicing_configs import get_bloom_config from transformers import PretrainedConfig use_hivemind_log_handler("in_root_logger") logger = get_logger(__name__) class QuantType(Enum): NONE = 0 INT8 = 1 # 8-bit as in the LLM.int8() paper NF4 = 2 # 4-bit as in the QLoRA paper def convert_block( block: nn.Module, block_index: int, config: PretrainedConfig, tensor_parallel_devices: Sequence[torch.device], output_device: torch.device, quant_type: QuantType, freeze: bool = True, adapters: Optional[Sequence[str]] = None, **kwargs, ) -> tp.TensorParallel: """ Optimize a transformer block for use in a Grid server, apply tensor parallelism and/or LLM.8bit quantization :note: some optimizations will modify the input block in-place! :param block: a single transformer block, either pre-trained or newly initialized :param config: HF transformers config for the full model :param tensor_parallel_devices: if specified, use tensor parallelism to split the model between these devices :note: if there is only a single device, model wil still be wrapped with TensorParallel (for uniformity) :param output_device: if tensor_parallel_devices is True, output :param quant_type: quantization type :param freeze: if True (default), make all module parameters non-trainable :return: a module that acts like the original block, but runs with all specified optimizations """ if freeze: block.requires_grad_(False) block = make_tensor_parallel(block, config, tensor_parallel_devices, output_device=output_device) if quant_type != QuantType.NONE: block = quantize_module(block, quant_type=quant_type) for shard, device in zip(block.module_shards, block.devices): shard.to(device) if adapters: from grid.utils.peft import add_adapter_to_block, create_lora_adapter, load_peft create_lora_adapter(block, quant_type=quant_type) for adapter_name in adapters: adapter_config, adapter_state_dict = load_peft( adapter_name, block_idx=block_index, **kwargs, ) add_adapter_to_block(block, block_index, adapter_name, adapter_config, adapter_state_dict) return block def quantize_module(model: nn.Module, *, quant_type: QuantType) -> nn.Module: # Import bitsandbytes only when necessary, so Grid runs on platforms not supported by bitsandbytes import bitsandbytes as bnb for n, module in model.named_children(): if len(list(module.children())) > 0: quantize_module(module, quant_type=quant_type) if isinstance(module, torch.nn.Linear) and n not in ["lm_head", "score"]: assert module.weight.device.type == "cpu", f"expected linear layers on CPU, got {module.weight.device}" if quant_type == QuantType.INT8: model._modules[n] = bnb.nn.Linear8bitLt( module.in_features, module.out_features, module.bias is not None, has_fp16_weights=False, threshold=6.0, # Default from the LLM.int8() paper ) model._modules[n].weight = bnb.nn.Int8Params( module.weight.data, requires_grad=False, has_fp16_weights=False ).to(module.weight.dtype) elif quant_type == QuantType.NF4: compress_statistics = True model._modules[n] = bnb.nn.LinearNF4( module.in_features, module.out_features, module.bias is not None, compress_statistics=compress_statistics, ) model._modules[n].weight = bnb.nn.Params4bit( module.weight.data, requires_grad=False, quant_type="nf4", blocksize=64, compress_statistics=compress_statistics, ).to(module.weight.dtype) else: raise ValueError(f"Unsupported quant_type='{quant_type}'") model._modules[n].bias = module.bias return model def make_tensor_parallel( block: nn.Module, model_config: PretrainedConfig, devices: Sequence[torch.device], output_device: torch.device ) -> nn.Module: if model_config.model_type == "bloom": tp_config = get_bloom_config(model_config, devices) del tp_config.state_rules[re.compile(".*word_embeddings.weight$")] else: if len(devices) > 1: logger.warning("Tensor parallelism is not tested for models other than BLOOM yet, proceed with caution") tp_config = None tp_block = tp.TensorParallel(block, devices, config=tp_config, output_device=output_device, delay_init=True) total_heads = 0 for tp_shard in tp_block.module_shards: for submodule in tp_shard.modules(): if isinstance(submodule, model_config.attn_class): total_heads += submodule.num_heads assert total_heads == model_config.num_attention_heads return tp_block def check_device_balance(devices: Sequence[torch.device]): if not all(device.type == "cuda" for device in devices): logger.warning("Running tensor parallelism on non-GPU devices; proceed at your own risk") return unique_device_capabilities = set(map(torch.cuda.get_device_capability, devices)) if len(unique_device_capabilities) > 1: logger.warning( f"Found GPUs with uneven capabilities: {unique_device_capabilities}. " f"Using GPUs with different performance will cause the server to wait for the slowest GPU." ) memory_per_device = tuple(torch.cuda.get_device_properties(device).total_memory for device in devices) used_memory = min(memory_per_device) * len(memory_per_device) wasted_memory_rate = (sum(memory_per_device) - used_memory) / sum(memory_per_device) if wasted_memory_rate > 0.05: logger.warning( f"GPU devices have highly uneven memory, {wasted_memory_rate * 100:.2f}% memory is wasted. " f"Consider running high-memory GPUs in a separate server." )
TheGrid-main
grid/utils/convert_block.py
import os from typing import Union def always_needs_auth(model_name: Union[str, os.PathLike, None]) -> bool: loading_from_repo = model_name is not None and not os.path.isdir(model_name) return loading_from_repo and model_name.startswith("meta-llama/Llama-2-")
TheGrid-main
grid/utils/hf_auth.py
from abc import ABC, abstractmethod from typing import Tuple import torch TokenIds = torch.Tensor HypoIds = torch.Tensor class DecodingAlgorithm(ABC): """ An abstract class for decoding algorithms. Describes the base function of those algorithms: they have to select new tokens and provide the corresponding hypotheses. """ @abstractmethod def __call__(self, logits: torch.Tensor) -> Tuple[TokenIds, HypoIds]: """ :param logits: A tensor of shape (batch_size, seq_length, vocab_size) :return: A tuple of selected token ids and corresponding hypotheses. The shape of the token ids is (batch_size, seq_length), and the shape of the hypotheses is (batch_size) """ pass class GreedyAlgorithm(DecodingAlgorithm): """ The simplest algorithm for decoding. It selects the most probable token. """ def __call__(self, logits: torch.Tensor) -> Tuple[TokenIds, HypoIds]: """ Returns the most probable token. The second returned object is always a range of integers from 0 to batch_size - 1. """ return logits.max(-1)[1].unsqueeze(1), torch.arange(logits.size(0)) class SamplingAlgorithm(DecodingAlgorithm): def __init__(self, temperature: float = 1.0): self.temperature = temperature def sample(self, logits: torch.Tensor, indices_to_remove: torch.Tensor) -> Tuple[TokenIds, HypoIds]: """ :param logits: A tensor of shape (batch_size * num_hypos, vocab_size) :param indices_to_remove: A bool tensor of shape (batch_size * num_hypos, vocab_size) :return: A tuple of selected token ids and corresponding hypotheses. The shape of the token ids is (batch_size, seq_length), and the shape of the hypotheses is (batch_size). """ logits[indices_to_remove] = -float("Inf") probs = torch.softmax(logits / self.temperature, -1) return torch.multinomial(probs, num_samples=1), torch.arange(logits.size(0)) def __call__(self, logits: torch.Tensor) -> Tuple[TokenIds, HypoIds]: indices_to_remove = torch.full_like(logits, False, dtype=torch.bool) return self.sample(logits, indices_to_remove) class TopKAlgorithm(SamplingAlgorithm): def __init__(self, top_k: int, temperature: float = 1.0) -> None: super().__init__(temperature=temperature) self.top_k = top_k def __call__(self, logits: torch.Tensor) -> Tuple[TokenIds, HypoIds]: indices_to_remove = logits < torch.topk(logits, self.top_k, dim=-1)[0][..., -1, None] return self.sample(logits, indices_to_remove) class NucleusAlgorithm(SamplingAlgorithm): def __init__(self, top_p: float, temperature: float = 1.0) -> None: super().__init__(temperature=temperature) self.top_p = top_p def __call__(self, logits: torch.Tensor) -> Tuple[TokenIds, HypoIds]: sorted_logits, sorted_indices = torch.sort(logits, descending=False, dim=-1) probs = torch.softmax(sorted_logits / self.temperature, -1) cumulative_probs = torch.cumsum(probs, dim=-1) sorted_indices_to_remove = cumulative_probs <= (1 - self.top_p) indices_to_remove = sorted_indices_to_remove.scatter(1, sorted_indices, sorted_indices_to_remove) return self.sample(logits, indices_to_remove) class BeamSearchAlgorithm(DecodingAlgorithm): def __init__(self, num_beams: int, batch_size: int) -> None: self.num_beams = num_beams self.batch_size = batch_size self._batch_beams = [list() for _ in range(batch_size)] def __call__(self, logits: torch.Tensor): sorted_logits, sorted_indices = torch.sort(logits, descending=True, dim=-1) probs = torch.log_softmax(sorted_logits, -1) if len(self._batch_beams[0]) > 0: for batch_idx in range(self.batch_size): new_beams = [] cur_beams = self._batch_beams[batch_idx] for beam_idx in range(len(cur_beams)): probs_idx = batch_idx + beam_idx * self.batch_size new_beam = cur_beams[beam_idx] for hypo_idx in range(self.num_beams): new_beams.append( (new_beam[0] + probs[probs_idx, hypo_idx].item(), beam_idx * self.num_beams + hypo_idx) ) self._batch_beams[batch_idx] = sorted(new_beams, reverse=True)[: self.num_beams] else: for batch_idx in range(self.batch_size): for beam_idx in range(self.num_beams): self._batch_beams[batch_idx].append((probs[batch_idx, beam_idx].item(), beam_idx)) return_hypos = [] return_tokens = [] for batch_idx in range(self.batch_size): cur_beam = self._batch_beams[batch_idx] return_hypos.append(list()) return_tokens.append(list()) for beam in cur_beam: beam_idx = beam[1] // self.num_beams hypo_idx = batch_idx + beam_idx * self.batch_size token_idx = beam[1] % self.num_beams return_hypos[-1].append(hypo_idx) return_tokens[-1].append([sorted_indices[hypo_idx, token_idx].item()]) return_hypos = [hypo_idx for hypo_indexes in zip(*return_hypos) for hypo_idx in hypo_indexes] return_tokens = [token_idx for token_indexes in zip(*return_tokens) for token_idx in token_indexes] return torch.tensor(return_tokens), torch.tensor(return_hypos)
TheGrid-main
grid/utils/generation_algorithms.py
import asyncio import math import threading import time from functools import partial from typing import Dict, Sequence import hivemind from hivemind.proto import dht_pb2 from hivemind.utils.logging import get_logger logger = get_logger(__name__) async def ping( peer_id: hivemind.PeerID, _dht: hivemind.DHT, node: hivemind.dht.DHTNode, *, wait_timeout: float = 5, ) -> float: try: ping_request = dht_pb2.PingRequest(peer=node.protocol.node_info) start_time = time.perf_counter() await node.protocol.get_stub(peer_id).rpc_ping(ping_request, timeout=wait_timeout) return time.perf_counter() - start_time except Exception: logger.debug(f"Failed to ping {peer_id}:", exc_info=True) return math.inf async def ping_parallel(peer_ids: Sequence[hivemind.PeerID], *args, **kwargs) -> Dict[hivemind.PeerID, float]: rpc_infos = await asyncio.gather(*[ping(peer_id, *args, **kwargs) for peer_id in peer_ids]) return dict(zip(peer_ids, rpc_infos)) class PingAggregator: def __init__(self, dht: hivemind.DHT, *, ema_alpha: float = 0.2, expiration: float = 300): self.dht = dht self.ema_alpha = ema_alpha self.expiration = expiration self.ping_emas = hivemind.TimedStorage() self.lock = threading.Lock() def ping(self, peer_ids: Sequence[hivemind.PeerID], **kwargs) -> None: current_rtts = self.dht.run_coroutine(partial(ping_parallel, peer_ids, **kwargs)) logger.debug(f"Current RTTs: {current_rtts}") with self.lock: expiration = hivemind.get_dht_time() + self.expiration for peer_id, rtt in current_rtts.items(): prev_rtt = self.ping_emas.get(peer_id) if prev_rtt is not None and prev_rtt.value != math.inf: rtt = self.ema_alpha * rtt + (1 - self.ema_alpha) * prev_rtt.value # Exponential smoothing self.ping_emas.store(peer_id, rtt, expiration) def to_dict(self) -> Dict[hivemind.PeerID, float]: with self.lock, self.ping_emas.freeze(): smoothed_rtts = {peer_id: rtt.value for peer_id, rtt in self.ping_emas.items()} logger.debug(f"Smothed RTTs: {smoothed_rtts}") return smoothed_rtts
TheGrid-main
grid/utils/ping.py
from grid.models.bloom import * from grid.models.llama import *
TheGrid-main
grid/models/__init__.py
import os from typing import Optional, Union from hivemind import get_logger from transformers.models.bloom import BloomConfig from transformers.models.bloom.modeling_bloom import BloomAttention from grid.client.lm_head import LMHeadConfig from grid.client.ptune import PTuneConfig from grid.client.routing.sequence_manager import SequenceManagerConfig from grid.models.bloom.block import WrappedBloomBlock logger = get_logger(__name__) class DistributedBloomConfig(BloomConfig, SequenceManagerConfig, PTuneConfig, LMHeadConfig): block_class = WrappedBloomBlock attn_class = BloomAttention block_prefix = "h" num_key_value_groups = 1 @classmethod def from_pretrained( cls, model_name_or_path: Union[str, os.PathLike, None], *args, dht_prefix: Optional[str] = None, **kwargs ): logger.info("Make sure you follow the BLOOM's terms of use: https://bit.ly/bloom-license") loading_from_repo = model_name_or_path is not None and not os.path.isdir(model_name_or_path) if loading_from_repo and dht_prefix is None: # We need "-grid" for backward compatibility with Grid < 1.2.0 dht_prefix = str(model_name_or_path) + "-grid" logger.info(f"Using DHT prefix: {dht_prefix}") return super().from_pretrained(model_name_or_path, *args, dht_prefix=dht_prefix, **kwargs)
TheGrid-main
grid/models/bloom/config.py
from grid.models.bloom.config import DistributedBloomConfig from grid.models.bloom.model import ( DistributedBloomForCausalLM, DistributedBloomForSequenceClassification, DistributedBloomModel, ) from grid.utils.auto_config import register_model_classes register_model_classes( config=DistributedBloomConfig, model=DistributedBloomModel, model_for_causal_lm=DistributedBloomForCausalLM, model_for_sequence_classification=DistributedBloomForSequenceClassification, )
TheGrid-main
grid/models/bloom/__init__.py
from typing import Optional import hivemind import torch import torch.nn as nn from hivemind.utils.logging import get_logger from transformers.modeling_outputs import BaseModelOutputWithPastAndCrossAttentions from transformers.models.bloom import BloomForCausalLM, BloomForSequenceClassification, BloomModel, BloomPreTrainedModel from grid.client.from_pretrained import FromPretrainedMixin from grid.client.lm_head import LMHead from grid.client.ptune import PTuneMixin from grid.client.remote_generation import RemoteGenerationMixin from grid.client.remote_sequential import RemoteSequential from grid.models.bloom.config import DistributedBloomConfig logger = get_logger(__name__) class DistributedBloomModel(FromPretrainedMixin, PTuneMixin, BloomModel): """BloomModel, but all transformer layers are hosted by the swarm""" _keys_to_ignore_on_load_missing = PTuneMixin._keys_to_ignore_on_load_missing _keys_to_ignore_on_load_unexpected = [r"^h\."] config_class = DistributedBloomConfig def __init__(self, config: DistributedBloomConfig, *, dht: Optional[hivemind.DHT] = None): n_layer, config.num_hidden_layers = config.num_hidden_layers, 0 # Prevent initialization super().__init__(config) assert len(self.h) == 0 config.num_hidden_layers = n_layer self.h = RemoteSequential(config, dht=dht) self.requires_grad_(False) # Forbid accumulate grads for embeddings and layernorm self.init_prompts(config) def forward( self, input_ids: Optional[torch.LongTensor] = None, inputs_embeds: Optional[torch.Tensor] = None, attention_mask: Optional[torch.Tensor] = None, **kwargs, ): assert attention_mask is None, f"{self.__class__.__name__} does not support attention masks right now" for k, v in kwargs.items(): if not (v is None or v is False): logger.debug(f"Extra keyword arguments are not yet supported (got {k} = {v})") if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time") elif input_ids is not None: input_shape = input_ids.size() input_ids = input_ids.view(-1, input_shape[-1]) elif inputs_embeds is not None: input_shape = inputs_embeds.size()[:-1] else: raise ValueError("You have to specify either input_ids or inputs_embeds") if inputs_embeds is None: inputs_embeds = self.word_embeddings(input_ids) if self.config.tuning_mode and "ptune" in self.config.tuning_mode: batch_size = inputs_embeds.shape[0] prompts, intermediate_prompts = self.get_prompt(batch_size) inputs_embeds = torch.cat([prompts, inputs_embeds], dim=1) hidden_states = self.word_embeddings_layernorm(inputs_embeds) output_shape = input_shape + (hidden_states.size(-1),) if self.config.tuning_mode and "ptune" in self.config.tuning_mode: hidden_states = self.h(hidden_states, prompts=intermediate_prompts) else: hidden_states = self.h(hidden_states) # Remove prefix if self.config.tuning_mode and "ptune" in self.config.tuning_mode: hidden_states = hidden_states[:, self.pre_seq_len :] # Add last hidden state hidden_states = self.ln_f(hidden_states) hidden_states = hidden_states.view(output_shape) return BaseModelOutputWithPastAndCrossAttentions( last_hidden_state=hidden_states, past_key_values=None, hidden_states=None, attentions=None, ) class DistributedBloomForCausalLM(FromPretrainedMixin, RemoteGenerationMixin, BloomForCausalLM): _keys_to_ignore_on_load_missing = DistributedBloomModel._keys_to_ignore_on_load_missing _keys_to_ignore_on_load_missing += [r"^lm_head\."] # Missing since they are shared with input embeddings _keys_to_ignore_on_load_unexpected = DistributedBloomModel._keys_to_ignore_on_load_unexpected config_class = DistributedBloomConfig def __init__(self, config: DistributedBloomConfig): BloomPreTrainedModel.__init__(self, config) self.transformer = DistributedBloomModel(config) self.lm_head = LMHead(config) # Initialize weights and apply final processing self.post_init() def get_output_embeddings(self): return self.lm_head class DistributedBloomForSequenceClassification(FromPretrainedMixin, BloomForSequenceClassification): _keys_to_ignore_on_load_missing = DistributedBloomModel._keys_to_ignore_on_load_missing _keys_to_ignore_on_load_unexpected = DistributedBloomModel._keys_to_ignore_on_load_unexpected config_class = DistributedBloomConfig def __init__(self, config: DistributedBloomConfig): BloomPreTrainedModel.__init__(self, config) self.num_labels = config.num_labels self.transformer = DistributedBloomModel(config) self.score = nn.Linear(config.hidden_size, config.num_labels, bias=False) # Initialize weights and apply final processing self.post_init()
TheGrid-main
grid/models/bloom/model.py
""" Bloom intermediate layer Based on https://github.com/huggingface/transformers/commit/ca2a55e9dfb245527b5e1c954fec6ffbb7aef07b See commit history for authorship. """ from typing import Optional, Tuple import torch from transformers.models.bloom.modeling_bloom import BloomBlock, BloomModel, build_alibi_tensor class WrappedBloomBlock(BloomBlock): def forward( self, hidden_states: torch.Tensor, *args, attention_mask: Optional[torch.Tensor] = None, alibi: Optional[torch.Tensor] = None, layer_past: Optional[Tuple[torch.Tensor, torch.Tensor]] = None, **kwargs ): assert attention_mask is None, "Non-causal attention masks are not supported yet" batch_size, seq_length = hidden_states.shape[:2] past_length = 0 if layer_past is None else layer_past[0].shape[-1] seq_length_with_past = seq_length + past_length attention_mask = torch.ones((batch_size, seq_length_with_past), device=hidden_states.device) if alibi is None: alibi = build_alibi_tensor(attention_mask, num_heads=self.num_heads, dtype=hidden_states.dtype) attention_mask = BloomModel._prepare_attn_mask(None, attention_mask, (batch_size, seq_length), past_length) return super().forward( hidden_states, *args, attention_mask=attention_mask, alibi=alibi, layer_past=layer_past, **kwargs )
TheGrid-main
grid/models/bloom/block.py
import os from typing import Optional, Union from hivemind import get_logger from transformers.models.llama import LlamaConfig from transformers.models.llama.modeling_llama import LlamaAttention from grid.client.lm_head import LMHeadConfig from grid.client.ptune import PTuneConfig from grid.client.routing.sequence_manager import SequenceManagerConfig from grid.models.llama.block import WrappedLlamaBlock logger = get_logger(__name__) class DistributedLlamaConfig(LlamaConfig, SequenceManagerConfig, PTuneConfig, LMHeadConfig): block_class = WrappedLlamaBlock attn_class = LlamaAttention block_prefix = "model.layers" @property def num_key_value_groups(self): return self.num_attention_heads // self.num_key_value_heads @classmethod def from_pretrained( cls, model_name_or_path: Union[str, os.PathLike, None], *args, dht_prefix: Optional[str] = None, **kwargs ): logger.info( "Make sure you follow the LLaMA's terms of use: " "https://bit.ly/llama2-license for LLaMA 2, https://bit.ly/llama-license for LLaMA 1" ) loading_from_repo = model_name_or_path is not None and not os.path.isdir(model_name_or_path) if loading_from_repo and dht_prefix is None: dht_prefix = str(model_name_or_path) dht_prefix = dht_prefix.split("/")[-1] # Use only repo name to merge blocks hosted by different accounts if not dht_prefix.endswith("-hf"): dht_prefix += "-hf" logger.info(f"Using DHT prefix: {dht_prefix}") result = super().from_pretrained(model_name_or_path, *args, dht_prefix=dht_prefix, **kwargs) config = result[0] if isinstance(result, tuple) else result config.pretraining_tp = 1 # This may give less accurate results but it doesn't matter if we use quantization return result
TheGrid-main
grid/models/llama/config.py
from grid.models.llama.config import DistributedLlamaConfig from grid.models.llama.model import ( DistributedLlamaForCausalLM, DistributedLlamaForSequenceClassification, DistributedLlamaModel, ) from grid.utils.auto_config import register_model_classes register_model_classes( config=DistributedLlamaConfig, model=DistributedLlamaModel, model_for_causal_lm=DistributedLlamaForCausalLM, model_for_sequence_classification=DistributedLlamaForSequenceClassification, )
TheGrid-main
grid/models/llama/__init__.py
from typing import Optional import hivemind import torch import torch.nn as nn from hivemind.utils.logging import get_logger from transformers.modeling_outputs import BaseModelOutputWithPast from transformers.models.llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel from grid.client.from_pretrained import FromPretrainedMixin from grid.client.lm_head import LMHead from grid.client.ptune import PTuneMixin from grid.client.remote_generation import RemoteGenerationMixin from grid.client.remote_sequential import RemoteSequential from grid.models.llama.config import DistributedLlamaConfig logger = get_logger(__name__) class DistributedLlamaModel(FromPretrainedMixin, PTuneMixin, LlamaModel): """LlamaModel, but all transformer layers are hosted by the swarm""" _keys_to_ignore_on_load_missing = PTuneMixin._keys_to_ignore_on_load_missing _keys_to_ignore_on_load_unexpected = [r"^model\.layers\."] config_class = DistributedLlamaConfig def __init__(self, config: DistributedLlamaConfig, *, dht: Optional[hivemind.DHT] = None): n_layer, config.num_hidden_layers = config.num_hidden_layers, 0 # Prevent initialization super().__init__(config) assert len(self.layers) == 0 config.num_hidden_layers = n_layer self.layers = RemoteSequential(config, dht=dht) self.requires_grad_(False) # Forbid accumulate grads for embeddings and layernorm self.init_prompts(config) def forward( self, input_ids: Optional[torch.LongTensor] = None, inputs_embeds: Optional[torch.Tensor] = None, attention_mask: Optional[torch.Tensor] = None, **kwargs, ) -> BaseModelOutputWithPast: assert attention_mask is None, f"{self.__class__.__name__} does not support attention masks right now" for k, v in kwargs.items(): if not (v is None or v is False): logger.debug(f"Extra keyword arguments are not yet supported (got {k} = {v})") if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time") elif input_ids is not None: input_shape = input_ids.size() input_ids = input_ids.view(-1, input_shape[-1]) elif inputs_embeds is not None: input_shape = inputs_embeds.size()[:-1] else: raise ValueError("You have to specify either input_ids or inputs_embeds") if inputs_embeds is None: inputs_embeds = self.embed_tokens(input_ids) if self.config.tuning_mode and "ptune" in self.config.tuning_mode: batch_size = inputs_embeds.shape[0] prompts, intermediate_prompts = self.get_prompt(batch_size) inputs_embeds = torch.cat([prompts, inputs_embeds], dim=1) hidden_states = inputs_embeds output_shape = input_shape + (hidden_states.size(-1),) if self.config.tuning_mode and "ptune" in self.config.tuning_mode: hidden_states = self.layers(hidden_states, prompts=intermediate_prompts) else: hidden_states = self.layers(hidden_states) # Remove prefix if self.config.tuning_mode and "ptune" in self.config.tuning_mode: hidden_states = hidden_states[:, self.pre_seq_len :] # Add last hidden state hidden_states = self.norm(hidden_states) hidden_states = hidden_states.view(output_shape) return BaseModelOutputWithPast( last_hidden_state=hidden_states, past_key_values=None, hidden_states=None, attentions=None, ) @property def word_embeddings(self) -> nn.Embedding: # For compatibility with RemoteGenerationMixin return self.embed_tokens @property def word_embeddings_layernorm(self) -> nn.Module: # For compatibility with RemoteGenerationMixin return nn.Identity() @property def h(self) -> RemoteSequential: # For compatibility with RemoteGenerationMixin return self.layers @property def ln_f(self) -> nn.Module: # For compatibility with RemoteGenerationMixin return self.norm class DistributedLlamaForCausalLM(FromPretrainedMixin, RemoteGenerationMixin, LlamaForCausalLM): _keys_to_ignore_on_load_missing = DistributedLlamaModel._keys_to_ignore_on_load_missing _keys_to_ignore_on_load_unexpected = DistributedLlamaModel._keys_to_ignore_on_load_unexpected config_class = DistributedLlamaConfig def __init__(self, config: DistributedLlamaConfig): LlamaPreTrainedModel.__init__(self, config) self.model = DistributedLlamaModel(config) self.pretraining_tp = config.pretraining_tp self.vocab_size = config.vocab_size self.lm_head = LMHead(config) # Initialize weights and apply final processing self.post_init() def get_output_embeddings(self): return self.lm_head @property def transformer(self) -> DistributedLlamaModel: # For compatibility with RemoteGenerationMixin return self.model class DistributedLlamaForSequenceClassification(FromPretrainedMixin, LlamaForSequenceClassification): _keys_to_ignore_on_load_missing = DistributedLlamaModel._keys_to_ignore_on_load_missing _keys_to_ignore_on_load_unexpected = DistributedLlamaModel._keys_to_ignore_on_load_unexpected config_class = DistributedLlamaConfig def __init__(self, config): LlamaPreTrainedModel.__init__(self, config) self.num_labels = config.num_labels self.model = DistributedLlamaModel(config) self.score = nn.Linear(config.hidden_size, self.num_labels, bias=False) # Initialize weights and apply final processing self.post_init() @property def transformer(self) -> DistributedLlamaModel: # For compatibility with RemoteGenerationMixin return self.model
TheGrid-main
grid/models/llama/model.py
""" LLaMA intermediate layer Based on https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py See commit history for authorship. """ from typing import Optional, Tuple import torch from transformers.models.llama.modeling_llama import LlamaDecoderLayer, LlamaModel class WrappedLlamaBlock(LlamaDecoderLayer): def forward( self, hidden_states: torch.Tensor, *args, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, layer_past: Optional[Tuple[torch.Tensor]] = None, use_cache: bool = False, **kwargs, ) -> Tuple[torch.FloatTensor, Optional[Tuple[torch.FloatTensor, torch.FloatTensor]]]: batch_size, seq_length, _ = hidden_states.shape seq_length_with_past = seq_length past_key_values_length = 0 past_key_value = layer_past if past_key_value is not None: past_key_values_length = past_key_value[0].shape[2] seq_length_with_past = seq_length_with_past + past_key_values_length past_key_value = self._reorder_cache_from_bloom_to_llama(past_key_value, batch_size, past_key_values_length) if position_ids is None: device = hidden_states.device position_ids = torch.arange( past_key_values_length, seq_length + past_key_values_length, dtype=torch.long, device=device ) position_ids = position_ids.unsqueeze(0).view(-1, seq_length) else: position_ids = position_ids.view(-1, seq_length).long() # embed positions if attention_mask is None: attention_mask = torch.ones( (batch_size, seq_length_with_past), dtype=torch.bool, device=hidden_states.device ) attention_mask = LlamaModel._prepare_decoder_attention_mask( None, attention_mask, (batch_size, seq_length), hidden_states, past_key_values_length ) outputs = super().forward( hidden_states, *args, attention_mask=attention_mask, position_ids=position_ids, past_key_value=past_key_value, use_cache=use_cache, **kwargs, ) if use_cache: present_key_value = outputs[-1] present_key_value = self._reorder_cache_from_llama_to_bloom( present_key_value, batch_size, seq_length_with_past ) outputs = outputs[:-1] + (present_key_value,) return outputs def _reorder_cache_from_bloom_to_llama( self, key_value: Tuple[torch.Tensor], batch_size: int, seq_length: int ) -> Tuple[torch.Tensor]: key_states, value_states = key_value key_states = key_states.permute(0, 2, 1) key_states = key_states.view( batch_size, self.self_attn.num_key_value_heads, seq_length, self.self_attn.head_dim ) value_states = value_states.view(*key_states.shape) return (key_states, value_states) def _reorder_cache_from_llama_to_bloom( self, key_value: Tuple[torch.Tensor], batch_size: int, seq_length: int ) -> Tuple[torch.Tensor]: key_states, value_states = key_value value_states = value_states.view( batch_size * self.self_attn.num_key_value_heads, seq_length, self.self_attn.head_dim ) key_states = key_states.view(*value_states.shape) key_states = key_states.permute(0, 2, 1) return (key_states, value_states)
TheGrid-main
grid/models/llama/block.py
""" A copy of run_dht.py from hivemind with the ReachabilityProtocol added: https://github.com/learning-at-home/hivemind/blob/master/hivemind/hivemind_cli/run_dht.py This script may be used for launching lightweight CPU machines serving as bootstrap nodes to a Grid swarm. This may be eventually merged to the hivemind upstream. """ import time from argparse import ArgumentParser from secrets import token_hex from hivemind.dht import DHT, DHTNode from hivemind.utils.logging import get_logger, use_hivemind_log_handler from hivemind.utils.networking import log_visible_maddrs from grid.server.reachability import ReachabilityProtocol use_hivemind_log_handler("in_root_logger") logger = get_logger(__name__) async def report_status(dht: DHT, node: DHTNode): logger.info( f"{len(node.protocol.routing_table.uid_to_peer_id) + 1} DHT nodes (including this one) " f"are in the local routing table " ) logger.debug(f"Routing table contents: {node.protocol.routing_table}") logger.info(f"Local storage contains {len(node.protocol.storage)} keys") logger.debug(f"Local storage contents: {node.protocol.storage}") # Contact peers and keep the routing table healthy (remove stale PeerIDs) await node.get(f"heartbeat_{token_hex(16)}", latest=True) def main(): parser = ArgumentParser() parser.add_argument( "--initial_peers", nargs="*", help="Multiaddrs of the peers that will welcome you into the existing DHT. " "Example: /ip4/203.0.113.1/tcp/31337/p2p/XXXX /ip4/203.0.113.2/tcp/7777/p2p/YYYY", ) parser.add_argument( "--host_maddrs", nargs="*", default=["/ip4/0.0.0.0/tcp/0", "/ip6/::/tcp/0"], help="Multiaddrs to listen for external connections from other DHT instances. " "Defaults to all IPv4 interfaces and the TCP protocol: /ip4/0.0.0.0/tcp/0", ) parser.add_argument( "--announce_maddrs", nargs="*", help="Visible multiaddrs the host announces for external connections from other DHT instances", ) parser.add_argument( "--use_ipfs", action="store_true", help='Use IPFS to find initial_peers. If enabled, you only need to provide the "/p2p/XXXX" ' "part of the multiaddrs for the initial_peers " "(no need to specify a particular IPv4/IPv6 host and port)", ) parser.add_argument( "--identity_path", help="Path to a private key file. If defined, makes the peer ID deterministic. " "If the file does not exist, writes a new private key to this file.", ) parser.add_argument( "--no_relay", action="store_false", dest="use_relay", help="Disable circuit relay functionality in libp2p (see https://docs.libp2p.io/concepts/nat/circuit-relay/)", ) parser.add_argument( "--use_auto_relay", action="store_true", help="Look for libp2p relays to reach peers behind NATs/firewalls" ) parser.add_argument( "--refresh_period", type=int, default=30, help="Period (in seconds) for fetching the keys from DHT" ) args = parser.parse_args() dht = DHT( start=True, initial_peers=args.initial_peers, host_maddrs=args.host_maddrs, announce_maddrs=args.announce_maddrs, use_ipfs=args.use_ipfs, identity_path=args.identity_path, use_relay=args.use_relay, use_auto_relay=args.use_auto_relay, ) log_visible_maddrs(dht.get_visible_maddrs(), only_p2p=args.use_ipfs) ReachabilityProtocol.attach_to_dht(dht, await_ready=True) while True: dht.run_coroutine(report_status, return_future=False) time.sleep(args.refresh_period) if __name__ == "__main__": main()
TheGrid-main
grid/cli/run_dht.py
TheGrid-main
grid/cli/__init__.py
import argparse import configargparse from hivemind.proto.runtime_pb2 import CompressionType from hivemind.utils.limits import increase_file_limit from hivemind.utils.logging import get_logger from humanfriendly import parse_size from grid.constants import DTYPE_MAP, PUBLIC_INITIAL_PEERS from grid.server.server import Server from grid.utils.convert_block import QuantType from grid.utils.version import validate_version logger = get_logger(__name__) def main(): # fmt:off parser = configargparse.ArgParser(default_config_files=["config.yml"], formatter_class=argparse.ArgumentDefaultsHelpFormatter) parser.add('-c', '--config', required=False, is_config_file=True, help='config file path') group = parser.add_mutually_exclusive_group(required=True) group.add_argument('--converted_model_name_or_path', type=str, default=None, help="path or name of a pretrained model, converted with cli/convert_model.py") group.add_argument('model', nargs='?', type=str, help="same as --converted_model_name_or_path") parser.add_argument("--public_name", type=str, default=None, help="Public name to be reported in the leaderboard") group = parser.add_mutually_exclusive_group(required=False) group.add_argument("--token", type=str, default=None, help="Hugging Face hub auth token for .from_pretrained()") group.add_argument("--use_auth_token", action="store_true", dest="token", help="Read token saved by `huggingface-cli login") parser.add_argument('--num_blocks', type=int, default=None, help="The number of blocks to serve") parser.add_argument('--block_indices', type=str, default=None, help="Specific block indices to serve") parser.add_argument('--dht_prefix', type=str, default=None, help="Announce all blocks with this DHT prefix") parser.add_argument('--port', type=int, required=False, help='Port this server listens to. ' 'This is a simplified way to set the --host_maddrs and --announce_maddrs options (see below) ' 'that sets the port across all interfaces (IPv4, IPv6) and protocols (TCP, etc.) ' 'to the same number. Default: a random free port is chosen for each interface and protocol') parser.add_argument('--public_ip', type=str, required=False, help='Your public IPv4 address, which is visible from the Internet. ' 'This is a simplified way to set the --announce_maddrs option (see below).' 'Default: server announces IPv4/IPv6 addresses of your network interfaces') parser.add_argument("--no_auto_relay", action="store_false", dest="use_auto_relay", help="Do not look for libp2p relays to become reachable if we are behind NAT/firewall") parser.add_argument('--host_maddrs', nargs='+', required=False, help='Multiaddrs to listen for external connections from other peers') parser.add_argument('--announce_maddrs', nargs='+', required=False, help='Visible multiaddrs the host announces for external connections from other peers') parser.add_argument('--daemon_startup_timeout', type=float, default=60, help='Timeout for the libp2p daemon connecting to initial peers') parser.add_argument('--compression', type=str, default='NONE', required=False, help='Tensor compression communication') parser.add_argument('--num_handlers', type=int, default=8, required=False, help='server will use this many processes to handle incoming requests') parser.add_argument('--prefetch_batches', type=int, default=1, required=False, help='Pre-form this many subsequent batches while GPU is processing the current one') parser.add_argument('--sender_threads', type=int, default=1, required=False, help='Use this many threads to pass results/exceptions from Runtime to Pools') parser.add_argument('--inference_max_length', type=int, default=None, help='Maximum total sequence length permitted per inference, defaults to 16384 tokens. ' 'Default: 2048 for most models, 8192 for models with multi-query attention (e.g., Llama-2-70b)') parser.add_argument('--min_batch_size', type=int, default=1, help='Minimum required batch size for all operations (in total tokens)') parser.add_argument('--max_batch_size', type=int, default=None, help='The total number of tokens in the same batch will not exceed this value. ' 'Default: 2048 for most models, 8192 for models with multi-query attention (e.g., Llama-2-70b)') parser.add_argument('--max_chunk_size_bytes', type=int, default=256 * 1024 * 1024, help='Maximum size of activation tensor processed in one go; larger tensors are split into chunks') parser.add_argument('--attn_cache_tokens', type=int, default=None, help='The number of past attention key/value pairs that will be stored between inference steps. ' 'Default: 8192 for most models, 32768 for models with multi-query attention (e.g., Llama-2-70b)') parser.add_argument('--cache_dir', type=str, default=None, help='Path to a directory in which a downloaded pretrained model configuration should be cached if the standard cache should not be used.') parser.add_argument("--max_disk_space", type=str, default=None, help="Maximal disk space used for caches. Example: 50GB, 100GiB (GB != GiB here). " "Default: unlimited. " "For Agora/bloom-grid, this default means that the server may use up to " "min(free_disk_space, 350GB) in the worst case, which happens when the server runs " "for a long time and caches all model blocks after a number of rebalancings. " "However, this worst case is unlikely, expect the server to consume " "the disk space equal to 2-4x of your GPU memory on average.") parser.add_argument('--device', type=str, default=None, required=False, help='all blocks will use this device in torch notation; default: cuda if available else cpu') parser.add_argument("--torch_dtype", type=str, choices=DTYPE_MAP.keys(), default="auto", help="Use this dtype to store block weights and do computations. " "By default, respect the dtypes in the pre-trained state dict.") parser.add_argument('--alloc_timeout', type=float, default=1, help='If the cache is full, the server will wait for this number of seconds hoping that some memory will be freed ' 'before rejecting the request') parser.add_argument('--revision', type=str, default=None, help="The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a git-based system for storing models" "and other artifacts on huggingface.co, so `revision` can be any identifier allowed by git.") parser.add_argument('--throughput', type=lambda value: value if value in ['auto', 'eval'] else float(value), default='auto', help='Expected server throughput (a float measured in RPS). ' 'If set to "auto" (default), the script evaluates network and compute throughput ' 'on the first run and uses these estimates for future runs. ' 'If set to "eval", the script re-evaluates the throughput and overrides the cache.') parser.add_argument('--update_period', type=float, required=False, default=120, help='Server will report blocks to DHT once in this many seconds') parser.add_argument('--expiration', type=float, required=False, default=None, help='DHT entries will expire after this many seconds') parser.add_argument('--request_timeout', type=float, required=False, default=3 * 60, help='Timeout (in seconds) for the whole rpc_forward/rpc_backward/rpc_forward_stream/rpc_backward_stream request') parser.add_argument('--session_timeout', type=float, required=False, default=30 * 60, help='Timeout (in seconds) for the whole inference session') parser.add_argument('--step_timeout', type=float, required=False, default=5 * 60, help="Timeout (in seconds) for waiting the next step's inputs inside an inference session") group = parser.add_mutually_exclusive_group() group.add_argument('--initial_peers', type=str, nargs='*', required=False, default=PUBLIC_INITIAL_PEERS, help='Multiaddrs of one or more DHT peers from the target swarm. Default: connects to the public swarm') group.add_argument('--new_swarm', action='store_true', help='Start a new private swarm (i.e., do not connect to any initial peers)') parser.add_argument('--increase_file_limit', action='store_true', help='On *nix, this will increase the max number of processes ' 'a server can spawn before hitting "Too many open files"; Use at your own risk.') parser.add_argument('--stats_report_interval', type=int, required=False, help='Interval between two reports of batch processing performance statistics') parser.add_argument('--custom_module_path', type=str, required=False, help='Path of a file with custom nn.modules, wrapped into special decorator') parser.add_argument('--identity_path', type=str, required=False, help='Path to identity file to be used in P2P') parser.add_argument("--balance_quality", type=float, default=0.75, help="Rebalance the swarm if its throughput is worse than this share of the optimal " "throughput. Use 0.0 to disable rebalancing, values > 1.0 to force rebalancing " "on each check for debugging purposes.") parser.add_argument("--mean_balance_check_period", type=float, default=60, help="Check the swarm's balance every N seconds (and rebalance it if necessary)") parser.add_argument('--quant_type', type=str, default=None, choices=[choice.name.lower() for choice in QuantType], help="Quantize blocks to 8-bit (int8 from the LLM.int8() paper) or " "4-bit (nf4 from the QLoRA paper) formats to save GPU memory. " "Default: 'int8' if GPU is available, 'none' otherwise") parser.add_argument("--tensor_parallel_devices", nargs='+', default=None, help= "Split each block between the specified GPUs such that each device holds a portion of every " "weight matrix. See https://huggingface.co/transformers/v4.9.0/parallelism.html#tensor-parallelism") parser.add_argument("--skip_reachability_check", action='store_true', help="Skip checking this server's reachability via health.grid.dev " "when connecting to the public swarm. If you connect to a private swarm, " "the check is skipped by default. Use this option only if you know what you are doing") parser.add_argument("--adapters", nargs='+', default=(), help="List of pre-loaded LoRA adapters that can be used for inference or training") # fmt:on args = vars(parser.parse_args()) args.pop("config", None) args["converted_model_name_or_path"] = args.pop("model") or args["converted_model_name_or_path"] host_maddrs = args.pop("host_maddrs") port = args.pop("port") if port is not None: assert host_maddrs is None, "You can't use --port and --host_maddrs at the same time" else: port = 0 if host_maddrs is None: host_maddrs = [f"/ip4/0.0.0.0/tcp/{port}", f"/ip6/::/tcp/{port}"] announce_maddrs = args.pop("announce_maddrs") public_ip = args.pop("public_ip") if public_ip is not None: assert announce_maddrs is None, "You can't use --public_ip and --announce_maddrs at the same time" assert port != 0, "Please specify a fixed non-zero --port when you use --public_ip (e.g., --port 31337)" announce_maddrs = [f"/ip4/{public_ip}/tcp/{port}"] args["startup_timeout"] = args.pop("daemon_startup_timeout") if args.pop("increase_file_limit"): increase_file_limit() compression_type = args.pop("compression").upper() compression = getattr(CompressionType, compression_type) max_disk_space = args.pop("max_disk_space") if max_disk_space is not None: max_disk_space = parse_size(max_disk_space) assert isinstance( max_disk_space, (int, type(None)) ), "Unrecognized value for --max_disk_space. Correct examples: 1.5GB or 1500MB or 1572864000 (bytes)" if args.pop("new_swarm"): args["initial_peers"] = [] quant_type = args.pop("quant_type") if quant_type is not None: args["quant_type"] = QuantType[quant_type.upper()] validate_version() server = Server( **args, host_maddrs=host_maddrs, announce_maddrs=announce_maddrs, compression=compression, max_disk_space=max_disk_space, ) try: server.run() except KeyboardInterrupt: logger.info("Caught KeyboardInterrupt, shutting down") finally: server.shutdown() if __name__ == "__main__": main()
TheGrid-main
grid/cli/run_server.py
from __future__ import annotations import asyncio import itertools import time import uuid from typing import AsyncIterator, List, Optional, Tuple import torch from hivemind import ( MSGPackSerializer, anext, deserialize_torch_tensor, get_logger, nested_flatten, serialize_torch_tensor, ) from hivemind.moe.client.remote_expert_worker import RemoteExpertWorker from hivemind.p2p import P2P from hivemind.proto import runtime_pb2 from grid.client.routing.sequence_manager import RemoteSequenceManager, SequenceManagerConfig, maybe_log_traceback from grid.data_structures import CHAIN_DELIMITER, ModuleUID, RemoteSpanInfo, RPCInfo from grid.server.handler import TransformerConnectionHandler from grid.utils.misc import DUMMY, is_dummy logger = get_logger(__name__) class _ServerInferenceSession: """ An interface to a single multi-step *inference* session for a a set of blocks on a specific server. :note: This class is *not* fault-tolerant out of the box. """ def __init__( self, config: SequenceManagerConfig, span: RemoteSpanInfo, uid: ModuleUID, rpc_info: RPCInfo, inputs_queue: asyncio.Queue, outputs_aiter: AsyncIterator, *, max_length: int, **metadata, ): self.config = config self.span, self.uid, self.rpc_info = span, uid, rpc_info self.num_blocks = uid.count(CHAIN_DELIMITER) + 1 self._inputs_queue: asyncio.Queue[runtime_pb2.ExpertRequest] = inputs_queue self._outputs_stream: AsyncIterator[runtime_pb2.ExpertResponse] = outputs_aiter self.session_id = str(uuid.uuid4()) self.session_metadata = dict(max_length=max_length, **metadata) self.stepped = False self.closed = False self._position = 0 self.history = None # Used in case of server failures to regenerate attention caches on new servers self.next_session = None @classmethod async def create( cls, config: SequenceManagerConfig, p2p: P2P, span: RemoteSpanInfo, uid: ModuleUID, rpc_info: RPCInfo, **metadata, ) -> _ServerInferenceSession: """Create a new session for a given remote module. This code is meant to be run inside RemoteExpertWorker""" stub = TransformerConnectionHandler.get_stub(p2p, span.peer_id) inputs_queue = asyncio.Queue() outputs_stream = await asyncio.wait_for( stub.rpc_inference(cls._read_inputs_from_queue(inputs_queue)), config.request_timeout, ) return cls(config, span, uid, rpc_info, inputs_queue, outputs_stream, **metadata) @staticmethod async def _read_inputs_from_queue(queue: asyncio.Queue, input_timeout: Optional[float] = None) -> AsyncIterator: while True: next_input_message = await asyncio.wait_for(queue.get(), input_timeout) yield next_input_message if not next_input_message.uid and not next_input_message.tensors: break # this message means "done sending" def step( self, inputs: torch.Tensor, prompts: Optional[torch.Tensor] = None, hypo_ids: Optional[torch.Tensor] = None, *, step_id: str, ) -> torch.Tensor: """ Inference step: send a chunk of input tensors and receive a chunk of outputs :prompts: optional DEEP prompts, added to a prefix of each layer's outputs, if specified, deep prompts should have shape [num_layers, batch_size, prefix_len, hid_size] """ if self.closed: raise Exception("Session is closed, cannot perform step") n_input_tokens = inputs.shape[1] if self.history is None: self.history = inputs elif self.history.shape[1] == self._position: self.history = torch.cat([self.history, inputs[:, -n_input_tokens:]], dim=1) assert self.history.shape[1] == self._position + n_input_tokens, ( f"Broken input cache: span={self.span} shape={self.history.shape} " f"position={self._position} n_input_tokens={n_input_tokens}" ) if not self.stepped: inputs = self.history # Pass full inputs including prefix else: inputs = inputs[:, -n_input_tokens:] # No need to pass prefix further if prompts is None or is_dummy(prompts): prompts = DUMMY else: assert prompts.ndim == 4, "deep prompts should have shape [num_blocks, batch_size, prefix_len, hid_size]" assert prompts.shape[0] == self.num_blocks assert prompts.shape[1] in (inputs.shape[0], 1) assert prompts.shape[2] <= inputs.shape[1] assert prompts.shape[3] == inputs.shape[2] if hypo_ids is None or is_dummy(hypo_ids): hypo_ids = DUMMY else: assert len(hypo_ids) == len(inputs) assert hypo_ids.dtype == torch.int64 # serialize inputs and put them into the queue input_tensors = (inputs, prompts, hypo_ids) request_metadata = dict(session_id=self.session_id, step_id=step_id) if not self.stepped: request_metadata.update(self.session_metadata) elif self.config.use_server_to_server: next_servers = self._collect_next_servers() if next_servers: request_metadata["next_servers"] = next_servers outputs_serialized = RemoteExpertWorker.run_coroutine( self._step( runtime_pb2.ExpertRequest( uid=self.uid, tensors=[ serialize_torch_tensor(tensor.to(proto.dtype), proto.compression) for tensor, proto in zip(input_tensors, nested_flatten(self.rpc_info["inference_schema"])) ], metadata=MSGPackSerializer.dumps(request_metadata), ) ) ) outputs = list(map(deserialize_torch_tensor, outputs_serialized.tensors)) assert ( outputs[0].shape == inputs.shape ), f"output activation shape is different from input shape: {outputs[0].shape} != {inputs.shape}" self._position += n_input_tokens return outputs[0] def _collect_next_servers(self) -> List[Tuple[str, str, int, int]]: next_servers = [] session = self.next_session while session is not None and session.stepped: next_servers.append( (session.span.peer_id.to_base58(), session.session_id, session.span.start, session.span.end) ) session = session.next_session return next_servers async def _step(self, inputs_serialized: runtime_pb2.ExpertRequest) -> runtime_pb2.ExpertResponse: """Inference step on serialized data. This code is meant to be run inside RemoteExpertWorker""" await self._inputs_queue.put(inputs_serialized) self.stepped = True return await asyncio.wait_for(anext(self._outputs_stream), self.config.request_timeout) def close(self): """Finish a given inference session, close the underlying connection""" if self._outputs_stream is None: return # already closed RemoteExpertWorker.run_coroutine(self._aclose_stream()) self._outputs_stream = self._inputs_queue = None self.closed = True async def _aclose_stream(self): """Close the inference session. This code is meant to be run inside RemoteExpertWorker""" if self._outputs_stream is None: return # already closed if self.stepped: await self._inputs_queue.put(runtime_pb2.ExpertRequest()) # empty request will trigger end of session try: await anext(self._outputs_stream) except StopAsyncIteration: pass def __del__(self): self.close() def __enter__(self): assert not self.closed return self def __exit__(self, *exc_details): self.close() class InferenceSession: """ An interface to a multi-step *inference* session for a sequence of remote transformer blocks """ def __init__(self, sequence_manager: RemoteSequenceManager, max_length: int): self._sequence_manager = sequence_manager self._closed = False self._server_sessions = [] self._position = 0 self._max_length = max_length self.last_token_id = None @property def num_blocks(self) -> int: return len(self._sequence_manager) @property def position(self) -> int: return self._position def _enter_server_sessions(self, chosen_spans: List[RemoteSpanInfo]) -> List[_ServerInferenceSession]: server_sessions = [] try: for span in chosen_spans: span_uids = CHAIN_DELIMITER.join(self._sequence_manager.block_uids[span.start : span.end]) metadata = self._sequence_manager.get_request_metadata("rpc_inference", span_uids, peer_id=span.peer_id) session = RemoteExpertWorker.run_coroutine( _ServerInferenceSession.create( self._sequence_manager.config, self._sequence_manager.state.p2p, span, span_uids, rpc_info=self._sequence_manager.rpc_info, max_length=self._max_length, **metadata, ) ) server_sessions.append(session) session.__enter__() return server_sessions except: self._exit_server_sessions(server_sessions) raise def _exit_server_sessions(self, server_sessions: List[_ServerInferenceSession]) -> None: for session in reversed(server_sessions): try: session.__exit__(None, None, None) except Exception: logger.debug("Caught exception while closing connection to server:", exc_info=True) def __enter__(self) -> "InferenceSession": assert not self._closed and not self._server_sessions return self def step(self, inputs: torch.Tensor, prompts: Optional[torch.Tensor] = None, **kwargs) -> torch.Tensor: assert not self._closed if torch.is_grad_enabled(): logger.warning("Running inference session with grad enabled. Gradients will *not* be propagated correctly.") if prompts is None or is_dummy(prompts): prompts = DUMMY else: assert prompts.ndim == 4, "deep prompts should have shape [num_blocks, batch_size, prefix_len, hid_size]" assert prompts.shape[0] == self.num_blocks inputs_device = inputs.device inputs_dtype = inputs.dtype inputs = inputs.cpu() prompts = prompts.cpu() step_id = str(uuid.uuid4()) n_input_tokens = inputs.shape[1] if self._position + n_input_tokens > self._max_length: raise ValueError( f"Maximum length exceeded: prefix {self._position} + current {n_input_tokens} exceeds pre-allocated maximum {self._max_length}" ) server_idx = 0 block_idx = 0 while block_idx < self.num_blocks: for attempt_no in itertools.count(): logger.debug(f"Inference: block {block_idx}, attempt {attempt_no}") server_session = None try: if not self._server_sessions or attempt_no >= 1: self._update_sequence(server_idx, block_idx, attempt_no) server_session = self._server_sessions[server_idx] inputs = server_session.step( inputs, prompts[server_session.span.start : server_session.span.end], step_id=step_id, **kwargs ) server_idx += 1 block_idx = server_session.span.end self._sequence_manager.on_request_success(server_session.span.peer_id) break except Exception as e: self._sequence_manager.on_request_failure( server_session.span.peer_id if server_session is not None else None ) if attempt_no + 1 == self._sequence_manager.config.max_retries: raise delay = self._sequence_manager.get_retry_delay(attempt_no) logger.warning( f"Caught exception when running inference via {server_session.span if server_session is not None else None} " f"(retry in {delay:.0f} sec): {repr(e)}" ) maybe_log_traceback(e) time.sleep(delay) self._position += n_input_tokens outputs = inputs[:, -n_input_tokens:] outputs = outputs.to(device=inputs_device, dtype=inputs_dtype) return outputs def _update_sequence(self, server_idx: int, block_idx: int, attempt_no: int) -> int: # If there is a failed server session, this code closes it self._exit_server_sessions(self._server_sessions[server_idx : server_idx + 1]) n_prev_spans = len(self._server_sessions) update_end = self._server_sessions[server_idx].span.end if server_idx < n_prev_spans else self.num_blocks if attempt_no >= 1: logger.info( f"Due to a server failure, remote attention caches " f"from block {block_idx} to {update_end} will be regenerated" ) updated_spans = self._sequence_manager.make_sequence( block_idx, update_end, mode="min_latency", cache_tokens_needed=self._max_length ) # make_sequence() could return a longer sequence updated_spans[-1].end = min(updated_spans[-1].end, update_end) updated_sessions = self._enter_server_sessions(updated_spans) logger.debug(f"Found path from block {block_idx} to {update_end} via {len(updated_spans)} servers") # If there is a failed span, this code replaces it, otherwise it just adds new ones if server_idx < n_prev_spans: updated_sessions[0].history = self._server_sessions[server_idx].history self._server_sessions[server_idx : server_idx + 1] = updated_sessions # Update links to the next server session for direct server-to-server communication via rpc_push() for i in range(max(server_idx - 1, 0), min(server_idx + len(updated_spans), len(self._server_sessions) - 1)): self._server_sessions[i].next_session = self._server_sessions[i + 1] def close(self, *exc_details): """Finish a given inference session, close the underlying connection""" if not self._closed: self._exit_server_sessions(self._server_sessions) self._server_sessions.clear() self._closed = True def __exit__(self, *exc_details): self.close(*exc_details) def __del__(self): self.close()
TheGrid-main
grid/client/inference_session.py
from __future__ import annotations from typing import Optional, Union import torch from hivemind import DHT, get_logger from torch import nn from grid.client.inference_session import InferenceSession from grid.client.routing.sequence_manager import RemoteSequenceManager, SequenceManagerConfig from grid.client.sequential_autograd import _RemoteSequentialAutogradFunction from grid.data_structures import UID_DELIMITER from grid.utils.misc import DUMMY logger = get_logger(__name__) class RemoteSequential(nn.Module): """ A sequence of transformer blocks hosted by the swarm. """ def __init__( self, config: SequenceManagerConfig, *, sequence_manager: Optional[RemoteSequenceManager] = None, dht: Optional[DHT] = None, start_block: Optional[int] = None, end_block: Optional[int] = None, **kwargs, ): super().__init__() self.config = config assert sequence_manager is None or ( dht is None and start_block is None and end_block is None ), "`dht`, `start_block`, and `end_block` have no effect when you provide a custom `sequence_manager`" if sequence_manager is None: if start_block is None: start_block = 0 if end_block is None: end_block = self.config.num_hidden_layers block_uids = tuple(f"{config.dht_prefix}{UID_DELIMITER}{i}" for i in range(start_block, end_block)) sequence_manager = RemoteSequenceManager(config, block_uids, dht=dht, **kwargs) self.sequence_manager = sequence_manager def forward(self, inputs: torch.Tensor, prompts: torch.Tensor = DUMMY): assert inputs.ndim == 3, "inputs must be a tensor of shape [batch_size, seq_length, hidden_size]" assert inputs.shape[1] <= 2048, "The sequence length is capped at 2048 tokens in this version" outputs = _RemoteSequentialAutogradFunction.apply(inputs, prompts, self.sequence_manager) return outputs def __getitem__(self, ix: Union[int, slice]) -> RemoteSequential: return RemoteSequential( self.config, sequence_manager=self.sequence_manager[ix], ) def __iter__(self): for block_index in range(len(self)): yield self[block_index] def __len__(self): return len(self.sequence_manager) def inference_session(self, **kwargs) -> InferenceSession: return InferenceSession(self.sequence_manager, **kwargs) def extra_repr(self) -> str: return f"modules={self.sequence_manager.block_uids[0]}..{self.sequence_manager.block_uids[-1]}"
TheGrid-main
grid/client/remote_sequential.py
TheGrid-main
grid/client/__init__.py
""" Utility functions that call RPC forward or backward on a single remote server """ import asyncio from typing import Iterable, List, Optional, Sequence, Tuple import torch from hivemind import nested_compare, nested_flatten, nested_pack, serialize_torch_tensor from hivemind.compression.serialization import deserialize_tensor_stream, deserialize_torch_tensor from hivemind.p2p import StubBase from hivemind.p2p.p2p_daemon_bindings.control import DEFAULT_MAX_MSG_SIZE, MAX_UNARY_PAYLOAD_SIZE from hivemind.proto import runtime_pb2 from hivemind.utils.asyncio import aiter_with_timeout, iter_as_aiter from hivemind.utils.streaming import split_for_streaming from grid.data_structures import ModuleUID, RPCInfo async def _forward_unary( uid: str, serialized_tensors: Iterable[runtime_pb2.Tensor], stub, timeout: float, **kwargs ) -> List[torch.Tensor]: outputs: runtime_pb2.ExpertResponse = await stub.rpc_forward( runtime_pb2.ExpertRequest(uid=uid, tensors=list(serialized_tensors), **kwargs), timeout=timeout, ) return [deserialize_torch_tensor(t) for t in outputs.tensors] async def _backward_unary( uid: str, serialized_tensors: Iterable[runtime_pb2.Tensor], stub, timeout: float, **kwargs ) -> List[torch.Tensor]: grad_inputs: runtime_pb2.ExpertResponse = await stub.rpc_backward( runtime_pb2.ExpertRequest(uid=uid, tensors=list(serialized_tensors), **kwargs), timeout=timeout, ) return [deserialize_torch_tensor(t) for t in grad_inputs.tensors] async def _forward_stream( uid: str, serialized_tensors: Iterable[runtime_pb2.Tensor], stub, timeout: float, **kwargs ) -> List[torch.Tensor]: parts = ( runtime_pb2.ExpertRequest(uid=uid, tensors=[part], **kwargs) for tensor in serialized_tensors for part in split_for_streaming(tensor, DEFAULT_MAX_MSG_SIZE) ) outputs = await asyncio.wait_for(stub.rpc_forward_stream(iter_as_aiter(parts)), timeout) outputs = aiter_with_timeout(outputs, timeout) return await deserialize_tensor_stream(msg.tensors async for msg in outputs) async def _backward_stream( uid: str, serialized_tensors: Iterable[runtime_pb2.Tensor], stub, timeout: float, **kwargs ) -> List[torch.Tensor]: parts = ( runtime_pb2.ExpertRequest(uid=uid, tensors=[part], **kwargs) for tensor in serialized_tensors for part in split_for_streaming(tensor, DEFAULT_MAX_MSG_SIZE) ) grad_inputs = await asyncio.wait_for(stub.rpc_backward_stream(iter_as_aiter(parts)), timeout) grad_inputs = aiter_with_timeout(grad_inputs, timeout) return await deserialize_tensor_stream(msg.tensors async for msg in grad_inputs) async def run_remote_forward( uid: ModuleUID, stub: StubBase, rpc_info: RPCInfo, *inputs: torch.Tensor, timeout: float, metadata: Optional[bytes] = None, **kwargs, ) -> Tuple[torch.Tensor, ...]: """ Serializes input tensors and calls "rpc_forward" on a remote server. Mostly adapted from https://github.com/learning-at-home/hivemind/blob/7a7c93aefffc9494c39e7b170c07cb06d8c09c4c/hivemind/moe/client/expert.py#L198 but without RemoteExpertWorker.run_coroutine() call that leads to deadlock here. """ # Note: *inputs are flattened input tensors that follow the expert's info['input_schema'] # detach to avoid pickling the computation graph assert len(kwargs) == len(rpc_info["keyword_names"]), f"Keyword args should be {rpc_info['keyword_names']}" kwargs = {key: kwargs[key] for key in rpc_info["keyword_names"]} # Note: we put keyword arguments in the same order as on a server to prevent f(a=1, b=2) != f(b=2, a=1) errors forward_inputs = (inputs, kwargs) # Modify forward_schema to support prompts args_schema, kwargs_schema = rpc_info["forward_schema"] # TODO: rm this assert when support arbitrary number of input tensors assert len(args_schema) == 1 and len(inputs) == 2 forward_schema_with_prompts = (tuple(args_schema * len(inputs)), kwargs_schema) if not nested_compare(forward_inputs, forward_schema_with_prompts): raise TypeError("Inputs do not match expert input schema. Did you pass the right number of parameters?") forward_inputs = nested_flatten(forward_inputs) inputs = tuple(tensor.cpu().detach() for tensor in forward_inputs) # Asynchronous serialization loop = asyncio.get_running_loop() serialized_tensors = await asyncio.gather( *( loop.run_in_executor(None, serialize_torch_tensor, tensor.to(proto.dtype), proto.compression) for tensor, proto in zip(inputs, nested_flatten(forward_schema_with_prompts)) ) ) # call RPC on remote server size = sum(t.element_size() * t.nelement() for t in inputs) forward_fn = _forward_stream if size > MAX_UNARY_PAYLOAD_SIZE // 2 else _forward_unary # Hotfix: we use "// 2" since hivemind==1.1.5 serializes bfloat16 tensors in float32, so they take 2x more space deserialized_outputs = await forward_fn(uid, serialized_tensors, stub, timeout, metadata=metadata, **kwargs) return nested_pack(deserialized_outputs, structure=rpc_info["outputs_schema"]) async def run_remote_backward( uid: ModuleUID, stub: StubBase, rpc_info: RPCInfo, inputs: torch.Tensor, grad_outputs: List[torch.Tensor], *extra_tensors: torch.Tensor, timeout: float, metadata: Optional[bytes] = None, **kwargs, ) -> Sequence[torch.Tensor]: """ Serializes grad outputs and calls "rpc_backward" on a remote server. Mostly adapted from https://github.com/learning-at-home/hivemind/blob/7a7c93aefffc9494c39e7b170c07cb06d8c09c4c/hivemind/moe/client/expert.py#L221 but without RemoteExpertWorker.run_coroutine() call that leads to deadlock here. """ grad_outputs_cpu = tuple(tensor.cpu() for tensor in grad_outputs) inputs_and_grad_outputs = tuple(nested_flatten((inputs, grad_outputs_cpu, *extra_tensors))) # Modify forward_schema to support prompts args_schema, kwargs_schema = rpc_info["forward_schema"] assert len(args_schema) == 1 and isinstance(inputs, torch.Tensor) # TODO generalize this prompts_schema = next(iter(args_schema)) backward_schema = tuple(nested_flatten((rpc_info["forward_schema"], rpc_info["outputs_schema"], prompts_schema))) # Asynchronous serialization loop = asyncio.get_running_loop() serialized_tensors = await asyncio.gather( *( loop.run_in_executor(None, serialize_torch_tensor, tensor.to(proto.dtype), proto.compression) for tensor, proto in zip(inputs_and_grad_outputs, backward_schema) ) ) size = sum(t.element_size() * t.nelement() for t in inputs_and_grad_outputs) backward_fn = _backward_stream if size > MAX_UNARY_PAYLOAD_SIZE // 2 else _backward_unary # Hotfix: we use "// 2" since hivemind==1.1.5 serializes bfloat16 tensors in float32, so they take 2x more space deserialized_grad_inputs = await backward_fn(uid, serialized_tensors, stub, timeout, metadata=metadata, **kwargs) return deserialized_grad_inputs
TheGrid-main
grid/client/remote_forward_backward.py
import dataclasses from contextlib import contextmanager from typing import Optional import torch import torch.nn as nn from hivemind import get_logger from transformers import PretrainedConfig from grid.utils.misc import DUMMY logger = get_logger(__name__) @dataclasses.dataclass class PTuneConfig: pre_seq_len: int = 0 # a number of tokens for prompt tuning. tuning_mode: Optional[str] = None # fine-tuning regime, one of [None, "ptune", "deep_ptune"] class PTuneMixin: _keys_to_ignore_on_load_missing = [r"(intermediate_)?prompt_embeddings\.weight$"] def init_prompts(self, config: PretrainedConfig) -> None: if config.tuning_mode and "ptune" in config.tuning_mode: assert config.pre_seq_len > 0, "The number of prefix tokens must be > 0" self.pre_seq_len = config.pre_seq_len self.prefix_tokens = torch.arange(self.pre_seq_len).long() with force_non_empty_weights(): # Prompt embeddings and their optimizer stats are kept in float32 to increase ptune quality self.prompt_embeddings = nn.Embedding(self.pre_seq_len, config.hidden_size, dtype=torch.float32) if config.tuning_mode == "deep_ptune": self.intermediate_prompt_embeddings = nn.Embedding( self.pre_seq_len, config.num_hidden_layers * config.hidden_size, # ^-- TODO: should be num_hidden_layers - 1 dtype=torch.float32, ) elif config.tuning_mode: raise NotImplementedError(f"{self.tuning_mode} mode is not supported for now") def get_prompt(self, batch_size): prefix_tokens = self.prefix_tokens.unsqueeze(0).expand(batch_size, -1) prefix_tokens = prefix_tokens.to(self.word_embeddings.weight.device) prompts = self.prompt_embeddings(prefix_tokens) if self.config.tuning_mode == "deep_ptune": intermediate_prompts = self.intermediate_prompt_embeddings(prefix_tokens) intermediate_prompts = intermediate_prompts.view( batch_size, self.pre_seq_len, self.config.num_hidden_layers, self.config.hidden_size # TODO: should be num_hidden_layers - 1 ) intermediate_prompts = intermediate_prompts.permute([2, 0, 1, 3]) else: intermediate_prompts = DUMMY dtype = self.word_embeddings.weight.dtype return prompts.to(dtype), intermediate_prompts.to(dtype) _original_register_parameter = nn.Module.register_parameter @contextmanager def force_non_empty_weights(): """ This context manager allows to bypass the accelerate.init_empty_weights() context manager (that forces all nn.Parameters to be PyTorch's meta tensors) used when low_cpu_mem_usage=True. The transformers library should replace all meta tensors by empty tensors by itself but this feature does not work due to a bug ([1] fails if `add_prefix_to_model == True`). [1] https://github.com/huggingface/transformers/blob/ab9fe45236cd99b8797df78219438f8f6662bb42/src/transformers/modeling_utils.py#L2515 """ try: possibly_patched_register_parameter = nn.Module.register_parameter nn.Module.register_parameter = _original_register_parameter yield finally: nn.Module.register_parameter = possibly_patched_register_parameter
TheGrid-main
grid/client/ptune.py
import dataclasses import platform from typing import Union import psutil import torch import torch.nn.functional as F import torch.utils.checkpoint from hivemind import get_logger from torch import nn from transformers import PretrainedConfig logger = get_logger(__name__) @dataclasses.dataclass class LMHeadConfig: # This settings matter for running the client with dtype bfloat16 on CPU. # If the CPU doesn't support AVX512, chunked_forward() significantly speeds up computations. use_chunked_forward: Union[str, bool] = "auto" chunked_forward_step: int = 16384 class LMHead(nn.Module): def __init__(self, config: PretrainedConfig): super().__init__() if not config.tie_word_embeddings: self.weight = nn.Parameter(torch.zeros(config.vocab_size, config.hidden_size)) self.weight.requires_grad = False else: self.weight = None # Will be set to get_input_embeddings().weight during loading the model self.bias = None self.in_features = config.hidden_size # Similar to nn.Linear attributes self.out_features = config.vocab_size self.use_chunked_forward = config.use_chunked_forward if self.use_chunked_forward == "auto": if platform.machine() == "x86_64": # Import of cpufeature may crash on non-x86_64 machines from cpufeature import CPUFeature # If the CPU supports AVX512, plain bfloat16 is ~10x faster than chunked_forward(). # Otherwise, it's ~8x slower. self.use_chunked_forward = not (CPUFeature["AVX512f"] and CPUFeature["OS_AVX512"]) else: self.use_chunked_forward = True self.chunked_forward_step = config.chunked_forward_step self._bf16_warning_shown = False def forward(self, hidden_states): if ( self.weight.dtype in [torch.float16, torch.bfloat16] and self.weight.device.type == "cpu" and self.use_chunked_forward ): lm_logits = self.chunked_forward(hidden_states) else: # Switch dtype in case word_embeddings are fp16/bf16 hidden_states = hidden_states.to(self.weight.dtype) lm_logits = F.linear(hidden_states, self.weight) return lm_logits def chunked_forward(self, hidden_states): """Splits word embeddings on chunks and iteratively casts them into fp32 to perform matmul more efficiently on CPU. chunked_forward_step: provides trade-off between efficiency and extra memory consumption. """ assert self.chunked_forward_step > 0, "Chunk size for chunked forward must be positive" if not self._bf16_warning_shown: if self.weight.numel() * 4 < 0.9 * psutil.virtual_memory().total: logger.warning( "Running the client with dtype bfloat16 on CPU may be slow, since your CPU doesn't support AVX512. " "Consider loading the model with torch_dtype='float32'" ) self._bf16_warning_shown = True hidden_states = hidden_states.float() output = torch.empty(*hidden_states.shape[:-1], self.out_features) for i in range(0, self.out_features, self.chunked_forward_step): chunk = self.weight[i : i + self.chunked_forward_step].float() output[..., i : i + self.chunked_forward_step] = F.linear(hidden_states, chunk) return output
TheGrid-main
grid/client/lm_head.py
import contextlib from typing import List, Optional import torch from hivemind.utils.logging import get_logger from grid.client.inference_session import InferenceSession from grid.utils.generation_algorithms import ( BeamSearchAlgorithm, DecodingAlgorithm, GreedyAlgorithm, NucleusAlgorithm, SamplingAlgorithm, TopKAlgorithm, ) from grid.utils.generation_constraints import ABCBloomConstraint, EosConstraint logger = get_logger(__name__) class RemoteGenerationMixin: """ A class containing all functions for auto-regressive text generation, to be used as a mixin in [`BloomForCausalLM`]. The class exposes can be used for: - *greedy decoding*. - *multinomial, top-k and top-p sampling*. - *beam-search decoding* This class is similar to transformer's [`generation_utils.GenerationMixin`], it can be used instead of it. However, it has some differences for remote usage. """ def inference_session(self, **kwargs) -> InferenceSession: """ Returns an inference session for the model's RemoteSequential module. :param max_length: Maximal expected length of inference results. Servers use this parameter to calculate the size of attention caches allocated to this client. """ return self.transformer.h.inference_session(**kwargs) @torch.inference_mode() def generate( self, inputs: Optional[torch.Tensor] = None, *, do_sample: Optional[bool] = None, temperature: float = 1.0, top_k: Optional[int] = None, top_p: Optional[float] = None, num_beams: Optional[int] = 1, bos_token_id: Optional[int] = None, eos_token_id: Optional[int] = None, pad_token_id: Optional[int] = None, max_length: Optional[int] = None, max_new_tokens: Optional[int] = None, decoding_algorithm: Optional[DecodingAlgorithm] = None, provided_constraints: List[ABCBloomConstraint] = [], num_return_sequences: Optional[int] = None, session: Optional[InferenceSession] = None, ) -> torch.LongTensor: """ Generates sequences of token ids for models with a language modeling head. :param inputs: The input tokens to the model. :param do_sample: Whether to sample from the model predictions or take the argmax. :param temperature: The temperature to use for sampling. :param top_k: The number of results to return. :param top_p: The cumulative probability of results to return. :param num_beams: The number of beams to use for beam search. :param bos_token_id: The id of the beginning of sentence token. :param eos_token_id: The id of the end of sentence token. :param pad_token_id: The id of the padding token. :param max_length: The maximum number of tokens in the output (including input tokens). :param max_new_tokens: The maximum number of tokens to generate. :param decoding_algorithm: The decoding algorithm to use. :param provided_constraints: A list of constraints to use. :param num_return_sequences: How many hypothesis from the beam will be in output. """ prefix_length = 0 if inputs is None else inputs.size(1) prefix_length += self.config.pre_seq_len bos_token_id = bos_token_id if bos_token_id is not None else self.config.bos_token_id pad_token_id = pad_token_id if pad_token_id is not None else self.config.pad_token_id eos_token_id = eos_token_id if eos_token_id is not None else self.config.eos_token_id assert (max_length is None) != (max_new_tokens is None), "please set max_length or max_new_tokens (not both)" if max_length is not None and max_new_tokens is None: max_new_tokens = max_length - prefix_length assert max_new_tokens > 0, f"Provided max_length is less than prefix size: {max_length} < {inputs.size(1)}" elif max_length is None and max_new_tokens is not None: max_length = prefix_length + max_new_tokens resuming_session = session is not None and session.last_token_id is not None if num_beams > 1 and resuming_session: raise NotImplementedError( "Resuming inference session in .generate() along with beam search is not supported yet" ) if inputs is not None: assert isinstance(inputs, torch.Tensor) and inputs.ndim == 2, "inputs must be a 2d tensor [batch, length]" if resuming_session: inputs = torch.cat([session.last_token_id, inputs], dim=1) else: if resuming_session: inputs = session.last_token_id else: assert bos_token_id is not None, "You have to provide a bos_token_id if you do not provide inputs" inputs = torch.tensor([[bos_token_id]] * num_beams, dtype=torch.long, device=self.device) batch_size = inputs.size(0) if decoding_algorithm is None: if do_sample: decoding_algorithm = self._choose_sample_algorithm(temperature, top_k, top_p) elif num_beams is not None and num_beams > 1: decoding_algorithm = BeamSearchAlgorithm(num_beams, batch_size=batch_size) else: if top_k is not None or top_p is not None: logger.warning("You passed top_k or top_p but did not pass do_sample=True. Running greedy sampling") decoding_algorithm = GreedyAlgorithm() if num_beams > 1: inputs = torch.cat([inputs] * num_beams, dim=0) if batch_size > 1: # TODO: resolve padding problem logger.warning( f"You set batch_size {batch_size} within beam search generation. " f"Be careful, results on sequences with different length may be padded wrong way" ) if num_return_sequences is None: num_return_sequences = 1 assert num_return_sequences <= num_beams, ( "You want more sequences than the beam has." " Check num_return_sequences: {num_return_sequences} and num_beams: {num_beams}." ) constraints = self._get_constraints( inputs=inputs, eos_token_id=eos_token_id, pad_token_id=pad_token_id, provided_constraints=provided_constraints, ) if session is None: context_manager = self.inference_session(max_length=max_length) else: context_manager = contextlib.nullcontext(session) # Doesn't actually enter session or exit from it with context_manager as session: outputs = [] # Find samples with padded inputs. # They will be changed before all of the samples have right length. if torch.any(inputs == pad_token_id): # TODO: move to prepare_inputs outputs += [inputs[:, : inputs.size(1) - (inputs == pad_token_id).sum(-1).max()]] else: outputs += [inputs] last_token_id = None seq_idx = outputs[0].size(1) hypo_ids = torch.arange(outputs[0].size(0)) while True: hidden_state = self.transformer.word_embeddings(outputs[-1]) intermediate_prompts = None if self.config.pre_seq_len > 0 and len(outputs) == 1: prompts, intermediate_prompts = self.transformer.get_prompt(hidden_state.size(0)) hidden_state = torch.cat([prompts, hidden_state], dim=1) hidden_state = self.transformer.word_embeddings_layernorm(hidden_state) hidden_state = session.step(hidden_state, prompts=intermediate_prompts, hypo_ids=hypo_ids)[:, -1] hidden_state = self.transformer.ln_f(hidden_state) lm_logits = self.lm_head(hidden_state) for constraint in constraints: lm_logits = constraint(last_token_id, lm_logits, hypo_ids) last_token_id, hypo_ids = decoding_algorithm(lm_logits) # If some samples were padded, change only these samples if seq_idx < inputs.size(1): pad_token_mask = inputs[:, seq_idx : seq_idx + 1] == pad_token_id last_token_id = (~pad_token_mask) * inputs[ :, seq_idx : seq_idx + 1 ] + pad_token_mask * last_token_id # TODO: refactor outputs if num_beams > 1: for i in range(len(outputs), 1, -1): outputs[i - 1] = outputs[i - 1][hypo_ids] outputs.append(last_token_id) session.last_token_id = last_token_id seq_idx += 1 if torch.all(last_token_id == eos_token_id) or len(outputs) > max_new_tokens: break outputs = torch.cat(outputs, dim=-1) if resuming_session: outputs = outputs[:, 1:] if num_beams > 1: pre_return_idx = [ torch.arange(idx, num_return_sequences * batch_size, batch_size) for idx in range(batch_size) ] return_idx = torch.cat(pre_return_idx, dim=0) outputs = outputs[return_idx] return outputs def greedy_search( self, input_ids: torch.LongTensor, max_length: Optional[int] = None, pad_token_id: Optional[int] = None, eos_token_id: Optional[int] = None, provided_constraints: List[ABCBloomConstraint] = [], ) -> torch.LongTensor: """ Generates sequences of token ids for models with a language modeling head. Uses greedy search. :param input_ids: The input tokens to the model. :param max_length: The maximum length of the sequence to generate. :param pad_token_id: The id of the padding token. :param eos_token_id: The id of the end of sentence token. :param provided_constraints: A list of constraints to use. """ return self.generate( inputs=input_ids, max_new_tokens=max_length, pad_token_id=pad_token_id, eos_token_id=eos_token_id, decoding_algorithm=GreedyAlgorithm(), provided_constraints=provided_constraints, ) def sample( self, input_ids: torch.LongTensor, temperature: float = 1.0, top_k: Optional[int] = None, top_p: Optional[float] = None, max_length: Optional[int] = None, pad_token_id: Optional[int] = None, eos_token_id: Optional[int] = None, provided_constraints: List[ABCBloomConstraint] = [], ) -> torch.LongTensor: """ Generates sequences of token ids for models with a language modeling head. Uses multinomial sampling. If top_k is provided, uses top_k sampling. If top_p is provided, uses nucleus sampling. :param: input_ids: The input tokens to the model. :param: temperature: The temperature to use for sampling. :param: top_k: The number of samples to use for top_k sampling. :param: top_p: The probability of using top_p sampling. :param: max_length: The maximum length of the sequence to generate. :param: pad_token_id: The id of the padding token. :param: eos_token_id: The id of the end of sentence token. :param: provided_constraints: A list of constraints to use. """ return self.generate( inputs=input_ids, max_new_tokens=max_length, pad_token_id=pad_token_id, eos_token_id=eos_token_id, decoding_algorithm=self._choose_sample_algorithm(temperature, top_k, top_p), provided_constraints=provided_constraints, ) def beam_search( self, input_ids: torch.LongTensor, num_beams: int = 1, max_length: Optional[int] = None, pad_token_id: Optional[int] = None, eos_token_id: Optional[int] = None, provided_constraints: List[ABCBloomConstraint] = [], ) -> torch.LongTensor: """ Generates sequences of token ids for models with a language modeling head. Uses beam search. :param input_ids: The input tokens to the model. :param num_beams: The number of beams to use. :param max_length: The maximum length of the sequence to generate. :param pad_token_id: The id of the padding token. :param eos_token_id: The id of the end of sentence token. :param provided_constraints: A list of constraints to use. """ decoding_algorithm = BeamSearchAlgorithm( num_beams=num_beams, batch_size=input_ids.size(0), ) return self.generate( inputs=input_ids, num_beams=num_beams, max_new_tokens=max_length, pad_token_id=pad_token_id, eos_token_id=eos_token_id, decoding_algorithm=decoding_algorithm, provided_constraints=provided_constraints, ) def beam_sample( self, input_ids: torch.LongTensor, max_length: Optional[int] = None, pad_token_id: Optional[int] = None, eos_token_id: Optional[int] = None, provided_constraints: List[ABCBloomConstraint] = [], ) -> torch.LongTensor: raise NotImplementedError def group_beam_search( self, input_ids: torch.LongTensor, max_length: Optional[int] = None, pad_token_id: Optional[int] = None, eos_token_id: Optional[int] = None, provided_constraints: List[ABCBloomConstraint] = [], ) -> torch.LongTensor: raise NotImplementedError def _choose_sample_algorithm( self, temperature: float = 1.0, top_k: Optional[int] = None, top_p: Optional[float] = None, ) -> DecodingAlgorithm: if (top_k is not None) and (top_p is not None): raise ValueError("You have to provide only top_k or top_p for sampling") if top_k is not None: return TopKAlgorithm(top_k, temperature) elif top_p is not None: return NucleusAlgorithm(top_p, temperature) else: return SamplingAlgorithm(temperature) def _get_constraints( self, inputs: Optional[torch.Tensor] = None, eos_token_id: Optional[int] = None, pad_token_id: Optional[int] = None, provided_constraints: List[ABCBloomConstraint] = [], ) -> List[ABCBloomConstraint]: constraints = [] constraints.extend(provided_constraints) constraints.append(EosConstraint(inputs, eos_token_id, pad_token_id)) return constraints
TheGrid-main
grid/client/remote_generation.py
""" A PyTorch autograd function that runs forward/backward on a sequence of remote servers in a fault-tolerant manner """ import asyncio import itertools from collections import deque from typing import List, Optional, Sequence, Tuple import torch from hivemind import MSGPackSerializer from hivemind.moe.client.remote_expert_worker import RemoteExpertWorker from hivemind.utils.logging import get_logger from grid.client.remote_forward_backward import run_remote_backward, run_remote_forward from grid.client.routing.sequence_manager import RemoteSequenceManager, maybe_log_traceback from grid.data_structures import CHAIN_DELIMITER, RemoteSpanInfo from grid.server.handler import TransformerConnectionHandler from grid.utils.misc import DUMMY, is_dummy logger = get_logger(__name__) MAX_TOKENS_IN_BATCH = 1024 async def sequential_forward( inputs: torch.Tensor, prompts: torch.Tensor, sequence_manager: RemoteSequenceManager, start_index: int = 0, end_index: Optional[int] = None, ) -> Tuple[torch.Tensor, Sequence[torch.Tensor], Sequence[RemoteSpanInfo]]: """ Constructs a routing path from <start_index> to <end_index>. Performs chained forward for each subsequence of blocks on the path. If some subsequence fails, reconstructs the remaining path and tries to finish the forward. """ assert isinstance(inputs, torch.Tensor) and inputs.ndim == 3, f"{type(inputs)}: {inputs.ndim}" inputs_device = inputs.device inputs_dtype = inputs.dtype inputs = inputs.cpu() prompts = prompts.cpu() end_index = end_index if end_index is not None else len(sequence_manager.block_uids) assert start_index >= 0 and end_index <= len(sequence_manager.block_uids) assert is_dummy(prompts) or len(prompts) == len( sequence_manager.block_uids ) # should be n_layers - 1 but add extra prompts for convenience sequences = deque() intermediate_inputs = [] done_sequences = [] block_idx = start_index while block_idx < end_index: for attempt_no in itertools.count(): logger.debug(f"Forward: block {block_idx}, attempt {attempt_no}") span = None try: if not sequences or attempt_no >= 1: sequences = deque(sequence_manager.make_sequence(block_idx, end_index, mode="max_throughput")) # make_sequence() could return a longer sequence sequences[-1].end = min(sequences[-1].end, end_index) logger.debug(f"Found path from block {block_idx} to {end_index} via {len(sequences)} servers") span = sequences.popleft() stub = TransformerConnectionHandler.get_stub(sequence_manager.state.p2p, span.peer_id) inputs_and_prompts = [inputs, prompts[span.start : span.end]] span_uids = CHAIN_DELIMITER.join(sequence_manager.block_uids[span.start : span.end]) metadata = sequence_manager.get_request_metadata("rpc_forward", span_uids, *inputs_and_prompts) (outputs,) = await run_remote_forward( span_uids, stub, sequence_manager.rpc_info, *inputs_and_prompts, timeout=sequence_manager.config.request_timeout, metadata=MSGPackSerializer.dumps(metadata), ) assert isinstance(outputs, torch.Tensor) assert outputs.shape == inputs.shape, f"Expected output {inputs.shape}, got {outputs.shape}" # Save intermediate inputs and subsequences if the forward is already done for them intermediate_inputs.append(inputs) done_sequences.append(span) inputs = outputs block_idx = span.end sequence_manager.on_request_success(span.peer_id) break except Exception as e: sequence_manager.on_request_failure(span.peer_id if span is not None else None) if attempt_no + 1 == sequence_manager.config.max_retries: raise delay = sequence_manager.get_retry_delay(attempt_no) logger.warning( f"Caught exception when running forward via {span} (retry in {delay:.0f} sec): {repr(e)}" ) maybe_log_traceback(e) await asyncio.sleep(delay) outputs = inputs.to(device=inputs_device, dtype=inputs_dtype) intermediate_inputs = [tensor.to(device=inputs_device, dtype=inputs_dtype) for tensor in intermediate_inputs] return outputs, intermediate_inputs, done_sequences async def sequential_backward( grad_outputs: Sequence[torch.Tensor], intermediate_inputs: List[torch.Tensor], prompts: torch.Tensor, forward_sequences: List[RemoteSpanInfo], sequence_manager: RemoteSequenceManager, ) -> Tuple[Sequence[torch.Tensor], torch.Tensor]: """ Performs chained backward for each forward subsequence. If some subsequence fails, reconstructs the particular sub-path and recovers the backward. """ assert len(intermediate_inputs) == len(forward_sequences) grad_outputs_device = grad_outputs[0].device if grad_outputs else None grad_outputs_dtype = grad_outputs[0].dtype if grad_outputs else None prompts_device = prompts.device prompts_dtype = prompts.dtype grad_outputs = [tensor.cpu() for tensor in grad_outputs] intermediate_inputs = [tensor.cpu() for tensor in intermediate_inputs] prompts = prompts.cpu() grad_prompts_reversed = [] while len(forward_sequences) > 0 and len(intermediate_inputs) > 0: inputs = intermediate_inputs.pop() span = forward_sequences.pop() for attempt_no in itertools.count(): logger.debug(f"Backward: block {span.end - 1}, attempt {attempt_no}") try: if attempt_no >= 1: _, backup_inputs, backup_sequences = await sequential_forward( inputs, prompts, sequence_manager, start_index=span.start, end_index=span.end ) assert len(backup_inputs) == len(backup_sequences) assert backup_sequences[0].start == span.start assert backup_sequences[-1].end == span.end intermediate_inputs.extend(backup_inputs) forward_sequences.extend(backup_sequences) inputs = intermediate_inputs.pop() span = forward_sequences.pop() span_uids = CHAIN_DELIMITER.join(sequence_manager.block_uids[span.start : span.end]) stub = TransformerConnectionHandler.get_stub(sequence_manager.state.p2p, span.peer_id) metadata = sequence_manager.get_request_metadata( "rpc_backward", span_uids, *inputs, *grad_outputs, peer_id=span.peer_id ) grad_outputs, *span_grad_prompts = await run_remote_backward( span_uids, stub, sequence_manager.rpc_info, inputs, grad_outputs, prompts[span.start : span.end], timeout=sequence_manager.config.request_timeout, metadata=MSGPackSerializer.dumps(metadata), ) grad_outputs = [grad_outputs] grad_prompts_reversed.extend(span_grad_prompts) sequence_manager.on_request_success(span.peer_id) break except Exception as e: sequence_manager.on_request_failure(span.peer_id if span is not None else None) if attempt_no + 1 == sequence_manager.config.max_retries: raise delay = sequence_manager.get_retry_delay(attempt_no) logger.warning( f"Caught exception when running backward via {span} (retry in {delay:.0f} sec): {repr(e)}" ) maybe_log_traceback(e) await asyncio.sleep(delay) # For now, we do not support mixed dummy and grad prompts # Concat in num_layer dimension grad_prompts = torch.cat(grad_prompts_reversed[::-1], dim=0) if grad_prompts_reversed else None if grad_outputs_dtype is not None: grad_outputs = [tensor.to(device=grad_outputs_device, dtype=grad_outputs_dtype) for tensor in grad_outputs] if grad_prompts is not None: grad_prompts = grad_prompts.to(device=prompts_device, dtype=prompts_dtype) return grad_outputs, grad_prompts async def _gather_forward(input_batches, prompt_batches, sequence_manager): """Wrapper for asyncio.gather to perform parallel sequential forwards""" return await asyncio.gather( *[ sequential_forward(input_batch, prompt_batch, sequence_manager) for input_batch, prompt_batch in zip(input_batches, prompt_batches) ] ) async def _gather_backward( grad_output_batches, intermediate_input_batches, prompt_batches, forward_sequences, sequence_manager ): """Wrapper for asyncio.gather to perform parallel sequential backwards""" return await asyncio.gather( *[ sequential_backward((grad_output,), input_batch, prompt_batch, spans, sequence_manager) for grad_output, input_batch, prompt_batch, spans in zip( grad_output_batches, intermediate_input_batches, prompt_batches, forward_sequences ) ] ) class _RemoteSequentialAutogradFunction(torch.autograd.Function): """ PyTorch autograd function that provides forward and backward calls for the entire sequence of remote transformer blocks. This function splits input data into batches with <MAX_TOKENS_IN_BATCH> and performs efficient parallel processing. """ @staticmethod def forward(ctx, inputs: torch.Tensor, prompts: torch.Tensor, sequence_manager: RemoteSequenceManager): batch_size = max(MAX_TOKENS_IN_BATCH // inputs.shape[1], 1) input_batches: Sequence[torch.Tensor] = inputs.detach().split(batch_size) if is_dummy(prompts): prompt_batches = [DUMMY] * len(input_batches) else: prompt_batches: Sequence[torch.Tensor] = prompts.detach().split(batch_size, dim=1) sequence_manager.rpc_info # lazy init outputs = RemoteExpertWorker.run_coroutine(_gather_forward(input_batches, prompt_batches, sequence_manager)) assert len(outputs) == len(input_batches) output_batches = [output[0] for output in outputs] intemediate_input_batches = [output[1] for output in outputs] sequences_for_batches = [output[2] for output in outputs] ctx.prompt_batches = prompt_batches ctx.sequence_manager = sequence_manager ctx.intemediate_input_batches = intemediate_input_batches ctx.sequences_for_batches = sequences_for_batches return torch.cat(output_batches, dim=0) @staticmethod def backward(ctx, grad_outputs: torch.Tensor): intermediate_input_batches: List[Sequence[torch.Tensor]] = ctx.intemediate_input_batches forward_sequences: List[Sequence[RemoteSpanInfo]] = ctx.sequences_for_batches ctx.sequence_manager.rpc_info # lazy init batch_size = max(MAX_TOKENS_IN_BATCH // grad_outputs.shape[1], 1) grad_output_batches: Sequence[torch.Tensor] = grad_outputs.split(batch_size) assert len(intermediate_input_batches) == len(grad_output_batches) == len(forward_sequences) outputs = RemoteExpertWorker.run_coroutine( _gather_backward( grad_output_batches, intermediate_input_batches, ctx.prompt_batches, forward_sequences, ctx.sequence_manager, ) ) grad_input_batches = [output[0][0] for output in outputs] grad_prompt_batches = [output[1] for output in outputs] grad_inputs = torch.cat(grad_input_batches, dim=0) dummy_grad_prompts = [grad_prompt is None for grad_prompt in grad_prompt_batches] grad_prompts = torch.cat(grad_prompt_batches, dim=1) if not any(dummy_grad_prompts) else None return (grad_inputs, grad_prompts, None)
TheGrid-main
grid/client/sequential_autograd.py
import contextlib import json import os import re import tempfile import threading from typing import List, Optional, Tuple, Union import torch from hivemind.utils.logging import get_logger from transformers import BloomPreTrainedModel, modeling_utils from grid.utils.version import get_compatible_model_repo logger = get_logger(__name__) class FromPretrainedMixin: @classmethod def from_pretrained( cls, model_name_or_path: Union[str, os.PathLike, None], *args, low_cpu_mem_usage: Optional[bool] = None, torch_dtype: Optional[Union[str, torch.dtype]] = None, **kwargs, ): model_name_or_path = get_compatible_model_repo(model_name_or_path) if low_cpu_mem_usage is None: low_cpu_mem_usage = True if torch_dtype is None: # torch_dtype=None gives torch.float32 in transformers>=4.26.0. In contrast, # torch_dtype="auto" attempts to (1) use config.torch_dtype (if exists), (2) use dtype of the weights. torch_dtype = "auto" with ignore_keys(cls._keys_to_ignore_on_load_unexpected): return super().from_pretrained( model_name_or_path, *args, low_cpu_mem_usage=low_cpu_mem_usage, torch_dtype=torch_dtype, **kwargs ) from_pretrained.__doc__ = BloomPreTrainedModel.from_pretrained.__doc__.replace( "low_cpu_mem_usage(`bool`, *optional*)", "low_cpu_mem_usage(`bool`, *optional*, defaults to `True` in Grid)", ).replace( "torch_dtype (`str` or `torch.dtype`, *optional*)", 'torch_dtype (`str` or `torch.dtype`, *optional*, defaults to `"auto"` in Grid)', ) _shard_config = threading.local() _shard_config.ignored_keys = None @contextlib.contextmanager def ignore_keys(patterns: List[str]): try: prev_patterns = _shard_config.ignored_keys _shard_config.ignored_keys = patterns yield finally: _shard_config.ignored_keys = prev_patterns def patched_get_checkpoint_shard_files( pretrained_model_name_or_path, index_filename, *args, **kwargs ) -> Tuple[List[str], dict]: """Same as modeling_utils.get_checkpoint_shard_files(), but does not download shards for the ignored keys.""" should_ignore_keys = _shard_config.ignored_keys is not None tempdir_ctx = tempfile.TemporaryDirectory() if should_ignore_keys else contextlib.nullcontext() with tempdir_ctx as tempdir: if should_ignore_keys: with open(index_filename) as f: index = json.load(f) n_original_shards = len(set(index["weight_map"].values())) index["weight_map"] = { param_name: filename for param_name, filename in index["weight_map"].items() if all(re.search(pattern, param_name) is None for pattern in _shard_config.ignored_keys) } n_loaded_shards = len(set(index["weight_map"].values())) logger.debug(f"Loading {n_loaded_shards} shards out of {n_original_shards}") # Replace the original index with a patched JSON, where ignored keys are removed index_filename = os.path.join(tempdir, "pytorch_model.bin.index.json") with open(index_filename, "w") as f: json.dump(index, f) return original_get_checkpoint_shard_files(pretrained_model_name_or_path, index_filename, *args, **kwargs) original_get_checkpoint_shard_files = modeling_utils.get_checkpoint_shard_files modeling_utils.get_checkpoint_shard_files = patched_get_checkpoint_shard_files
TheGrid-main
grid/client/from_pretrained.py
"""Client-side functions responsible for choosing the best server, """
TheGrid-main
grid/client/routing/__init__.py
from __future__ import annotations import asyncio import dataclasses import itertools import logging import random import threading import time from typing import Any, Collection, Dict, List, Optional, Sequence, Union from weakref import WeakMethod import dijkstar import numpy as np from hivemind import DHT, P2P, MSGPackSerializer, PeerID from hivemind.dht.node import Blacklist from hivemind.moe.client.remote_expert_worker import RemoteExpertWorker from hivemind.proto import runtime_pb2 from hivemind.utils.logging import get_logger import grid.dht_utils from grid.client.routing.sequence_info import RemoteSequenceInfo from grid.client.routing.spending_policy import NoSpendingPolicy from grid.constants import PUBLIC_INITIAL_PEERS from grid.data_structures import ModuleUID, RemoteSpanInfo, ServerState from grid.server.handler import TransformerConnectionHandler from grid.utils.ping import PingAggregator from grid.utils.random import sample_up_to logger = get_logger(__name__) @dataclasses.dataclass class SequenceManagerConfig: initial_peers: Sequence[str] = tuple(PUBLIC_INITIAL_PEERS) # a list of initial peers for hivemind DHT dht_prefix: Optional[str] = None # a prefix for all dht keys that correspond to this model (default: model name) daemon_startup_timeout: int = 60 # timeout for the libp2p daemon connecting to initial peers show_route: Union[str, bool] = "inference" # show chosen route through servers. one of [False, "inference", True] allowed_servers: Optional[Collection[Union[PeerID, str]]] = None # if defined, send requests only to these servers use_server_to_server: bool = True # Use direct server-to-server communication request_timeout: float = 3 * 60 # timeout for forward/backward/inference requests update_period: float = 60 # refresh DHT information once in this many seconds max_retries: Optional[int] = None # max number retries before the client raises an exception (default: inf) min_backoff: float = 1 # after a repeated failure, sleep for this many seconds times 2 ** (num_failures - 1) max_backoff: float = 60 # limit maximal sleep time between retries to this value ban_timeout: float = 15 # when a remote peer fails to respond, prevent routing to that peer for this many seconds active_adapter: Optional[str] = None # name of active LoRA adapter (usually, Hugging Face repo) max_pinged: int = 5 # max servers to ping from each sequence side, per update ping_timeout: float = 2 # max time to wait for pings, per update @dataclasses.dataclass class SequenceManagerState: p2p: P2P = None sequence_info: Optional[RemoteSequenceInfo] = None rpc_info: Optional[dict] = None banned_peers: Optional[Blacklist] = None def __getitem__(self, ix: Union[int, slice]) -> SequenceManagerState: return dataclasses.replace(self, sequence_info=self.sequence_info[ix]) def __len__(self) -> int: return len(self.sequence_info) class RemoteSequenceManager: """ Sequence manager is a thread that keeps track of remote servers that hold the specified sequence of blocks. TL;DR it tells you, which peers you should ask to get a specific layer. It is used in RemoteSequential. When created, RemoteSequenceManager looks up which servers serve necessary layers by reading from DHT. Using this information, sequence manager can form sequences of servers that collectively have the full sequence. To form such a sequence, call .make_sequence with the appropriate optimization policy (see make_sequence docstr). :note: RemoteSequenceManager takes up some CPU and network I/O to operate in background. It is recommended to avoid running redundant sequence managers for the same set of layers. """ def __init__( self, config: SequenceManagerConfig, block_uids: Sequence[ModuleUID], *, dht: Optional[DHT] = None, state: Optional[SequenceManagerState] = None, ): assert config.initial_peers or dht is not None, "Please specify `config.initial_peers` or `dht`" assert config.dht_prefix, "Could not find dht_prefix in config, please create model with dht_prefix=..." assert len(block_uids) > 0, "Sequences must contain at least one block" self.config = config if state is None: state = SequenceManagerState() self.state = state if dht is None: dht = DHT( initial_peers=config.initial_peers, client_mode=True, num_workers=32, startup_timeout=config.daemon_startup_timeout, start=True, ) assert isinstance(dht, DHT) and dht.is_alive(), "`dht` must be a running hivemind.DHT instance" self.dht = dht if state.p2p is None: state.p2p = RemoteExpertWorker.run_coroutine(dht.replicate_p2p()) self.lock_changes = threading.Lock() self._thread = _SequenceManagerUpdateThread(config.update_period, WeakMethod(self._update)) self._thread_start_lock = threading.Lock() self.policy = NoSpendingPolicy() self.ping_aggregator = PingAggregator(dht) if state.banned_peers is None: state.banned_peers = Blacklist(base_time=config.ban_timeout, backoff_rate=2.0) if state.sequence_info is None: state.sequence_info = RemoteSequenceInfo.make_empty(block_uids) if state.sequence_info.last_updated_time is not None: assert block_uids == state.sequence_info.block_uids self._thread.ready.set() # no need to await the first dht fetch self._need_latest_infos = True def make_sequence( self, start_index: int = 0, end_index: Optional[int] = None, *, mode: str, cache_tokens_needed: Optional[int] = None, ) -> List[RemoteSpanInfo]: """ Form a sequence of remote servers that collectively serve all consecutive layers :param start_index: optional index of the first module in a sequence, default = the first of block_uids :param end_index: optional index of the last module (non-inclusive), default = after last of block uids :param mode: one of ["max_throughput", "min_latency"] """ with self._thread_start_lock: if not self.is_alive(): self._thread.start() if not self.ready.is_set(): self.update(wait=True) # this will await an existing update or trigger a new one (if not updating) end_index = end_index if end_index is not None else len(self) if mode == "min_latency": span_sequence = self._make_sequence_with_min_latency( start_index, end_index, cache_tokens_needed=cache_tokens_needed ) elif mode == "max_throughput": span_sequence = self._make_sequence_with_max_throughput(start_index, end_index) else: raise RuntimeError(f"Unexpected mode {mode}") if self.config.show_route is True or (mode == "min_latency" and self.config.show_route == "inference"): route_repr = " => ".join( [f"{span.start}:{span.end} via …{str(span.peer_id)[-6:]}" for span in span_sequence] ) logger.info(f"Route found: {route_repr}") return span_sequence def _make_sequence_with_min_latency( self, start_index: int, end_index: int, *, cache_tokens_needed: Optional[int] ) -> List[RemoteSpanInfo]: if start_index == end_index: return [] with self.lock_changes: missing_blocks = [ block_idx for block_idx in range(start_index, end_index) if not self.state.sequence_info.spans_containing_block[block_idx] ] if missing_blocks: raise MissingBlocksError(missing_blocks) server_infos = { span.peer_id: span.server_info for block_idx in range(start_index, end_index) for span in self.state.sequence_info.spans_containing_block[block_idx] } graph = self._build_inference_graph(start_index, end_index, cache_tokens_needed=cache_tokens_needed) path = dijkstar.find_path(graph, "start", "end") logger.debug(f"Path info: {path}") if start_index == 0 and end_index == len(self): logger.debug(f"Expected speed: {1 / path.total_cost:.1f} steps/sec") span_sequence = [] for peer_id, block_idx in path.nodes[1:-1]: if not span_sequence or span_sequence[-1].peer_id != peer_id: span_sequence.append(RemoteSpanInfo(peer_id, block_idx, block_idx, server_infos[peer_id])) else: span_sequence[-1].end = block_idx # Remove empty spans that can appear if we don't force to go to the end of each server and network delay # don't follow triangle inequality (delay(A, B) + delay(B, C) < delay(A, C)) due to measurement errors span_sequence = [span for span in span_sequence if span.length > 0] return span_sequence def _build_inference_graph( self, start_index: int, end_index: int, *, cache_tokens_needed: Optional[int], overhead_delay: float = 0.018, # Serialization overhead (empirically measured) default_inference_rps: float = 300, # If inference RPS unknown alloc_delay: float = 10, # If not enough cache left, we penalize the edge ) -> dijkstar.Graph: missing_blocks = [ block_idx for block_idx in range(start_index, end_index) if not self.state.sequence_info.spans_containing_block[block_idx] ] if missing_blocks: raise MissingBlocksError(missing_blocks) client_server_rtts = self.ping_aggregator.to_dict() graph = dijkstar.Graph() # Clent -> server network delays for span in self.state.sequence_info.spans_containing_block[start_index]: delay = self._rtt_to_delay(client_server_rtts.get(span.peer_id)) delay += overhead_delay if not self._has_cache_for(span, cache_tokens_needed): delay += alloc_delay graph.add_edge("start", (span.peer_id, start_index), delay) # Server -> client network delays for span in self.state.sequence_info.spans_containing_block[end_index - 1]: delay = self._rtt_to_delay(client_server_rtts.get(span.peer_id)) graph.add_edge((span.peer_id, end_index), "end", delay) # Server -> server network delays for block_idx in range(start_index + 1, end_index): for cur_span in self.state.sequence_info.spans_containing_block[block_idx - 1]: if cur_span.end != block_idx: # If we choose a server, we force to go to the end of it before switching to a new one # to avoid O(N^2) graphs for N servers continue for next_span in self.state.sequence_info.spans_containing_block[block_idx]: rtt = None if cur_span.server_info.next_pings is not None: rtt = cur_span.server_info.next_pings.get(next_span.peer_id.to_base58()) delay = self._rtt_to_delay(rtt) delay += overhead_delay if not self._has_cache_for(next_span, cache_tokens_needed): delay += alloc_delay graph.add_edge((cur_span.peer_id, block_idx), (next_span.peer_id, block_idx), delay) # Compute delays for span in self.state.sequence_info.spans_by_priority: for block_idx in range(max(span.start, start_index), min(span.end, end_index)): inference_rps = span.server_info.inference_rps if inference_rps is None: inference_rps = default_inference_rps graph.add_edge((span.peer_id, block_idx), (span.peer_id, block_idx + 1), 1.0 / inference_rps) return graph @staticmethod def _rtt_to_delay( rtt: float, *, default_delay: float = 0.15, # If network delay unknown max_delay: float = 5, # If unreachable, we don't want to discard the edge completely ) -> float: if rtt is None: return default_delay return min(rtt / 2, max_delay) @staticmethod def _has_cache_for(span: RemoteSpanInfo, cache_tokens_needed: Optional[int] = None) -> bool: if cache_tokens_needed is None or span.server_info.cache_tokens_left is None: return True # Here, `span` contains all blocks hosted by a server - but we won't necessarily run all of them through # this particular server in our path. It is difficult to estimate how many blocks we'll use at this stage, # so we assume that we'll use all of them (the worst case for the cache size) and get a pessimistic estimate. # This is okay since false positives are more costly than false negatives here. return cache_tokens_needed * 2 * span.length <= span.server_info.cache_tokens_left def _make_sequence_with_max_throughput( self, start_index: int, end_index: int, *, relay_penalty: float = 0.5 ) -> List[RemoteSpanInfo]: span_sequence = [] current_index = start_index while current_index < end_index: candidate_spans = self.state.sequence_info.spans_containing_block[current_index] if not candidate_spans: raise MissingBlocksError(current_index) span_weights = np.array( [ span.server_info.throughput * (1 if not span.server_info.using_relay else relay_penalty) for span in candidate_spans ], dtype=np.float64, ) chosen_span = np.random.choice(candidate_spans, p=span_weights / span_weights.sum()) assert chosen_span.start <= current_index < chosen_span.end span_sequence.append(dataclasses.replace(chosen_span, start=current_index)) current_index = chosen_span.end return span_sequence def __getitem__(self, ix: Union[int, slice]) -> RemoteSequenceManager: """Get a RemoteSequenceManager for a sub-sequence of blocks""" assert isinstance(ix, (int, slice)) if not isinstance(ix, slice): ix = slice(int(ix), int(ix) + 1, 1) return type(self)(self.config, self.block_uids[ix], dht=self.dht, state=self.state[ix]) def update(self, *, wait: bool): """Run an asynchronous update in background as soon as possible""" self.ready.clear() self._thread.trigger.set() if wait: self.ready.wait() def _update(self): """Perform an immediate and synchronous refresh, may take time""" new_block_infos = grid.dht_utils.get_remote_module_infos( self.dht, self.block_uids, active_adapter=self.config.active_adapter, latest=True ) for block_info in new_block_infos: if not block_info: continue # Apply whitelist, if defined if self.config.allowed_servers is not None: block_info.servers = { peer_id: server_info for peer_id, server_info in block_info.servers.items() if peer_id in self.config.allowed_servers or str(peer_id) in self.config.allowed_servers } # Remove temporarily banned peers, unless there are no peers left valid_servers = { peer_id: server_info for peer_id, server_info in block_info.servers.items() if peer_id not in self.state.banned_peers } if len(valid_servers) < len(block_info.servers): if valid_servers: logger.debug( f"Kept {len(valid_servers)} out of {len(block_info.servers)} servers holding {block_info.uid}" ) block_info.servers = valid_servers else: # If we blacklisted all servers, the error may actually be client-caused logger.debug(f"All servers holding {block_info.uid} are blacklisted, ignoring blacklist") with self.lock_changes: self.state.sequence_info.update_(new_block_infos) first_servers = [span.peer_id for span in self.state.sequence_info.spans_containing_block[0]] last_servers = [span.peer_id for span in self.state.sequence_info.spans_containing_block[-1]] pinged_servers = set(sample_up_to(first_servers, self.config.max_pinged)) pinged_servers |= set(sample_up_to(last_servers, self.config.max_pinged)) self.ping_aggregator.ping(list(pinged_servers), wait_timeout=self.config.ping_timeout) self.ready.set() def on_request_failure(self, peer_id: Optional[PeerID]): """remove a given peer from the routing table. If the routing is no longer possible, trigger an update""" if peer_id is not None: logger.debug(f"Peer {peer_id} did not respond, banning it temporarily") self.state.banned_peers.register_failure(peer_id) with self.lock_changes: should_update = False for info in self.state.sequence_info.block_infos: info.servers.pop(peer_id, None) if not info.servers: should_update = True if should_update: self.ready.clear() self.update(wait=False) def on_request_success(self, peer_id: PeerID): """if peer has a failure streak, clear that streak""" self.state.banned_peers.register_success(peer_id) def __len__(self): return len(self.block_uids) @property def is_alive(self): return self._thread.is_alive @property def ready(self) -> threading.Event: return self._thread.ready @property def block_uids(self): return self.state.sequence_info.block_uids @property def rpc_info(self): """Return the rpc_info queried from one of the servers that hold the first block""" if self.state.rpc_info is not None: return self.state.rpc_info with self._thread_start_lock: if not self.is_alive(): self._thread.start() for attempt_no in itertools.count(): peer_id = None try: if not self.ready.is_set(): self.update(wait=True) active_servers = [ peer_id for peer_id, server in self.state.sequence_info.block_infos[0].servers.items() if server.state == ServerState.ONLINE ] if not active_servers: raise MissingBlocksError(0) peer_id = random.choice(active_servers) stub = TransformerConnectionHandler.get_stub(self.state.p2p, peer_id) outputs = RemoteExpertWorker.run_coroutine( stub.rpc_info(runtime_pb2.ExpertUID(uid=self.block_uids[0]), timeout=self.config.request_timeout) ) self.state.rpc_info = MSGPackSerializer.loads(outputs.serialized_info) self.on_request_success(peer_id) break except Exception as e: self.on_request_failure(peer_id) if attempt_no + 1 == self.config.max_retries: raise delay = self.get_retry_delay(attempt_no) logger.warning( f"Caught exception when gathering information from peer {peer_id} " f"(retry in {delay:.0f} sec): {repr(e)}" ) maybe_log_traceback(e) time.sleep(delay) return self.state.rpc_info def get_retry_delay(self, attempt_no: int) -> float: if attempt_no == 0: return 0 return min(self.config.min_backoff * 2 ** (attempt_no - 1), self.config.max_backoff) def get_request_metadata(self, protocol: str, *args, **kwargs) -> Optional[Dict[str, Any]]: """ :param protocol: one of "rpc_forward", "rpc_backward" or "rpc_inference" :param args: request-specific inputs, typically block uids and input tensors :param kwargs: additional request context, such as remote peer ID :returns: msgpack-serialized metadata dict that will be passed alongside a given request """ return dict(points=self.policy.get_points(protocol, *args, **kwargs), active_adapter=self.config.active_adapter) def shutdown(self): self._thread.shutdown() class _SequenceManagerUpdateThread(threading.Thread): def __init__(self, update_period: float, ref_update_manager: WeakMethod): super().__init__(daemon=True) self.ref_update_manager = ref_update_manager self.ready = threading.Event() self.trigger = threading.Event() self.update_period = update_period self.should_shutdown = False def run(self) -> None: while not self.should_shutdown: update_manager = self.ref_update_manager() if update_manager is None: logger.debug(f"{self.__class__.__name__} exited because the sequence manager no longer exists") break try: self.trigger.clear() update_manager() except Exception as e: logger.exception(e) finally: del update_manager self.trigger.wait(self.update_period) logger.debug(f"{self.__class__.__name__} thread exited") def shutdown(self, timeout: Optional[float] = None): self.should_shutdown = True self.trigger.set() if self.is_alive(): self.join(timeout) def __del__(self): self.shutdown() def maybe_log_traceback(exc: Exception): traceback_level = logging.DEBUG if str(exc) or isinstance(exc, asyncio.TimeoutError) else logging.WARNING logger.log(traceback_level, "See detailed traceback below:", exc_info=True) class MissingBlocksError(RuntimeError): def __init__(self, block_indices: Union[int, Sequence[int]]): super().__init__( f"No servers holding blocks {block_indices} are online. " f"You can check the public swarm's state at https://health.grid.dev " f"If there are not enough servers, please connect your GPU: " f"https://github.com/Agora-workshop/grid#connect-your-gpu-and-increase-grid-capacity " )
TheGrid-main
grid/client/routing/sequence_manager.py
import dataclasses import time from typing import Iterable, List, Optional, Sequence, Tuple, Type, TypeVar from hivemind import get_logger from grid.data_structures import ModuleUID, RemoteModuleInfo, RemoteSpanInfo, ServerState logger = get_logger(__name__) T = TypeVar("T") @dataclasses.dataclass class RemoteSequenceInfo: """ A dataclass that stores general information about which servers hold any given layer; - updated by RemoteSequenceManager in a background thread - accessed by routing strategies in .on_update :note: this class should *not* be modified by RoutingStrategy.on_update to avoid interference between strategies; Any metadata specific to one routing strategy, it should be stored inside that strategy. Any information that is used by most routing strategies should be moved from said strategies to this class. """ block_uids: Tuple[ModuleUID, ...] block_infos: Tuple[RemoteModuleInfo, ...] # note: the contents of RemoteModuleInfo can and will be updated spans_by_priority: List[RemoteSpanInfo] spans_containing_block: Tuple[List[RemoteSpanInfo], ...] last_updated_time: Optional[float] @classmethod def make_empty(cls: Type[T], block_uids: Iterable[ModuleUID]) -> T: block_uids = tuple(block_uids) empty_block_infos = tuple(RemoteModuleInfo(uid, {}) for uid in block_uids) empty_spans = tuple([] for _ in range(len(block_uids))) return cls(block_uids, empty_block_infos, [], empty_spans, last_updated_time=None) def __getitem__(self, ix: slice): assert isinstance(ix, slice) block_uids, block_infos = self.block_uids[ix], self.block_infos[ix] spans_by_priority, spans_containing_block = self.compute_spans(block_infos) return RemoteSequenceInfo( block_uids, block_infos, spans_by_priority, spans_containing_block, self.last_updated_time ) def __len__(self): return len(self.block_uids) def update_(self, new_block_infos: List[Optional[RemoteModuleInfo]]): assert len(new_block_infos) == len(self.block_uids) for block_index, (uid, info) in enumerate(zip(self.block_uids, new_block_infos)): if info is None: logger.debug(f"Found no block info for block {uid}") continue if not isinstance(info, RemoteModuleInfo): logger.warning(f"Unexpected dht entry type for {uid}: {info}") continue if not info.servers: logger.debug(f"Found no active peers for block {uid}") continue if info.uid != uid: logger.warning(f"The DHT entry for {uid} actually points to {info.uid}") continue self.block_infos[block_index].servers = info.servers self.spans_by_priority, self.spans_containing_block = self.compute_spans(self.block_infos) self.last_updated_time = time.perf_counter() @staticmethod def compute_spans(block_infos: Sequence[RemoteModuleInfo]): closed_spans = [] active_spans = {} for block_index, info in enumerate(block_infos): if info is not None: for peer_id, server_info in info.servers.items(): if server_info.state != ServerState.ONLINE: continue if peer_id not in active_spans: active_spans[peer_id] = RemoteSpanInfo( peer_id=peer_id, start=block_index, end=block_index + 1, server_info=server_info, ) else: # peer_id in active_spans active_spans[peer_id].end = block_index + 1 for peer_id in list(active_spans.keys()): if ( info is None or peer_id not in info.servers or info.servers[peer_id].state != ServerState.ONLINE or block_index == len(block_infos) - 1 ): closed_spans.append(active_spans.pop(peer_id)) assert not active_spans, f"spans: {active_spans}" closed_spans.sort(key=lambda span: span.length, reverse=True) spans_containing_block = tuple(list() for _ in range(len(block_infos))) for span in closed_spans: for block_index in range(span.start, span.end): spans_containing_block[block_index].append(span) return closed_spans, spans_containing_block
TheGrid-main
grid/client/routing/sequence_info.py
""" An interface for exchanging internal "BLOOM points" for higher priority compute requests. NOT IMPLEMENTED. The intent is to let Grid participants earn points by helping others while idle (e.g. at night), then use these points to run their own compute experiments faster. See Section 4 of https://arxiv.org/abs/2209.01188 for discussion. """ from abc import ABC, abstractmethod class SpendingPolicyBase(ABC): @abstractmethod def get_points(self, protocol: str, *args, **kwargs) -> float: pass class NoSpendingPolicy(SpendingPolicyBase): def get_points(self, protocol: str, *args, **kwargs) -> float: return 0.0
TheGrid-main
grid/client/routing/spending_policy.py
from setuptools import setup, find_packages setup( name = 'GeneSplice', packages = find_packages(exclude=[]), version = '0.0.3', license='MIT', description = 'GeneSplice Model, Ultra-Long Rage Genomic Expression Modelling', author = 'Kye Gomez', author_email = '[email protected]', long_description_content_type = 'text/markdown', url = 'https://github.com/kyegomez/GeneSplice', keywords = [ 'artificial intelligence', 'deep learning', 'transformers', 'attention mechanism', 'long context', 'genomics', 'pre-training' ], install_requires=[ 'torch', 'einops', 'transformers', 'accelerate', 'fairscale', 'timm', 'flash-attn', ], classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'Topic :: Scientific/Engineering :: Artificial Intelligence', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3.6', ], )
GeneSplice-main
setup.py
import torch import pytest from GeneSplice.model import GeneSplice, GeneSpliceTokenizer # Assuming the module name is GeneSplice def test_tokenizer_initialization(): tokenizer = GeneSpliceTokenizer() assert tokenizer is not None, "Tokenizer failed to initialize" def test_model_initialization(): model = GeneSplice() assert model is not None, "Model failed to initialize" def test_tokenization(): tokenizer = GeneSpliceTokenizer() text = "Hello, world!" tokens = tokenizer.tokenize_texts(text) assert tokens is not None, "Tokenization failed" assert tokens.shape[1] > 0, "Tokenization resulted in zero tokens" def test_model_forward(): model = GeneSplice() tokenizer = GeneSpliceTokenizer() text = "Hello, world!" tokens = tokenizer.tokenize_texts(text) output = model(tokens) assert output is not None, "Forward pass failed" assert output.shape == tokens.shape, "Output shape is different from input shape" if __name__ == "__main__": pytest.main()
GeneSplice-main
main.py
import torch import torch.nn as nn import torch.nn.functional as F from flash_attn.flash_attention import FlashMHA # Replace this with your correct GPU device device = "cuda:0" dtype=torch.float16 class DilatedAttention(nn.Module): def __init__(self, d_model, num_heads, dilation_rate, segment_size, dropout=0.0, casual=False): super(DilatedAttention, self).__init__() self.d_model = d_model self.num_heads = num_heads self.dilation_rate = dilation_rate self.segment_size = segment_size self.attention = FlashMHA(embed_dim=d_model, num_heads=num_heads, device=device, dtype=dtype) self.dropout = nn.Dropout(dropout) self.casual = casual def get_mask(self, i, j): return torch.ones((i, j), deice=device, dtype=torch.bool).triu(j - i + 2) def forward(self, x): batch_size, seq_len, _ = x.shape # Split and sparsify x = x.view(batch_size, -1, self.segment_size, self.d_model) x = x[:, :, :: self.dilation_rate, :] # Perform attention attn_output, _ = self.attention(x, x, x) #if casual create a mask and apply to the output if self.casual: mask = self.get_mask(attn_output.size(1), attn_output.size(1)) attn_output = attn_output.masked_fill(mask, float('-inf')) #apply dropout attn_output = self.dropout(attn_output) # Scatter and concatenate attn_output = attn_output.view(batch_size, -1, self.d_model) return attn_output class LongNetTransformer(nn.Module): def __init__(self, d_model, num_heads, dilation_rates, segment_sizes): super(LongNetTransformer, self).__init__() assert len(dilation_rates) == len(segment_sizes), "dilation_rates and segment_sizes should have the same length" self.d_model = d_model self.num_heads = num_heads self.dilation_rates = dilation_rates self.segment_sizes = segment_sizes self.dilated_attention_layers = nn.ModuleList( [DilatedAttention(d_model, num_heads, dilation_rate, segment_size)] for dilation_rate, segment_size in zip(dilation_rates, segment_sizes) ) def forward(self, x): #accumlate outputs from different layers outputs = [] #process each dilated attention layer for i in range(len(self.dilated_attention_layers)): output = self.dilated_attention_layers[i](x) outputs.append(output) #combine the outputs output = torch.sum(torch.stack(outputs), dim=0) return output
GeneSplice-main
GeneSplice/attention.py
# from GeneSplice.model import GeneSpliceTokenizer, GeneSplice from GeneSplice.training import Train from torchscale.torchscale.architecture.decoder import DecoderConfig, Decoder from torchscale.torchscale.component.embedding import PositionalEmbedding
GeneSplice-main
GeneSplice/__init__.py
import torch # from torchscale.torchscale.architecture.decoder import DecoderConfig, Decoder # from torchscale.torchscale.component.embedding import PositionalEmbedding from transformers import AutoTokenizer from torch.nn import Embedding, Module import bitsandbytes from GeneSplice import DecoderConfig, Decoder, PositionalEmbedding class GeneSpliceTokenizer: def __init__(self): self.tokenizer = AutoTokenizer.from_pretrained( "EleutherAI/gpt-neox-20b", eos_token="<eos>", pad_token="<pad>", extra_ids=0, model_max_length=8192 ) def tokenize_texts(self, texts): return self.tokenizer(texts, return_tensors="pt", padding=True, truncation=True).input_ids class GeneSplice(Module): def __init__(self): super().__init__() self.embed = bitsandbytes.nn.modules.Embedding( 320002, 2048, padding_idx=1 ) self.embed_positions = PositionalEmbedding( 2048, 2048, 1 ) self.output_projection = torch.nn.Linear( 2048, 32002, bias=False ) self.config = DecoderConfig( decoder_layers=24, decoder_embed_dim=2048, decoder_ffn_embed_dim=8192, decoder_attention_heads=32, dropout=0.1, activation_fn="gelu", attention_dropout=0.1, decoder_dilation_rate=4, decoder_segment_size=2, vocab_size=64007, ) self.decoder = Decoder( self.config, embed_tokens=self.embed, embed_positions=self.embed_positions, output_projection=self.output_projection ) def forward(self, text_tokens, **kwargs): model_input = self.decoder.forward_embedding(text_tokens)[0] return self.decoder(model_input, passed_x=model_input)[0]
GeneSplice-main
GeneSplice/model.py
import numpy as np import torch # This is the unfused version of StableAdamW. It is slower than the fused version (coming). class StableAdamWUnfused(torch.optim.Optimizer): def __init__( self, params, lr=0.002, weight_decay=0.2, betas=(0.9, 0.99), eps=1e-8, clip_thresh=1.0, precision="amp_bfloat16", custom_scalar=65536, ): beta1, beta2 = betas[0], betas[1] defaults = dict(lr=lr, weight_decay=weight_decay, beta1=beta1, beta2=beta2) super(StableAdamWUnfused, self).__init__(params, defaults) self.eps = eps self.d = clip_thresh # Set precision to "custom_fp16" if you want to use a fixed loss scalar, custom_scalar, which is divided out in the update step. # If you do this, call (custom_scalar * loss).backward() instead of loss.backward(). self.precision = precision self.custom_scaler = custom_scalar for group in self.param_groups: group["step"] = 1.0 print("Using StableAdamWUnfused-v1") def __setstate__(self, state): super(StableAdamWUnfused, self).__setstate__(state) def step(self, closure=None): loss = None if closure is not None: loss = closure() for group in self.param_groups: lr = group["lr"] weight_decay = group["weight_decay"] beta1 = group["beta1"] beta2 = group["beta2"] step = group["step"] for p in group["params"]: if p.grad is None: continue theta = p.data param_state = self.state[p] if self.precision == "custom_fp16": g = p.grad.data / self.custom_scaler if torch.any(torch.isnan(g) | torch.isinf(g)): continue else: g = p.grad.data if "exp_avg" not in param_state: v = param_state["exp_avg"] = torch.zeros_like(theta) u = param_state["exp_avg_sq"] = torch.zeros_like(theta) else: v = param_state["exp_avg"] u = param_state["exp_avg_sq"] beta1hat = beta1 * (1 - beta1 ** (step - 1)) / (1 - beta1**step) beta2hat = beta2 * (1 - beta2 ** (step - 1)) / (1 - beta2**step) v = v.mul_(beta1hat).add_(g, alpha=1.0 - beta1hat) u = u.mul_(beta2hat).addcmul_(g, g, value=1.0 - beta2hat) denominator = u.sqrt().add_(self.eps) # StableAdamW = AdamW + update clipping (https://arxiv.org/abs/1804.04235) applied tensor-wise. rms = ( torch.div( g.pow(2), torch.maximum(u, (self.eps**2) * torch.ones_like(u)) ) .mean() .sqrt() .item() ) theta = theta.mul_(1.0 - lr * weight_decay).addcdiv_( v, denominator, value=-lr * (1.0 / max(1.0, rms / self.d)) ) # save current params param_state["exp_avg"] = v param_state["exp_avg_sq"] = u group["step"] = step + 1
GeneSplice-main
GeneSplice/utils.py
import math import multiprocessing import os from datetime import timedelta from functools import partial from itertools import chain import torch from torch.distributed.fsdp import ( FullyShardedDataParallel, MixedPrecision, BackwardPrefetch, ShardingStrategy, ) from accelerate import Accelerator from accelerate.utils import (DummyOptim, DummyScheduler, InitProcessGroupKwargs) from datasets import concatenate_datasets, load_dataset from lion_pytorch import Lion from torch.nn import LayerNorm from torch.nn import LayerNorm from torch.distributed.algorithms._checkpoint.checkpoint_wrapper import ( CheckpointImpl, apply_activation_checkpointing, checkpoint_wrapper) from torch.distributed.fsdp.wrap import ( transformer_auto_wrap_policy, ) from torch.optim import AdamW from torch.utils.data import DataLoader from tqdm import tqdm from transformers import (AutoTokenizer, default_data_collator, get_cosine_schedule_with_warmup, get_linear_schedule_with_warmup, set_seed) # INTEGRATE GeneSplice selector + stable8bitfusedadam from GeneSplice.torchscale.torchscale.architecture.decoder import Decoder from GeneSplice.utils import StableAdamWUnfused from GeneSplice.model import GeneSplice, GeneSpliceTokenizer ############ SETUP CONFIG # import torch.distributed as dist # dist.init_process_group(backend='nccl', init_method="env://") ################ class CFG: BATCH_SIZE: int = 3 GRADIENT_ACCUMULATE_EVERY: int = 1 SEED: int = 42 LEARNING_RATE: float = 3e-4 WEIGHT_DECAY: float = 0.1 SEQ_LEN: int = 8192 NUM_CPU: int = multiprocessing.cpu_count() USE_DEEPSPEED: bool = True USE_FSDP: bool = False USE_PRETOKENIZED: bool = False USE_ACTIVATION_CHECKPOINTING: bool = False RESUME_FROM_CHECKPOINT: str = None CHECKPOINTING_STEPS: int = 1000 OUTPUT_DIR: str = "YOUR_OUTPUT_DIR" ENTITY_NAME: str = "YOUR_ENTITY_NAME" #wandb # helpers def print_num_params(model, accelerator: Accelerator): n_params = sum(p.numel() for p in model.parameters() if p.requires_grad) accelerator.print(f"Number of parameters in model: {n_params}") # activation checkpointing def activation_checkpointing( model: torch.nn.Module, offload_to_cpu: bool = False, accelerator: Accelerator = None, ): """ Apply activation checkpointing to a model. Args: model (Module): The model to which to apply activation checkpointing. offload_to_cpu (bool, optional): Whether to offload the activations to CPU. Defaults to False. accelerator (Accelerator, optional): The Accelerate library accelerator. Defaults to None. """ if accelerator is not None: accelerator.print(f"Using activation checkpointing") #maybe error here in decoder, use parallel transformer block check_fn = lambda submodule: isinstance(submodule, Decoder) non_reentrant_wrapper = partial( checkpoint_wrapper, offload_to_cpu=offload_to_cpu, checkpoint_impl=CheckpointImpl.NO_REENTRANT, ) apply_activation_checkpointing( model, checkpoint_wrapper_fn=non_reentrant_wrapper, check_fn=check_fn ) # FSDP def fsdp( model: torch.nn.Module, auto_wrap: bool = False, mp: str = "fp32", shard_strat: str = "NO_SHARD", ): """ This function wraps a given PyTorch model with the FullyShardedDataParallel (FSDP) wrapper to enable efficient data parallelism and model sharding. Args: model (torch.nn.Module): The original PyTorch model to be wrapped with FSDP. auto_wrap (bool, optional): If True, it enables automatic wrapping of the model's layers according to the transformer_auto_wrap_policy. Default is False. mp (str, optional): The mixed precision mode to be used. Can be 'bf16' for BFloat16, 'fp16' for Float16 or 'fp32' for Float32 precision. Default is 'fp32'. shard_strat (str, optional): The sharding strategy to be used. Can be 'SHARD_GRAD' for sharding at gradient computation, 'FULL_SHARD' for full model sharding or 'NO_SHARD' for no sharding. Default is 'NO_SHARD'. Raises: ValueError: If the provided mp (mixed precision mode) is not 'bf16', 'fp16' or 'fp32'. ValueError: If the provided shard_strat (sharding strategy) is not 'SHARD_GRAD', 'FULL_SHARD' or 'NO_SHARD'. Returns: torch.nn.Module: The input model wrapped with FSDP. """ if auto_wrap: GeneSplice_auto_wrap_policy = partial( transformer_auto_wrap_policy, transformer_layer_cls={ Decoder, }, ) else: GeneSplice_auto_wrap_policy = None if mp == "bf16": mp_fsdp = MixedPrecision( param_dtype=torch.bfloat16, # Gradient communication precision. reduce_dtype=torch.bfloat16, # Buffer precision. buffer_dtype=torch.bfloat16, ) elif mp == "fp16": mp_fsdp = MixedPrecision( param_dtype=torch.float16, # Gradient communication precision. reduce_dtype=torch.float16, # Buffer precision. buffer_dtype=torch.float16, ) elif mp == "fp32": mp_fsdp = MixedPrecision( param_dtype=torch.float32, # Gradient communication precision. reduce_dtype=torch.float32, # Buffer precision. buffer_dtype=torch.float32, ) else: raise ValueError( "Invalid scheduler_type. Expected 'bf16', 'fp16' or 'fp32', got: {}".format( mp ) ) if shard_strat == "SHARD_GRAD": sharding_strat_fsdp = ShardingStrategy.SHARD_GRAD_OP elif shard_strat == "FULL_SHARD": sharding_strat_fsdp = ShardingStrategy.FULL_SHARD elif shard_strat == "NO_SHARD": sharding_strat_fsdp = ShardingStrategy.NO_SHARD else: raise ValueError( "Invalid scheduler_type. Expected 'SHARD_GRAD', 'FULL_SHARD' or 'NO_SHARD', got: {}".format( shard_strat ) ) model = FullyShardedDataParallel( model, auto_wrap_policy=GeneSplice_auto_wrap_policy, mixed_precision=mp_fsdp, backward_prefetch=BackwardPrefetch.BACKWARD_PRE, sharding_strategy=sharding_strat_fsdp, forward_prefetch=True, use_orig_params=True, ) return model # learning rate scheduler def get_lr_scheduler_with_warmup( optimizer: torch.optim.Optimizer, scheduler_type: str, num_warmup_steps: int, max_train_steps: int, grad_accumulate_every: int = 1, accelerator: Accelerator = None, ): """ Get a learning rate scheduler with warmup. Args: optimizer (Optimizer): The optimizer for which to create the learning rate scheduler. scheduler_type (str): The type of learning rate scheduler to create, either "linear" or "cosine". num_warmup_steps (int): The number of warmup steps for the learning rate scheduler. max_train_steps (int): The maximum number of training steps. grad_accumulate_every (int, optional): The gradient accumulation factor. Defaults to 1. accelerator (Accelerator, optional): The Accelerate library accelerator. Defaults to None. Returns: The learning rate scheduler with warmup. Raises: ValueError: If scheduler_type is not "linear" or "cosine". """ NUM_WARMUP_STEPS = num_warmup_steps GRADIENT_ACCUMULATE_EVERY = grad_accumulate_every if accelerator is not None: accelerator.print(f"Using {scheduler_type} lr scheduler") if scheduler_type == "linear": return get_linear_schedule_with_warmup( optimizer=optimizer, num_warmup_steps=NUM_WARMUP_STEPS * GRADIENT_ACCUMULATE_EVERY, num_training_steps=max_train_steps * GRADIENT_ACCUMULATE_EVERY, ) elif scheduler_type == "cosine": return get_cosine_schedule_with_warmup( optimizer=optimizer, num_warmup_steps=NUM_WARMUP_STEPS * GRADIENT_ACCUMULATE_EVERY, num_training_steps=max_train_steps * GRADIENT_ACCUMULATE_EVERY, ) else: raise ValueError( "Invalid scheduler_type. Expected 'linear' or 'cosine', got: {}".format( scheduler_type ) ) # optimizers def decoupled_optimizer( model: torch.nn.Module, learning_rate: float, weight_decay: float, beta_1: float, beta_2: float, optimizer_type: str, use_fsdp: bool = True, accelerator: Accelerator = None, ): """ Decouples the optimizer from the training process. This function sets up the optimizer for the model by creating two groups of parameters: one for weight decay and one without weight decay. Then, it initializes the optimizer with these two groups of parameters. Args: model (Module): The model whose parameters are optimized. learning_rate (float): The learning rate for the optimizer. weight_decay (float): The weight decay for the optimizer. beta_1 (float): The exponential decay rate for the 1st moment estimates. beta_2 (float): The exponential decay rate for the 2nd moment estimates. optimizer_type (str): The type of the optimizer. Can be 'lion', 'adamw', or 'stable_adamw'. use_fsdp (bool, optional): If True, the optimizer will work with fully sharded data parallelism. Defaults to True. accelerator (Accelerator, optional): The accelerator from HuggingFace's Accelerate library. Defaults to None. Returns: Optimizer: The initialized optimizer. Raises: ValueError: If the optimizer type is not 'lion', 'adamw' or 'stable_adamw'. """ accelerator.print(f"Using {optimizer_type} optimizer") # Create an empty dictionary called param_dict to store the model's named parameters. param_dict = {} # Iterate over the model's named parameters and populate the param_dict with key-value pairs. for param_name, param in model.named_parameters(): param_dict[param_name] = param # Separate the model's named modules into two groups: decay and no_decay. # Create an empty list to store the names of the LayerNorm and Embedding layer weights with no weight decay. no_decay = [] if use_fsdp: exclude_module = "_fsdp_wrapped_module.token_emb" else: exclude_module = "token_emb" # Iterate through the named modules of the model. for module_name, module in model.named_modules(): # Check if the current module is an instance of any of the desired types (LayerNorm or torch.nn.Embedding). for ndim in [LayerNorm, torch.nn.Embedding]: if isinstance(module, ndim): # If torch.nn.Embedding, append its name with a ".weight" suffix to the no_decay list. if module_name == exclude_module: no_decay.append(f"{module_name}.weight") else: # If the module is an instance of LayerNorm no_decay.append(f"{module_name}.gamma") # Exit the inner loop since the desired module has been found. break # Create an empty list to store the names of the Linear layer weights with weight decay. decay = [] # Iterate through the named modules of the model. for module_name, module in model.named_modules(): # Check if the current module is an instance of the desired type (torch.nn.Linear). for ndim in [torch.nn.Linear]: if isinstance(module, ndim): # If the module is an instance of torch.nn.Linear, append its name with a ".weight" suffix to the decay list. decay.append(f"{module_name}.weight") # Exit the inner loop since the desired module has been found. break # Create two separate lists of model parameters: decay_param and no_decay_param. # The decay_param list contains the parameters that should have weight decay applied. # The no_decay_param list contains the parameters that should not have weight decay applied, excluding the 'to_logits.weight' parameter. # Create an empty list called decay_param to store the parameters with weight decay. decay_param = [] if use_fsdp: exclude_param = "_fsdp_wrapped_module.to_logits.weight" else: exclude_param = "to_logits.weight" # Iterate over the decay list, which contains the names of the parameters with weight decay. for param in decay: # Check if the current parameter is not 'to_logits.weight'. # Append the corresponding parameter from param_dict to the decay_param list. if param != exclude_param: decay_param.append(param_dict[param]) # Create an empty list called no_decay_param to store the parameters without weight decay. no_decay_param = [] # Iterate over the no_decay list, which contains the names of the parameters without weight decay. for param in no_decay: # Append the corresponding parameter from param_dict to the no_decay_param list. no_decay_param.append(param_dict[param]) # Create a list called grouped_params that contains two dictionaries. # The first dictionary has the decay_param list and the corresponding weight_decay value. # The second dictionary has the no_decay_param list and a weight_decay value of 0.0. grouped_params = [ {"params": decay_param, "weight_decay": weight_decay}, {"params": no_decay_param, "weight_decay": 0.0}, ] # Create a variable called optimizer that stores an instance of the optimizer. if optimizer_type == "lion": optimizer = Lion(grouped_params, lr=learning_rate, betas=(beta_1, beta_2),) elif optimizer_type == "adamw": optimizer = AdamW(grouped_params, lr=learning_rate, betas=(beta_1, beta_2),) elif optimizer_type == "deepspeed": optimizer = DummyOptim(grouped_params, lr=learning_rate, betas=(beta_1, beta_2),) elif optimizer_type == "stable_adamw": optimizer = StableAdamWUnfused( grouped_params, lr=learning_rate, betas=(beta_1, beta_2), ) else: raise ValueError( "Invalid optimizer_type. Expected 'lion', 'adamw', 'deepspeed' or 'stable_adamw', got: {}".format( optimizer_type ) ) # Return the optimizer. return optimizer # dataloaders def build_dataloaders(): """ Build data loaders for training. This function performs the following steps: 1. Load the tokenizer from the pretrained "EleutherAI/gpt-neox-20b" model. 2. Load the "openwebtext" dataset. 3. Tokenize the dataset, adding the end-of-sentence token to each text. 4. Process the tokenized dataset into chunks of a specified block size. Returns: Dataset: The processed dataset ready for training. """ tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neox-20b") dataset = load_dataset("openwebtext", split="train") tokenized_dataset = dataset.map( lambda example: tokenizer([t + tokenizer.eos_token for t in example["text"]]), batched=True, num_proc=CFG.NUM_CPU, remove_columns=["text"], ) block_size = CFG.SEQ_LEN # Main data processing function that will concatenate all texts from our dataset and generate chunks of block_size. def group_texts(examples): # Concatenate all texts. concatenated_examples = {k: list(chain(*examples[k])) for k in examples.keys()} total_length = len(concatenated_examples[list(examples.keys())[0]]) # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can # customize this part to your needs. if total_length >= block_size: total_length = (total_length // block_size) * block_size # Split by chunks of max_len. result = { k: [t[i : i + block_size] for i in range(0, total_length, block_size)] for k, t in concatenated_examples.items() } return result train_dataset = tokenized_dataset.map( group_texts, batched=True, num_proc=CFG.NUM_CPU, ) return train_dataset #switch to falconwebdataset def build_pre_tokenized(): d0 = load_dataset("conceptofmind/c4_0-to-20_neox_with_eos_8k", split="train") d1 = load_dataset("conceptofmind/c4_21-to-40_neox_with_eos_8k", split="train") d2 = load_dataset("conceptofmind/c4_41-to-60_neox_with_eos_8k", split="train") d3 = load_dataset("conceptofmind/c4_61-to-80_neox_with_eos_8k", split="train") d4 = load_dataset("conceptofmind/c4_81-to-100_neox_with_eos_8k", split="train") train_dataset = concatenate_datasets([d0, d1, d2, d3, d4]) return train_dataset def Train(): # accelerator timeout = InitProcessGroupKwargs(timeout=timedelta(seconds=1_000_000)) accelerator = Accelerator( gradient_accumulation_steps=CFG.GRADIENT_ACCUMULATE_EVERY, mixed_precision="fp16", log_with="wandb", kwargs_handlers=[timeout], ) # AcceleratorState().deepspeed_plugin.deepspeed_config['train_micro_batch_size_per_gpu'] = 4 #?????? accelerator.init_trackers( project_name="GeneSplice", config={ "batch_size": CFG.BATCH_SIZE, "gradient_accumulate_every": CFG.GRADIENT_ACCUMULATE_EVERY, "learning_rate": CFG.LEARNING_RATE, "seq_len": CFG.SEQ_LEN, }, init_kwargs={"wandb": {"entity": CFG.ENTITY_NAME}}, ) accelerator.print(f"Total GPUS: {accelerator.num_processes}") # set seed set_seed(CFG.SEED) GeneSplice.to(accelerator.device) print_num_params(model, accelerator) if CFG.USE_FSDP: model = fsdp( model, mp="fp16", shard_strat="SHARD_GRAD" ) if CFG.USE_ACTIVATION_CHECKPOINTING: activation_checkpointing(model, accelerator) model = accelerator.prepare(model) # dataloaders if CFG.USE_PRETOKENIZED: train_dataset = build_pre_tokenized() else: train_dataset = build_dataloaders() train_loader = DataLoader( train_dataset, batch_size=CFG.BATCH_SIZE, collate_fn=default_data_collator, ) # optimizer optim = decoupled_optimizer( model=model, learning_rate=CFG.LEARNING_RATE, weight_decay=CFG.WEIGHT_DECAY, beta_1=0.90, beta_2=0.95, optimizer_type='deepspeed', use_fsdp=True, accelerator=accelerator ) # Determine number of training steps max_train_steps = math.ceil(len(train_loader) / CFG.GRADIENT_ACCUMULATE_EVERY) accelerator.print(f"Max train steps: {max_train_steps}") # lr scheduler NUM_WARMUP_STEPS = int(max_train_steps * 0.01) accelerator.print(f"Num warmup steps: {NUM_WARMUP_STEPS}") if CFG.USE_DEEPSPEED: lr_scheduler = DummyScheduler( optim, total_num_steps=max_train_steps * accelerator.num_processes, warmup_num_steps=NUM_WARMUP_STEPS ) else: lr_scheduler = get_lr_scheduler_with_warmup( optimizer=optim, scheduler_type="cosine", num_warmup_steps=NUM_WARMUP_STEPS, max_train_steps=max_train_steps, grad_accumulate_every=CFG.GRADIENT_ACCUMULATE_EVERY, ) # prepare optim, train_loader, lr_scheduler = accelerator.prepare( optim, train_loader, lr_scheduler ) # checkpoint scheduler accelerator.register_for_checkpointing(lr_scheduler) # I do not know why Huggingface recommends recalculation of max_train_steps max_train_steps = math.ceil(len(train_loader) / CFG.GRADIENT_ACCUMULATE_EVERY) accelerator.print(f"Max train steps recalculated: {max_train_steps}") # Total batch size for logging total_batch_size = ( CFG.BATCH_SIZE * accelerator.num_processes * CFG.GRADIENT_ACCUMULATE_EVERY ) accelerator.print(f"Total batch size: {total_batch_size}") # resume training progress_bar = tqdm( range(max_train_steps), disable=not accelerator.is_local_main_process ) completed_steps = 0 if CFG.RESUME_FROM_CHECKPOINT: if CFG.RESUME_FROM_CHECKPOINT is not None or CFG.RESUME_FROM_CHECKPOINT != "": accelerator.print(f"Resuming from checkpoint {CFG.RESUME_FROM_CHECKPOINT}") accelerator.load_state(CFG.RESUME_FROM_CHECKPOINT) path = os.path.basename(CFG.RESUME_FROM_CHECKPOINT) training_difference = os.path.splitext(path)[0] # need to multiply `gradient_accumulation_steps` to reflect real steps resume_step = ( int(training_difference.replace("step_", "")) * CFG.GRADIENT_ACCUMULATE_EVERY ) if CFG.RESUME_FROM_CHECKPOINT and resume_step is not None: train_loader = accelerator.skip_first_batches(train_loader, resume_step) completed_steps += resume_step progress_bar.update(resume_step) # training model.train() for step, batch in enumerate(train_loader): with accelerator.accumulate(model): inputs = batch["input_ids"].to(accelerator.device) loss = model(inputs, return_loss=True) accelerator.backward(loss) accelerator.log({"loss": loss.item()}, step=step) if accelerator.sync_gradients: accelerator.clip_grad_norm_(model.parameters(), 1.0) optim.step() lr_scheduler.step() optim.zero_grad() if accelerator.sync_gradients: progress_bar.update(1) completed_steps += 1 if isinstance(CFG.CHECKPOINTING_STEPS, int): if completed_steps % CFG.CHECKPOINTING_STEPS == 0: output_dir = f"step_{completed_steps }" if CFG.OUTPUT_DIR is not None: output_dir = os.path.join(CFG.OUTPUT_DIR, output_dir) accelerator.save_state(output_dir) if completed_steps >= max_train_steps: break # end training # accelerator.print(f"Training Finished") accelerator.end_training() # save final model # accelerator.print(f"Saving model to {CFG.OUTPUT_DIR}") if CFG.OUTPUT_DIR is not None: accelerator.wait_for_everyone() unwrapped_model = accelerator.unwrap_model(model) with accelerator.main_process_first(): accelerator.save( unwrapped_model.state_dict(), f"{CFG.OUTPUT_DIR}/final/final_model.pt" ) if __name__ == "__main__": Train()
GeneSplice-main
GeneSplice/training.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] from io import open from setuptools import find_packages, setup setup( name="torchscale", version="0.2.0", author="TorchScale Team", author_email="[email protected]", description="Transformers at any scale", long_description=open("README.md", "r", encoding="utf-8").read(), long_description_content_type="text/markdown", keywords="Transformers at any scale", license="MIT", url="https://github.com/microsoft/torchscale", packages=find_packages(exclude=["*.tests", "*.tests.*", "tests.*", "tests"]), install_requires=["torch>=1.8", "fairscale==0.4.0", "timm==0.6.13"], python_requires=">=3.8.0", classifiers=[ "Programming Language :: Python :: 3", ], )
GeneSplice-main
GeneSplice/torchscale/setup.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/torchscale/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import numpy as np import torch import torch.nn as nn def fixed_pos_embedding(x): seq_len, dim = x.shape inv_freq = 1.0 / (10000 ** (torch.arange(0, dim) / dim)) sinusoid_inp = ( torch.einsum("i , j -> i j", torch.arange(0, seq_len, dtype=torch.float), inv_freq).to(x) ) return torch.sin(sinusoid_inp), torch.cos(sinusoid_inp) def rotate_every_two(x): x1 = x[:, :, ::2] x2 = x[:, :, 1::2] x = torch.stack((-x2, x1), dim=-1) return x.flatten(-2) # in einsum notation: rearrange(x, '... d j -> ... (d j)')\ def duplicate_interleave(m): """ A simple version of `torch.repeat_interleave` for duplicating a matrix while interleaving the copy. """ dim0 = m.shape[0] m = m.view(-1, 1) # flatten the matrix m = m.repeat(1, 2) # repeat all elements into the 2nd dimension m = m.view(dim0, -1) # reshape into a matrix, interleaving the copy return m def apply_rotary_pos_emb(x, sin, cos, scale=1): sin, cos = map(lambda t: duplicate_interleave(t * scale), (sin, cos)) # einsum notation for lambda t: repeat(t[offset:x.shape[1]+offset,:], "n d -> () n () (d j)", j=2) return (x * cos) + (rotate_every_two(x) * sin) class XPOS(nn.Module): def __init__( self, head_dim, scale_base=512 ): super().__init__() self.head_dim = head_dim self.scale_base = scale_base self.register_buffer( "scale", (torch.arange(0, head_dim, 2) + 0.4 * head_dim) / (1.4 * head_dim) ) def forward(self, x, offset=0, downscale=False): length = x.shape[1] min_pos = -(length + offset) // 2 max_pos = length + offset + min_pos scale = self.scale ** torch.arange(min_pos, max_pos, 1).to(self.scale).div(self.scale_base)[:, None] sin, cos = fixed_pos_embedding(scale) if scale.shape[0] > length: scale = scale[-length:] sin = sin[-length:] cos = cos[-length:] if downscale: scale = 1 / scale x = apply_rotary_pos_emb(x, sin, cos, scale) return x
GeneSplice-main
GeneSplice/torchscale/torchscale/component/xpos_relative_position.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import copy import torch import torch.nn as nn def MultiwayWrapper(args, module, dim=1): if args.multiway: return MultiwayNetwork(module, dim=dim) return module def set_split_position(position): def apply_fn(module): if hasattr(module, "split_position"): module.split_position = position return apply_fn class MultiwayNetwork(nn.Module): def __init__(self, module, dim=1): super().__init__() self.dim = dim self.A = module self.B = copy.deepcopy(module) self.B.reset_parameters() self.split_position = -1 def forward(self, x, **kwargs): if self.split_position == -1: return self.A(x, **kwargs) if self.split_position == 0: return self.B(x, **kwargs) x1, x2 = torch.split( x, [self.split_position, x.size(self.dim) - self.split_position], dim=self.dim, ) # x1, x2 = x[:self.split_position], x[self.split_position:] y1, y2 = self.A(x1, **kwargs), self.B(x2, **kwargs) return torch.cat([y1, y2], dim=self.dim) class MutliwayEmbedding(MultiwayNetwork): def __init__(self, modules, dim=1): super(MultiwayNetwork, self).__init__() self.dim = dim assert len(modules) == 2 self.A = modules[0] self.B = modules[1] self.split_position = -1
GeneSplice-main
GeneSplice/torchscale/torchscale/component/multiway_network.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import math import torch import torch.nn.functional as F from torch import nn try: from apex.normalization import FusedLayerNorm as LayerNorm except ModuleNotFoundError: from torch.nn import LayerNorm from .multiway_network import MultiwayWrapper from .xpos_relative_position import XPOS class MultiheadAttention(nn.Module): def __init__( self, args, embed_dim, num_heads, dropout=0.0, self_attention=False, encoder_decoder_attention=False, subln=False, ): super().__init__() self.args = args self.embed_dim = embed_dim self.num_heads = num_heads self.head_dim = embed_dim // num_heads self.scaling = self.head_dim**-0.5 self.self_attention = self_attention self.encoder_decoder_attention = encoder_decoder_attention assert self.self_attention ^ self.encoder_decoder_attention self.k_proj = MultiwayWrapper(args, nn.Linear(embed_dim, embed_dim, bias=True)) self.v_proj = MultiwayWrapper(args, nn.Linear(embed_dim, embed_dim, bias=True)) self.q_proj = MultiwayWrapper(args, nn.Linear(embed_dim, embed_dim, bias=True)) self.out_proj = MultiwayWrapper( args, nn.Linear(embed_dim, embed_dim, bias=True) ) self.inner_attn_ln = ( MultiwayWrapper(args, LayerNorm(self.embed_dim, eps=args.layernorm_eps)) if subln and self.self_attention else None ) self.dropout_module = torch.nn.Dropout(dropout) self.xpos = ( XPOS(self.head_dim, args.xpos_scale_base) if args.xpos_rel_pos and self.self_attention else None ) def reset_parameters(self): nn.init.xavier_uniform_(self.k_proj.weight, gain=1 / math.sqrt(2)) nn.init.xavier_uniform_(self.v_proj.weight, gain=1 / math.sqrt(2)) nn.init.xavier_uniform_(self.q_proj.weight, gain=1 / math.sqrt(2)) nn.init.xavier_uniform_(self.out_proj.weight) nn.init.constant_(self.out_proj.bias, 0.0) def forward( self, query, key, value, incremental_state=None, key_padding_mask=None, attn_mask=None, rel_pos=None, is_first_step=False, ): bsz, tgt_len, embed_dim = query.size() src_len = tgt_len assert embed_dim == self.embed_dim, f"query dim {embed_dim} != {self.embed_dim}" key_bsz, src_len, _ = key.size() assert key_bsz == bsz, f"{query.size(), key.size()}" assert value is not None assert bsz, src_len == value.shape[:2] q = self.q_proj(query) k = self.k_proj(key) v = self.v_proj(value) q *= self.scaling q = q.view(bsz, tgt_len, self.num_heads, self.head_dim).transpose(1, 2) k = k.view(bsz, src_len, self.num_heads, self.head_dim).transpose(1, 2) v = v.view(bsz, src_len, self.num_heads, self.head_dim).transpose(1, 2) q = q.reshape(bsz * self.num_heads, tgt_len, self.head_dim) k = k.reshape(bsz * self.num_heads, src_len, self.head_dim) v = v.reshape(bsz * self.num_heads, src_len, self.head_dim) if incremental_state is not None: if "prev_key" in incremental_state: prev_key = incremental_state["prev_key"].view( bsz * self.num_heads, -1, self.head_dim ) prev_value = incremental_state["prev_value"].view( bsz * self.num_heads, -1, self.head_dim ) k = torch.cat([prev_key, k], dim=1) v = torch.cat([prev_value, v], dim=1) incremental_state["prev_key"] = k.view( bsz, self.num_heads, -1, self.head_dim ) incremental_state["prev_value"] = v.view( bsz, self.num_heads, -1, self.head_dim ) src_len = k.size(1) if self.xpos is not None: if incremental_state is not None and not is_first_step: offset = src_len - 1 else: offset = 0 k = self.xpos(k, offset=0, downscale=True) q = self.xpos(q, offset=offset, downscale=False) attn_weights = torch.bmm(q, k.transpose(1, 2)) if attn_mask is not None: attn_weights = torch.nan_to_num(attn_weights) attn_mask = attn_mask.unsqueeze(0) attn_weights += attn_mask if key_padding_mask is not None: attn_weights = attn_weights.view(bsz, self.num_heads, tgt_len, src_len) attn_weights = attn_weights.masked_fill( key_padding_mask.unsqueeze(1).unsqueeze(2).to(torch.bool), float("-inf"), ) attn_weights = attn_weights.view(bsz * self.num_heads, tgt_len, src_len) if rel_pos is not None: rel_pos = rel_pos.view(attn_weights.size()) attn_weights = attn_weights + rel_pos attn_weights = F.softmax(attn_weights, dim=-1, dtype=torch.float32).type_as( attn_weights ) attn_probs = self.dropout_module(attn_weights) attn = torch.bmm(attn_probs, v) attn = attn.transpose(0, 1).reshape(tgt_len, bsz, embed_dim).transpose(0, 1) if self.inner_attn_ln is not None: attn = self.inner_attn_ln(attn) attn = self.out_proj(attn) attn_weights = attn_weights.view( bsz, self.num_heads, tgt_len, src_len ).transpose(1, 0) return attn, attn_weights
GeneSplice-main
GeneSplice/torchscale/torchscale/component/multihead_attention.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import math import torch import torch.nn as nn class RelativePositionBias(nn.Module): def __init__( self, bidirectional=True, num_buckets=32, max_distance=128, n_heads=12 ): super().__init__() self.bidirectional = bidirectional self.num_buckets = num_buckets self.max_distance = max_distance self.n_heads = n_heads self.relative_attention_bias = nn.Embedding(self.num_buckets, self.n_heads) @staticmethod def _relative_position_bucket( relative_position, bidirectional=True, num_buckets=32, max_distance=128 ): ret = 0 n = -relative_position if bidirectional: num_buckets //= 2 ret += (n < 0).to(torch.long) * num_buckets n = torch.abs(n) else: n = torch.max(n, torch.zeros_like(n)) max_exact = num_buckets // 2 is_small = n < max_exact val_if_large = max_exact + ( torch.log(n.float() / max_exact) / math.log(max_distance / max_exact) * (num_buckets - max_exact) ).to(torch.long) val_if_large = torch.min( val_if_large, torch.full_like(val_if_large, num_buckets - 1) ) ret += torch.where(is_small, n, val_if_large) return ret def compute_bias(self, qlen, klen, step=None): step = 0 if step is None else step context_position = torch.arange( step, step + qlen, dtype=torch.long, device=self.relative_attention_bias.weight.device, )[:, None] memory_position = torch.arange( klen, dtype=torch.long, device=self.relative_attention_bias.weight.device )[None, :] relative_position = memory_position - context_position # shape (qlen, klen) rp_bucket = self._relative_position_bucket( relative_position, # shape (qlen, klen) bidirectional=self.bidirectional, num_buckets=self.num_buckets, max_distance=self.max_distance, ) rp_bucket = rp_bucket.to(self.relative_attention_bias.weight.device) values = self.relative_attention_bias( rp_bucket ) # shape (qlen, klen, num_heads) values = values.permute([2, 0, 1]).unsqueeze( 0 ) # shape (1, num_heads, qlen, klen) return values def forward(self, batch_size, qlen, klen, step=None): # shape (batch * num_heads, qlen, klen) return ( self.compute_bias(qlen, klen, step) .repeat(batch_size, 1, 1, 1) .view(-1, qlen, klen) )
GeneSplice-main
GeneSplice/torchscale/torchscale/component/relative_position_bias.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import torch import torch.nn as nn import torch.nn.functional as F class VisionLanguageEmbedding(nn.Module): def __init__(self, text_embed, vision_embed): super().__init__() self.text_embed = text_embed self.vision_embed = vision_embed def forward(self, textual_tokens, visual_tokens, **kwargs): if textual_tokens is None: return self.vision_embed(visual_tokens) if visual_tokens is None: return self.text_embed(textual_tokens) x1 = self.vision_embed(visual_tokens) x2 = self.text_embed(textual_tokens) return torch.cat([x1, x2], dim=1) class VisionEmbedding(nn.Module): """Image to Patch Embedding""" def __init__( self, img_size=224, patch_size=16, in_chans=3, embed_dim=768, contain_mask_token=False, prepend_cls_token=False, ): super().__init__() img_size = (img_size, img_size) patch_size = (patch_size, patch_size) num_patches = (img_size[1] // patch_size[1]) * (img_size[0] // patch_size[0]) self.patch_shape = (img_size[0] // patch_size[0], img_size[1] // patch_size[1]) self.img_size = img_size self.patch_size = patch_size self.num_patches = num_patches self.proj = nn.Conv2d( in_chans, embed_dim, kernel_size=patch_size, stride=patch_size ) if contain_mask_token: self.mask_token = nn.Parameter(torch.zeros(1, 1, embed_dim)) else: self.mask_token = None if prepend_cls_token: self.cls_token = nn.Parameter(torch.zeros(1, 1, embed_dim)) else: self.cls_token = None def num_position_embeddings(self): if self.cls_token is None: return self.num_patches else: return self.num_patches + 1 def forward(self, x, masked_position=None, **kwargs): B, C, H, W = x.shape assert ( H == self.img_size[0] and W == self.img_size[1] ), f"Input image size ({H}*{W}) doesn't match model ({self.img_size[0]}*{self.img_size[1]})." x = self.proj(x).flatten(2).transpose(1, 2) batch_size, seq_len, _ = x.size() if masked_position is not None: assert self.mask_token is not None mask_token = self.mask_token.expand(batch_size, seq_len, -1) w = masked_position.unsqueeze(-1).type_as(mask_token) x = x * (1 - w) + mask_token * w if self.cls_token is not None: cls_tokens = self.cls_token.expand( batch_size, -1, -1 ) # stole cls_tokens impl from Phil Wang, thanks x = torch.cat((cls_tokens, x), dim=1) return x class TextEmbedding(nn.Embedding): def reset_parameters(self): nn.init.normal_(self.weight, mean=0, std=self.embedding_dim**-0.5) self._fill_padding_idx_with_zero() class PositionalEmbedding(nn.Embedding): def forward( self, x, positions=None, **kwargs, ): if positions is None: # being consistent with Fairseq, which starts from 2. positions = ( torch.arange(2, x.size(1) + 2, device=x.device).long().unsqueeze(0) ) return F.embedding( positions, self.weight, self.padding_idx, self.max_norm, self.norm_type, self.scale_grad_by_freq, self.sparse, )
GeneSplice-main
GeneSplice/torchscale/torchscale/component/embedding.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import torch.nn as nn from timm.models.layers import drop_path class DropPath(nn.Module): """Drop paths (Stochastic Depth) per sample (when applied in main path of residual blocks).""" def __init__(self, drop_prob=None): super(DropPath, self).__init__() self.drop_prob = drop_prob def forward(self, x): return drop_path(x, self.drop_prob, self.training) def extra_repr(self): return "p={}".format(self.drop_prob)
GeneSplice-main
GeneSplice/torchscale/torchscale/component/droppath.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/torchscale/component/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import torch import torch.nn as nn import torch.nn.functional as F try: from apex.normalization import FusedLayerNorm as LayerNorm except ModuleNotFoundError: from torch.nn import LayerNorm from .xmoe.global_groups import get_moe_group class set_torch_seed(object): def __init__(self, seed): assert isinstance(seed, int) self.rng_state = self.get_rng_state() torch.manual_seed(seed) if torch.cuda.is_available(): torch.cuda.manual_seed(seed) def get_rng_state(self): state = {"torch_rng_state": torch.get_rng_state()} if torch.cuda.is_available(): state["cuda_rng_state"] = torch.cuda.get_rng_state() return state def set_rng_state(self, state): torch.set_rng_state(state["torch_rng_state"]) if torch.cuda.is_available(): torch.cuda.set_rng_state(state["cuda_rng_state"]) def __enter__(self): return self def __exit__(self, *exc): self.set_rng_state(self.rng_state) def make_experts(args, embed_dim, expert_ffn_dim): world_size = ( 1 if not torch.distributed.is_initialized() else torch.distributed.get_world_size() ) expert_list = [] ddp_rank = args.ddp_rank start_seed = torch.randint(1000000, (1,)).item() # at least as many experts than gpus if args.moe_expert_count >= world_size: assert ( args.moe_expert_count % world_size == 0 ), f"{args.moe_expert_count}, {world_size}" local_moe_expert_count = args.moe_expert_count // world_size for i in range(local_moe_expert_count): with set_torch_seed(start_seed + ddp_rank * local_moe_expert_count + i): expert_list.append( FeedForwardNetwork( embed_dim, expert_ffn_dim, args.activation_fn, args.dropout, args.activation_dropout, args.layernorm_eps, args.subln, ) ) else: assert ( world_size % args.moe_expert_count == 0 ), f"{world_size}, {args.moe_expert_count}" moe_idx, _ = get_moe_group(args.moe_expert_count) with set_torch_seed(start_seed + moe_idx): expert_list.append( FeedForwardNetwork( embed_dim, expert_ffn_dim, args.activation_fn, args.dropout, args.activation_dropout, args.layernorm_eps, args.subln, ) ) experts = nn.ModuleList(expert_list) return experts def get_activation_fn(activation): if activation == "relu": return F.relu elif activation == "gelu": return F.gelu else: raise NotImplementedError class FeedForwardNetwork(nn.Module): def __init__( self, embed_dim, ffn_dim, activation_fn, dropout, activation_dropout, layernorm_eps, subln=False, ): super().__init__() self.embed_dim = embed_dim self.activation_fn = get_activation_fn(activation=str(activation_fn)) self.activation_dropout_module = torch.nn.Dropout(activation_dropout) self.dropout_module = torch.nn.Dropout(dropout) self.fc1 = nn.Linear(self.embed_dim, ffn_dim) self.fc2 = nn.Linear(ffn_dim, self.embed_dim) self.ffn_layernorm = LayerNorm(ffn_dim, eps=layernorm_eps) if subln else None def reset_parameters(self): self.fc1.reset_parameters() self.fc2.reset_parameters() if self.ffn_layernorm is not None: self.ffn_layernorm.reset_parameters() def forward(self, x): x_shape = x.shape x = x.reshape(-1, x.size(-1)) x = self.fc1(x) x = self.activation_fn(x.float()).type_as(x) x = self.activation_dropout_module(x) if self.ffn_layernorm is not None: x = self.ffn_layernorm(x) x = self.fc2(x) x = x.view(x_shape) x = self.dropout_module(x) return x
GeneSplice-main
GeneSplice/torchscale/torchscale/component/feedforward_network.py
import torch.distributed as dist def _find_my_group_index(grouped_ranks): my_rank = dist.get_rank() for i, group in enumerate(grouped_ranks): if my_rank in group: return i raise RuntimeError def get_moe_group(moe_expert_count=None): if dist.is_initialized(): if not hasattr(get_moe_group, "_moe_groups"): world_size = dist.get_world_size() if world_size <= moe_expert_count: assert moe_expert_count % world_size == 0 moe_groups = [[i] for i in range(world_size)] else: assert world_size % moe_expert_count == 0 ranks_per_group = world_size // moe_expert_count moe_groups = [ [i + j * moe_expert_count for j in range(ranks_per_group)] for i in range(moe_expert_count) ] get_moe_group._moe_expert_count = moe_expert_count get_moe_group._moe_group_idx = moe_groups get_moe_group._moe_groups = [dist.new_group(g) for g in moe_groups] my_group_idx = _find_my_group_index(get_moe_group._moe_group_idx) return my_group_idx, get_moe_group._moe_groups[my_group_idx] def get_all2all_group(moe_expert_count): if dist.is_initialized(): if not hasattr(get_all2all_group, "_all2all_groups"): world_size = dist.get_world_size() # more experts than world size if world_size <= moe_expert_count: assert moe_expert_count % world_size == 0 all2all_groups = [[i for i in range(world_size)]] # larger world than num experts else: assert world_size % moe_expert_count == 0 ranks_per_group = world_size // moe_expert_count all2all_groups = [ [i * moe_expert_count + j for j in range(moe_expert_count)] for i in range(ranks_per_group) ] get_all2all_group._all2all_group_idx = all2all_groups get_all2all_group._all2all_groups = [ dist.new_group(g) for g in all2all_groups ] my_group_idx = _find_my_group_index(get_all2all_group._all2all_group_idx) return get_all2all_group._all2all_groups[my_group_idx]
GeneSplice-main
GeneSplice/torchscale/torchscale/component/xmoe/global_groups.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/torchscale/component/xmoe/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Copyright (c) Facebook, Inc. and its affiliates. All rights reserved. # # This source code is licensed under the BSD license found in the # LICENSE file in the root directory of this source tree. # NOTE: This is a mirror of the code in # https://github.com/facebookresearch/fairscale/tree/master/fairscale/nn/moe import logging import time from typing import Any, Tuple, cast import torch import torch.distributed as dist from torch import Tensor from torch.nn import Module, ModuleList from .global_groups import get_all2all_group, get_moe_group try: from fairseq.modules.moe import MOELayer has_fairseq = True Base = MOELayer except ModuleNotFoundError: Base = Module has_fairseq = False try: # To enable Tutel MoE optimizations: # python3 -m pip install --user --upgrade git+https://github.com/microsoft/[email protected] from tutel import moe as tutel_moe has_tutel, fused_cumsum_sub_one = True, tutel_moe.fast_cumsum_sub_one except ModuleNotFoundError: has_tutel, fused_cumsum_sub_one = False, lambda mask: torch.cumsum(mask, dim=0) - 1 logger = logging.getLogger(__name__) # einsum dimensions: (g)roup, (s)equence, (e)xpert, (m)odel, (c)apacity # See https://arxiv.org/pdf/2006.16668.pdf for details. # Based on https://github.com/pytorch/pytorch/pull/40762 class _AllToAll(torch.autograd.Function): @staticmethod def forward(ctx: Any, group: dist.ProcessGroup, input: Tensor) -> Tensor: # type: ignore ctx.group = group input = input.contiguous() output = torch.empty_like(input) if torch.distributed.is_initialized(): dist.all_to_all_single(output, input, group=group) else: assert group is None output = input return output @staticmethod def backward(ctx: Any, *grad_output: Tensor) -> Tuple[None, Tensor]: return (None, _AllToAll.apply(ctx.group, *grad_output)) class MOELayer(Base): """MOELayer module which implements MixtureOfExperts as described in Gshard_. :: gate = Top2Gate(model_dim, num_experts) moe = MOELayer(gate, expert) output = moe(input) l_aux = moe.l_aux .. Gshard_: https://arxiv.org/pdf/2006.16668.pdf Args: gate (torch.nn.Module): gate network expert (torch.nn.Module): expert network """ def __init__(self, gate, experts, args): if has_fairseq: super(Base, self).__init__() else: super().__init__() self.gate = gate if type(experts) == ModuleList: self.experts = cast(ModuleList, experts) else: self.experts = ModuleList([experts]) _, self.expert_group = get_moe_group(args.moe_expert_count) self.all2all_group = get_all2all_group(args.moe_expert_count) self.world_size = dist.get_world_size(group=self.expert_group) self.all2all_size = dist.get_world_size(group=self.all2all_group) for p in experts.parameters(): p.expert = True # type: ignore self.num_local_experts = len(self.experts) self.args = args self.in_generation = False self.a2a_cuda_event_intervals = [] self.a2a_cpu_time_ms = 0.0 def forward(self, *input: Tensor, input_padding_mask=None, **kwargs: Any) -> Tensor: assert len(input) == 1, "only single input Tensor supported" input = input[0] assert ( len(input.shape) == 3 ), "input Tensor must have dimensions: (s)equence, (t)oken, (m)odel" if input_padding_mask is not None: assert ( len(input_padding_mask.shape) == 2 ), "input Tensor must have dimensions: (s)equence, (t)oken" assert input_padding_mask.shape[0] == input.shape[0] assert input_padding_mask.shape[1] == input.shape[1] # assert input.shape[0] % len(self.experts) == 0, "num tokens must be order of number of local experts" # Implement Algorithm 2 from GShard paper. d_model = input.shape[2] # Pad to expected batch size input_shape = list(input.shape) expected_bsz = ( getattr(self.args, "batch_size", 0) if self.training else getattr(self.args, "batch_size_valid", 0) ) # This indicates that --batch-size or --max-sentences is not specified if expected_bsz is None: expected_bsz = 0 # Note: Padding is not necessary at generation time at present # because all DDP workers process the same batch. Also, batch size at generation time # can be different from that present in the checkpoint state if ( not self.in_generation and expected_bsz != 0 and input_shape[0] != expected_bsz ): logger.warning( f"padding batch with unexpected size {input_shape[0]} (expected: {expected_bsz})" ) assert input_shape[0] < expected_bsz, f"{input_shape[0]} < {expected_bsz}" padded_input = torch.zeros( (expected_bsz, input_shape[1], input_shape[2]), dtype=input.dtype, layout=input.layout, device=input.device, ) padded_input[: input_shape[0], :, :] = input input = padded_input padded_input_padding_mask = torch.ones( ( expected_bsz, input_shape[1], ), dtype=torch.bool, device=input.device, ) if input_padding_mask is not None: padded_input_padding_mask[: input_shape[0], :] = input_padding_mask else: padded_input_padding_mask[: input_shape[0], :] = False input_padding_mask = padded_input_padding_mask # Reshape into S tokens by dropping sequence dimension. reshaped_input = input.reshape(-1, d_model) reshaped_input_shape = reshaped_input.shape reshaped_input_padding_mask = ( input_padding_mask.reshape(-1) if input_padding_mask is not None else None ) # Doing padding here when --max-tokens is specified and not --batch-size or --max-sentences # Pro of --max-tokens: more flexible for MT variable sequence lengths # Con of --max-tokens: extra all-reduce needed to figure out optimal padding without running OOM if expected_bsz == 0: expected_dim = reshaped_input_shape[0] * torch.ones( (1,), dtype=torch.long, device=input.device ) dist.all_reduce(expected_dim, group=dist.group.WORLD, op=dist.ReduceOp.MAX) expected_dim = int(expected_dim.item()) padded_input = torch.zeros( (expected_dim, reshaped_input_shape[1]), dtype=input.dtype, layout=input.layout, device=input.device, ) padded_input[: reshaped_input_shape[0], :] = reshaped_input reshaped_input = padded_input padded_input_padding_mask = torch.ones( (expected_dim,), dtype=torch.bool, device=padded_input.device ) if reshaped_input_padding_mask is not None: padded_input_padding_mask[ : reshaped_input_shape[0] ] = reshaped_input_padding_mask else: padded_input_padding_mask[: reshaped_input_shape[0]] = False reshaped_input_padding_mask = padded_input_padding_mask if has_tutel: l_aux, self.metadata, C, E, indices_, locations_, gates_ = self.gate( reshaped_input, reshaped_input_padding_mask ) S, M = reshaped_input.size(0), reshaped_input.size(1) if not hasattr(self, "_tutel_dispatcher"): self._tutel_dispatcher = tutel_moe.fast_dispatcher( E, C, M, dispatch_dtype=reshaped_input.dtype ) self._tutel_dispatcher.update(indices_, locations_, gates_, capacity=C) dispatched_input = self._tutel_dispatcher.encode(reshaped_input) else: l_aux, combine_weights, dispatch_mask, self.metadata = self.gate( reshaped_input, reshaped_input_padding_mask ) dispatch_mask = dispatch_mask.to(input.dtype).permute( 1, 2, 0 ) # S,E,C -> E,C,S E, C, S = dispatch_mask.size() M = reshaped_input.size(1) assert reshaped_input.size() == (S, M) # einsum("sec,sm->ecm") dispatched_input = torch.mm( dispatch_mask.view(E * C, S), reshaped_input ) # -> (E*C),M if self.all2all_size > 1: dispatched_input = self.all_to_all_wrapper(dispatched_input) # Re-shape after all-to-all: ecm -> gecm dispatched_input = dispatched_input.reshape( self.all2all_size, self.num_local_experts, -1, d_model ) chunks = dispatched_input.chunk(self.num_local_experts, dim=1) expert_outputs = [] for chunk, expert in zip(chunks, self.experts): expert_outputs += [expert(chunk)] expert_output = torch.cat(expert_outputs, dim=1) if self.all2all_size > 1: expert_output = self.all_to_all_wrapper(expert_output) # Re-shape back: gecm -> ecm expert_output = expert_output.reshape( self.all2all_size * self.num_local_experts, -1, d_model ) if has_tutel: combined_output = self._tutel_dispatcher.decode( expert_output.view(E * C, M) ) else: # einsum("sec,ecm->sm") combined_output = combine_weights.view(S, E * C).mm( expert_output.view(E * C, M) ) # Remove padding here when --max-tokens is specified and not --batch-size or --max-sentences combined_output = combined_output[: reshaped_input_shape[0], :] combined_output = combined_output.reshape(input.shape) combined_output = combined_output[: input_shape[0], :, :] self.record_all_to_all_stats() return combined_output, l_aux def prepare_for_inference_(self): self.in_generation = True def all_to_all_wrapper(self, input: Tensor): dummy_a2a = getattr(self.args, "dummy_a2a", False) if dummy_a2a: input = input.contiguous() output = input.detach().clone() return input # always record times, since it is not a lot of overhead # if we do not log it we simply clear it off in record_all_to_all_stats cuda_start = torch.cuda.Event(enable_timing=True) cuda_end = torch.cuda.Event(enable_timing=True) cpu_start = time.time() * 1000 cuda_start.record() output = _AllToAll.apply(self.all2all_group, input) cuda_end.record() cpu_end = time.time() * 1000 self.a2a_cpu_time_ms += cpu_end - cpu_start self.a2a_cuda_event_intervals.append((cuda_start, cuda_end)) return output def record_all_to_all_stats(self): # controlled via an argument as we want to minimize any impact from torch.cuda.synchronize() record_a2a_perf_stats = getattr(self.args, "record_a2a_perf_stats", False) if record_a2a_perf_stats: torch.cuda.synchronize() self.metadata["all_to_all_cpu_time_ms"] = self.a2a_cpu_time_ms a2a_cuda_time_ms = 0.0 for ev_start, ev_end in self.a2a_cuda_event_intervals: a2a_cuda_time_ms += ev_start.elapsed_time(ev_end) self.metadata["all_to_all_cuda_time_ms"] = a2a_cuda_time_ms # reset stats self.a2a_cpu_time_ms = 0.0 self.a2a_cuda_event_intervals = []
GeneSplice-main
GeneSplice/torchscale/torchscale/component/xmoe/moe_layer.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Copyright (c) Facebook, Inc. and its affiliates. All rights reserved. # # This source code is licensed under the BSD license found in the # LICENSE file in the root directory of this source tree. # Implementation of Top2Gating described in https://arxiv.org/pdf/2006.16668.pdf # Code is inspired by Top2GatingOnLogits from lingvo: # https://github.com/tensorflow/lingvo/blob/21b8106c5f1d30a196c98eedc441d4fd70833b11/lingvo/core/moe_layers.py#L477 # NOTE: This is a mirror of the code in # https://github.com/facebookresearch/fairscale/tree/master/fairscale/nn/moe import math from typing import Callable, Dict, Optional, Tuple import torch import torch.nn.functional as F from torch import Tensor from .moe_layer import fused_cumsum_sub_one, has_tutel # use a fixed temperature to compute balance loss TEMPERATURE_FOR_L_UAX = 0.07 # maximum capacity of 1 expert as a fraction of number of tokens in the batch # Note: setting this to 1.0 causes inference to significantly slow down EVAL_CAPACITY_TOKEN_FRACTION = 0.25 # logging SAMPLE_FRACTION = 0.2 def top1gating( logits: torch.Tensor, input_mask: Optional[torch.Tensor] = None, use_fp32=False, capacity_factor=1.0, eval_mode=False, moe_eval_capacity_token_fraction=EVAL_CAPACITY_TOKEN_FRACTION, use_xmoe=False, gate_obj=None, ) -> Tuple[Tensor, Tensor, Tensor, Dict]: """Implements Top2Gating on logits.""" metadata = {} if use_fp32: orig_dtype = logits.dtype logits = logits.float() gates = F.softmax(logits, dim=1) metadata["entropy_gating"] = entropy(probs=gates).mean().detach() # gates has shape of SE num_tokens = gates.shape[0] num_experts = gates.shape[1] if moe_eval_capacity_token_fraction > 0.0 and eval_mode: capacity = math.ceil(moe_eval_capacity_token_fraction * num_tokens) else: # capacity = capacity_factor * S/E capacity = int(capacity_factor * math.ceil(num_tokens / num_experts)) # Create a mask for 1st's expert per token indices1_s = torch.argmax(gates, dim=1) mask1 = one_hot(indices1_s, num_classes=num_experts, unsqueeze_indices=True) if input_mask is not None and input_mask.any(): nonpadding = ~input_mask mask1 = mask1 * nonpadding.unsqueeze(-1).to(mask1.dtype) # for logging (percent of tokens routed to each expert) expert1_hist = ( 100 * torch.histc( (indices1_s.squeeze() + 1), bins=num_experts, min=1, max=num_experts ) / num_tokens ) metadata["unused_expert1_count"] = (expert1_hist == 0).sum() expert1_hist = ( torch.sort(expert1_hist, dim=0, descending=True).values + torch.finfo(torch.float32).tiny ) sample_count = max(math.ceil(num_experts * SAMPLE_FRACTION), 1) metadata["expert1_balance_top"] = expert1_hist[:sample_count].sum() metadata["expert1_balance_bottom"] = expert1_hist[-sample_count:].sum() gates1_s = (gates * mask1).sum(dim=1) # Compute locations in capacity buffer locations1 = fused_cumsum_sub_one(mask1) # Compute l_aux me = torch.mean(gates, dim=0) ce = torch.mean(mask1.to(gates.dtype), dim=0) l_aux = torch.mean(me * ce) l_aux = l_aux * num_experts * num_experts if has_tutel: locations1_s = torch.sum(locations1 * mask1, dim=1) return ( l_aux, metadata, capacity, num_experts, [ indices1_s, ], [ locations1_s, ], [ gates1_s, ], ) # Remove locations outside capacity from mask mask1 = mask1 * torch.lt(locations1, capacity) # Store the capacity location for each token locations1_s = torch.sum(locations1 * mask1, dim=1) # Calculate combine_weights and dispatch_mask gates1 = gates1_s.unsqueeze(-1) * mask1.to(gates1_s.dtype) # einsum("s,se->se") # locations1_sc = num_tokens * capacity locations1_sc = one_hot(locations1_s, num_classes=capacity, unsqueeze_indices=True) combine1_sec = torch.bmm( # einsum("se,sc->sec") gates1.unsqueeze(-1), locations1_sc.to(gates1.dtype).unsqueeze(1), ) dispatch_mask = combine1_sec.bool() if use_fp32: return l_aux, combine1_sec.to(orig_dtype), dispatch_mask, metadata else: return l_aux, combine1_sec, dispatch_mask, metadata class Top1Gate(torch.nn.Module): """Gate module which implements Top2Gating as described in Gshard_. :: gate = Top2Gate(model_dim, num_experts) l_aux, combine_weights, dispatch_mask = gate(input) .. Gshard_: https://arxiv.org/pdf/2006.16668.pdf Args: model_dim (int): size of model embedding dimension num_experts (ints): number of experts in model """ wg: torch.nn.Linear def __init__( self, model_dim: int, num_experts: int, use_fp32=False, input_noise_type=None, capacity_factor=1.0, moe_eval_capacity_token_fraction=EVAL_CAPACITY_TOKEN_FRACTION, use_xmoe=False, ) -> None: # TODO: merge this to top2gate.py # super().__init__() if not use_xmoe: self.wg = torch.nn.Linear(model_dim, num_experts, bias=False) else: self.wg_reduction = torch.nn.Linear(model_dim, 16, bias=False) wg = torch.empty(num_experts, 16) torch.nn.init.orthogonal_(wg, gain=0.32) self.register_parameter("wg", torch.nn.Parameter(wg)) self.use_xmoe = use_xmoe self.use_fp32 = use_fp32 self.input_noise_type = input_noise_type self.capacity_factor = capacity_factor self.moe_eval_capacity_token_fraction = moe_eval_capacity_token_fraction def forward(self, input, mask=None): # type: ignore if self.use_xmoe: input = self.wg_reduction(input) with torch.no_grad(): wg_norm = self.wg.norm(p=2.0, dim=1, keepdim=True) self.wg.mul_(1.5 / wg_norm) logits = self._cosine(input, self.wg) logits = self._make_finite(logits) else: logits = self.wg(input) return top1gating( logits, mask, use_fp32=self.use_fp32, capacity_factor=self.capacity_factor, eval_mode=not self.training, moe_eval_capacity_token_fraction=self.moe_eval_capacity_token_fraction, use_xmoe=self.use_xmoe, gate_obj=self, ) def _make_finite(self, scores): ok = scores.isfinite() if not ok.all(): # NaNs here can break the assignment algorithm scores[~ok] = scores[ok].min() return scores def _get_gating_temperature(self, eps=1e-4): if self.gating_t.data.item() < eps: return eps return self.gating_t def _cosine(self, mat1, mat2, eps=1e-4): assert mat1.dim() == 2 assert mat2.dim() == 2 # mat1 = F.normalize(mat1, p=2.0, dim=1, eps=eps) mat2 = F.normalize(mat2.float(), p=2.0, dim=1, eps=eps) return mat1.float().matmul(mat2.transpose(0, 1)).type_as(mat1) gumbel_map: Dict[torch.device, Callable] = {} def gumbel_rsample(shape: Tuple, device: torch.device) -> Tensor: gumbel = gumbel_map.get(device) if gumbel is None: one = torch.tensor(1.0, device=device) zero = torch.tensor(0.0, device=device) gumbel = torch.distributions.gumbel.Gumbel(zero, one).rsample # type: ignore gumbel_map[device] = gumbel return gumbel(shape) def one_hot(indices: torch.Tensor, num_classes: int, unsqueeze_indices=False) -> Tensor: if unsqueeze_indices: indices = indices.unsqueeze(-1) assert indices.shape[-1] == 1, "last dimension of indices must be have size 1" output = torch.zeros( indices.shape[:-1] + (num_classes,), device=indices.device, dtype=indices.dtype ) output.scatter_(len(output.shape) - 1, indices, 1) return output def entropy(probs): logits = torch.distributions.utils.probs_to_logits(probs) p_log_p = probs * logits return -p_log_p.sum(-1) def top2gating( logits: torch.Tensor, input_mask: Optional[torch.Tensor] = None, use_fp32=False, second_expert_policy="sampling", normalize_gate_prob_before_dropping=False, eval_mode=False, moe_eval_capacity_token_fraction=0.25, batch_prioritized_routing=False, ) -> Tuple[Tensor, Tensor, Tensor]: """Implements Top2Gating on logits.""" metadata = {} if use_fp32: orig_dtype = logits.dtype logits = logits.float() gates = F.softmax(logits, dim=1) metadata["entropy_gating"] = entropy(probs=gates).mean().detach() # gates has shape of SE num_tokens = gates.shape[0] num_experts = gates.shape[1] if moe_eval_capacity_token_fraction > 0.0 and eval_mode: capacity = math.ceil(moe_eval_capacity_token_fraction * num_tokens) else: # capacity = 2S/E capacity = 2 * math.ceil(num_tokens / num_experts) # Create a mask for 1st's expert per token indices1_s = torch.argmax(gates, dim=1, keepdim=True) mask1 = one_hot(indices1_s, num_experts) if second_expert_policy == "sampling": # Create a mask for 2nd's expert per token using Gumbel-max trick # https://timvieira.github.io/blog/post/2014/07/31/gumbel-max-trick/ logits_w_noise = logits + gumbel_rsample(logits.shape, device=logits.device) else: logits_w_noise = logits # Replace top-expert with min value logits_except1 = logits_w_noise.masked_fill(mask1.bool(), float("-inf")) indices2_s = torch.argmax(logits_except1, dim=1, keepdim=True) mask2 = one_hot(indices2_s, num_experts) gates1_s = (gates * mask1).sum(dim=1) gates2_s = (gates * mask2).sum(dim=1) if normalize_gate_prob_before_dropping: # Normalize gate probabilities denom_s = gates1_s + gates2_s # Avoid divide-by-zero denom_s = torch.clamp(denom_s, min=torch.finfo(denom_s.dtype).eps) gates1_s = gates1_s / denom_s gates2_s = gates2_s / denom_s if second_expert_policy == "random": sampled = (2 * gates2_s) > torch.rand_like(gates2_s) mask2 = mask2 * sampled.repeat(num_experts, 1).transpose(1, 0) # Compute locations in capacity buffer if input_mask is not None and input_mask.any(): nonpadding = ~input_mask mask1 = mask1 * nonpadding.unsqueeze(-1).to(mask1.dtype) mask2 = mask2 * nonpadding.unsqueeze(-1).to(mask1.dtype) if batch_prioritized_routing: # if batch_prioritized_routing: importance_scores = -1 * gates.max(dim=1)[0] sorted_mask1 = mask1[importance_scores.argsort(dim=0)] sorted_cumsum1 = fused_cumsum_sub_one(sorted_mask1) * sorted_mask1 importance_sorted_locations1 = sorted_cumsum1[ importance_scores.argsort(dim=0).argsort(dim=0) ] sorted_mask2 = mask2[importance_scores.argsort(dim=0)] sorted_cumsum2 = fused_cumsum_sub_one(sorted_mask2) * sorted_mask2 importance_sorted_locations2 = sorted_cumsum2[ importance_scores.argsort(dim=0).argsort(dim=0) ] importance_sorted_locations2 += torch.sum(mask1, dim=0, keepdim=True) locations1, locations2 = ( importance_sorted_locations1, importance_sorted_locations2, ) else: locations1 = fused_cumsum_sub_one(mask1) locations2 = fused_cumsum_sub_one(mask2) # Update 2nd's location by accounting for locations of 1st locations2 += torch.sum(mask1, dim=0, keepdim=True) # Compute l_aux me = torch.mean(gates, dim=0) ce = torch.mean(mask1.to(gates.dtype), dim=0) l_aux = torch.mean(me * ce) l_aux = l_aux * num_experts * num_experts # for logging purposes metadata["overflow_expert1"] = ( 100 * torch.sum(mask1 * torch.ge(locations1, capacity)) / torch.sum(mask1) ) metadata["overflow_expert2"] = ( 100 * torch.sum(mask2 * torch.ge(locations2, capacity)) / torch.sum(mask2) ) # Remove locations outside capacity from mask mask1_, mask2_ = mask1, mask2 mask1 = mask1 * torch.lt(locations1, capacity) mask2 = mask2 * torch.lt(locations2, capacity) # for logging (percent of tokens routed to each expert) expert1_hist = ( 100 * torch.histc( (indices1_s.squeeze() + 1), bins=num_experts, min=1, max=num_experts ) / num_tokens ) metadata["unused_expert1_count"] = (expert1_hist == 0).sum() expert1_hist = ( torch.sort(expert1_hist, dim=0, descending=True).values + torch.finfo(torch.float32).tiny ) expert2_hist = ( 100 * torch.histc( (indices2_s.squeeze() + 1), bins=num_experts, min=1, max=num_experts ) / num_tokens ) metadata["unused_expert2_count"] = (expert2_hist == 0).sum() expert2_hist = ( torch.sort(expert2_hist, dim=0, descending=True).values + torch.finfo(torch.float32).tiny ) sample_count = max(math.ceil(num_experts * SAMPLE_FRACTION), 1) metadata["expert1_balance_top"] = expert1_hist[:sample_count].sum() metadata["expert1_balance_bottom"] = expert1_hist[-sample_count:].sum() metadata["expert2_balance_top"] = expert2_hist[:sample_count].sum() metadata["expert2_balance_bottom"] = expert2_hist[-sample_count:].sum() if not normalize_gate_prob_before_dropping: # Normalize gate probabilities gates1_s = (gates * mask1).sum(dim=1) gates2_s = (gates * mask2).sum(dim=1) denom_s = gates1_s + gates2_s # Avoid divide-by-zero denom_s = torch.clamp(denom_s, min=torch.finfo(denom_s.dtype).eps) gates1_s /= denom_s gates2_s /= denom_s if has_tutel: locations1_s = torch.sum(locations1 * mask1_, dim=1) locations2_s = torch.sum(locations2 * mask2_, dim=1) return ( l_aux, metadata, capacity, num_experts, [indices1_s, indices2_s], [locations1_s, locations2_s], [gates1_s, gates2_s], ) # Store the capacity location for each token locations1_s = torch.sum(locations1 * mask1, dim=1) locations2_s = torch.sum(locations2 * mask2, dim=1) # Calculate combine_weights and dispatch_mask gates1 = gates1_s.unsqueeze(-1) * mask1.to(gates1_s.dtype) # einsum("s,se->se") gates2 = gates2_s.unsqueeze(-1) * mask2.to(gates2_s.dtype) # einsum("s,se->se") locations1_sc = one_hot(locations1_s, num_classes=capacity, unsqueeze_indices=True) locations2_sc = one_hot(locations2_s, num_classes=capacity, unsqueeze_indices=True) combine1_sec = torch.bmm( # einsum("se,sc->sec") gates1.unsqueeze(-1), locations1_sc.to(gates1.dtype).unsqueeze(1), ) combine2_sec = torch.bmm( # einsum("se,sc->sec") gates2.unsqueeze(-1), locations2_sc.to(gates2.dtype).unsqueeze(1), ) combine_weights = combine1_sec + combine2_sec dispatch_mask = combine_weights.bool() if use_fp32: return l_aux, combine_weights.to(orig_dtype), dispatch_mask, metadata else: return l_aux, combine_weights, dispatch_mask, metadata class Top2Gate(torch.nn.Module): """Gate module which implements Top2Gating as described in Gshard_. :: gate = Top2Gate(model_dim, num_experts) l_aux, combine_weights, dispatch_mask = gate(input) .. Gshard_: https://arxiv.org/pdf/2006.16668.pdf Args: model_dim (int): size of model embedding dimension num_experts (ints): number of experts in model """ wg: torch.nn.Linear def __init__( self, model_dim: int, num_experts: int, use_fp32=False, second_expert_policy="sampling", normalize_gate_prob_before_dropping=False, moe_eval_capacity_token_fraction=0.25, batch_prioritized_routing=False, use_xmoe=False, ) -> None: super().__init__() if not use_xmoe: self.wg = torch.nn.Linear(model_dim, num_experts, bias=False) else: self.wg_reduction = torch.nn.Linear(model_dim, 16, bias=False) wg = torch.empty(num_experts, 16) torch.nn.init.orthogonal_(wg, gain=0.32) self.register_parameter("wg", torch.nn.Parameter(wg)) self.use_fp32 = use_fp32 self.second_expert_policy = second_expert_policy self.normalize_gate_prob_before_dropping = normalize_gate_prob_before_dropping self.moe_eval_capacity_token_fraction = moe_eval_capacity_token_fraction self.batch_prioritized_routing = batch_prioritized_routing self.use_xmoe = use_xmoe def forward(self, input, mask=None): # type: ignore if self.use_xmoe: input = self.wg_reduction(input) with torch.no_grad(): wg_norm = self.wg.norm(p=2.0, dim=1, keepdim=True) self.wg.mul_(1.5 / wg_norm) logits = self._cosine(input, self.wg) logits = self._make_finite(logits) else: logits = self.wg(input) return top2gating( logits, mask, use_fp32=self.use_fp32, second_expert_policy=self.second_expert_policy, normalize_gate_prob_before_dropping=self.normalize_gate_prob_before_dropping, eval_mode=not self.training, moe_eval_capacity_token_fraction=self.moe_eval_capacity_token_fraction, batch_prioritized_routing=self.batch_prioritized_routing, ) def _cosine(self, mat1, mat2, eps=1e-4): assert mat1.dim() == 2 assert mat2.dim() == 2 # mat1 = F.normalize(mat1, p=2.0, dim=1, eps=eps) mat2 = F.normalize(mat2.float(), p=2.0, dim=1, eps=eps) return mat1.float().matmul(mat2.transpose(0, 1)).type_as(mat1) def _make_finite(self, scores): ok = scores.isfinite() if not ok.all(): # NaNs here can break the assignment algorithm scores[~ok] = scores[ok].min() return scores
GeneSplice-main
GeneSplice/torchscale/torchscale/component/xmoe/routing.py
import math import numpy as np import torch import torch.nn as nn from fairscale.nn import checkpoint_wrapper, wrap from torchscale.architecture.utils import init_bert_params from torchscale.component.droppath import DropPath from torchscale.component.feedforward_network import FeedForwardNetwork, make_experts # from torchscale.component.multihead_attention import MultiheadAttention from LongNet.attention import DilatedAttention from torchscale.component.relative_position_bias import RelativePositionBias from torchscale.component.xmoe.moe_layer import MOELayer from torchscale.component.xmoe.routing import Top1Gate, Top2Gate try: from apex.normalization import FusedLayerNorm as LayerNorm except ModuleNotFoundError: from torch.nn import LayerNorm class DecoderLayer(nn.Module): def __init__( self, args, depth, is_moe_layer=False, is_encoder_decoder=False, ): super().__init__() self.args = args self.embed_dim = args.decoder_embed_dim self.dropout_module = torch.nn.Dropout(args.dropout) if args.drop_path_rate > 0: drop_path_prob = np.linspace(0, args.drop_path_rate, args.decoder_layers)[ depth ] self.drop_path = DropPath(drop_path_prob) else: self.drop_path = None self.self_attn = self.build_self_attention(self.embed_dim, args) self.normalize_before = args.decoder_normalize_before self.self_attn_layer_norm = LayerNorm(self.embed_dim, eps=args.layernorm_eps) if not is_encoder_decoder: self.encoder_attn = None self.encoder_attn_layer_norm = None else: self.encoder_attn = self.build_encoder_attention(self.embed_dim, args) self.encoder_attn_layer_norm = LayerNorm(self.embed_dim, eps=args.layernorm_eps) self.is_moe_layer = is_moe_layer self.ffn_dim = args.decoder_ffn_embed_dim if not self.is_moe_layer: self.ffn = self.build_ffn( self.embed_dim, self.args, ) else: if args.moe_top1_expert: gate = Top1Gate( self.embed_dim, args.moe_expert_count, use_fp32=args.moe_gating_use_fp32, moe_eval_capacity_token_fraction=args.moe_eval_capacity_token_fraction, use_xmoe=args.use_xmoe, ) else: gate = Top2Gate( self.embed_dim, args.moe_expert_count, args.moe_gating_use_fp32, args.moe_second_expert_policy, args.moe_normalize_gate_prob_before_dropping, args.moe_eval_capacity_token_fraction, use_xmoe=args.use_xmoe, ) experts = make_experts(args, self.embed_dim, self.ffn_dim) self.moe_layer = MOELayer(gate, experts, args) self.final_layer_norm = LayerNorm(self.embed_dim, eps=args.layernorm_eps) if args.deepnorm: if is_encoder_decoder: self.alpha = math.pow(3.0 * args.decoder_layers, 0.25) else: self.alpha = math.pow(2.0 * args.decoder_layers, 0.25) else: self.alpha = 1.0 def build_ffn(self, embed_dim, args): return FeedForwardNetwork( embed_dim, self.ffn_dim, args.activation_fn, args.dropout, args.activation_dropout, args.layernorm_eps, args.subln, ) def build_self_attention(self, embed_dim, args): return DilatedAttention( embed_dim, args.decoder_attention_heads, args.decoder_dilation_rate, args.decoder_segment_size, dropout=args.attention_dropout, #specify dropout ) def build_encoder_attention(self, embed_dim, args): return DilatedAttention( embed_dim, args.decoder_attention_heads, args.decoder_dilation_rate, args.decoder_segment_size, dropout=args.attention_dropout, #specify dropout ) def residual_connection(self, x, residual): return residual * self.alpha + x def forward( self, x, encoder_out=None, encoder_padding_mask=None, incremental_state=None, self_attn_mask=None, self_attn_padding_mask=None, self_attn_rel_pos=None, cross_attn_rel_pos=None, is_first_step=False, ): residual = x if self.normalize_before: x = self.self_attn_layer_norm(x) x, attn = self.self_attn( query=x, key=x, value=x, key_padding_mask=self_attn_padding_mask, incremental_state=incremental_state, attn_mask=self_attn_mask, rel_pos=self_attn_rel_pos, is_first_step=is_first_step, ) x = self.dropout_module(x) if self.drop_path is not None: x = self.drop_path(x) x = self.residual_connection(x, residual) if not self.normalize_before: x = self.self_attn_layer_norm(x) if self.encoder_attn is not None and encoder_out is not None: residual = x if self.normalize_before: x = self.encoder_attn_layer_norm(x) x, attn = self.encoder_attn( query=x, key=encoder_out, value=encoder_out, key_padding_mask=encoder_padding_mask, incremental_state=None, rel_pos=cross_attn_rel_pos, ) x = self.dropout_module(x) if self.drop_path is not None: x = self.drop_path(x) x = self.residual_connection(x, residual) if not self.normalize_before: x = self.encoder_attn_layer_norm(x) residual = x if self.normalize_before: x = self.final_layer_norm(x) if not self.is_moe_layer: x = self.ffn(x) l_aux = None else: x, l_aux = self.moe_layer(x) if self.drop_path is not None: x = self.drop_path(x) x = self.residual_connection(x, residual) if not self.normalize_before: x = self.final_layer_norm(x) return x, attn, None, l_aux class Decoder(nn.Module): def __init__( self, args, embed_tokens=None, embed_positions=None, output_projection=None, is_encoder_decoder=False, **kwargs ): super().__init__(**kwargs) self.args = args self.dropout_module = torch.nn.Dropout(args.dropout) embed_dim = args.decoder_embed_dim self.embed_dim = embed_dim self.embed_scale = 1.0 if args.no_scale_embedding else math.sqrt(embed_dim) self.embed_tokens = embed_tokens self.embed_positions = embed_positions if ( output_projection is None and not args.no_output_layer and args.vocab_size > 0 ): self.output_projection = self.build_output_projection(args) else: self.output_projection = output_projection if args.layernorm_embedding: self.layernorm_embedding = LayerNorm(embed_dim, eps=args.layernorm_eps) else: self.layernorm_embedding = None self.layers = nn.ModuleList([]) moe_freq = args.moe_freq for i in range(args.decoder_layers): is_moe_layer = moe_freq != 0 and (i + 1) % moe_freq == 0 self.layers.append( self.build_decoder_layer( args, depth=i, is_moe_layer=is_moe_layer, is_encoder_decoder=is_encoder_decoder, ) ) self.num_layers = len(self.layers) if args.decoder_normalize_before: self.layer_norm = LayerNorm(embed_dim, eps=args.layernorm_eps) else: self.layer_norm = None self.self_attn_relative_position = None self.cross_attn_relative_position = None if args.rel_pos_buckets > 0 and args.max_rel_pos > 0: self.self_attn_relative_position = RelativePositionBias( num_buckets=args.rel_pos_buckets, max_distance=args.max_rel_pos, n_heads=args.decoder_attention_heads, ) if is_encoder_decoder: self.cross_attn_relative_position = RelativePositionBias( num_buckets=args.rel_pos_buckets, max_distance=args.max_rel_pos, n_heads=args.decoder_attention_heads, ) if args.bert_init: self.apply(init_bert_params) if args.deepnorm: if is_encoder_decoder: init_scale = math.pow(12.0 * args.decoder_layers, 0.25) else: init_scale = math.pow(8.0 * args.decoder_layers, 0.25) for name, p in self.named_parameters(): if ( "fc1" in name or "fc2" in name or "out_proj" in name or "v_proj" in name ): p.data.div_(init_scale) if args.subln: if is_encoder_decoder: init_scale = math.sqrt(math.log(args.decoder_layers * 3)) else: init_scale = math.sqrt(math.log(args.decoder_layers * 2)) for name, p in self.named_parameters(): if "encoder_attn" in name: continue if ( "fc1" in name or "fc2" in name or "out_proj" in name or "v_proj" in name ): p.data.mul_(init_scale) def build_output_projection( self, args, ): if args.share_decoder_input_output_embed: output_projection = torch.nn.Linear( self.embed_tokens.weight.shape[1], self.embed_tokens.weight.shape[0], bias=False, ) output_projection.weight = self.embed_tokens.weight else: output_projection = torch.nn.Linear( args.decoder_embed_dim, args.vocab_size, bias=False ) torch.nn.init.normal_( output_projection.weight, mean=0, std=args.decoder_embed_dim**-0.5 ) return output_projection def build_decoder_layer( self, args, depth, is_moe_layer=False, is_encoder_decoder=False ): layer = DecoderLayer( args, depth, is_moe_layer=is_moe_layer, is_encoder_decoder=is_encoder_decoder, ) if args.checkpoint_activations: layer = checkpoint_wrapper(layer) if args.fsdp: layer = wrap(layer) return layer def forward_embedding( self, tokens, token_embedding=None, incremental_state=None, ): positions = None if self.embed_positions is not None: positions = self.embed_positions( tokens, incremental_state=incremental_state ) if incremental_state is not None and not self.is_first_step(incremental_state): tokens = tokens[:, -1:] if positions is not None: positions = positions[:, -1:] if token_embedding is None: token_embedding = self.embed_tokens(tokens) x = embed = self.embed_scale * token_embedding if positions is not None: x += positions if self.layernorm_embedding is not None: x = self.layernorm_embedding(x) x = self.dropout_module(x) return x, embed def is_first_step(self, incremental_state): if incremental_state is None: return False return incremental_state.get("is_first_step", False) def forward( self, prev_output_tokens, self_attn_padding_mask=None, encoder_out=None, incremental_state=None, features_only=False, return_all_hiddens=False, token_embeddings=None, **kwargs ): # embed tokens and positions x, _ = self.forward_embedding( prev_output_tokens, token_embeddings, incremental_state ) is_first_step = self.is_first_step(incremental_state) # relative position self_attn_rel_pos_bias = None slen = prev_output_tokens.size(1) if self.self_attn_relative_position is not None: self_attn_rel_pos_bias = self.self_attn_relative_position( batch_size=x.size(0), qlen=slen, klen=slen ) if incremental_state is not None and not is_first_step: self_attn_rel_pos_bias = self_attn_rel_pos_bias[-1:, :, :] cross_attn_rel_pos_bias = None if self.cross_attn_relative_position is not None: cross_attn_rel_pos_bias = self.cross_attn_relative_position( batch_size=x.size(0), qlen=slen, klen=encoder_out["encoder_out"].size(1), ) if incremental_state is not None and not is_first_step: cross_attn_rel_pos_bias = cross_attn_rel_pos_bias[-1:, :, :] # decoder layers inner_states = [x] if encoder_out is None: l_aux = [] else: l_aux = encoder_out["l_aux"] if "l_aux" in encoder_out else [] for idx, layer in enumerate(self.layers): if incremental_state is None or is_first_step: self_attn_mask = torch.triu( torch.zeros([x.size(1), x.size(1)]) .float() .fill_(float("-inf")) .type_as(x), 1, ) if is_first_step and incremental_state is not None: if idx not in incremental_state: incremental_state[idx] = {} else: self_attn_mask = None if idx not in incremental_state: incremental_state[idx] = {} x, layer_attn, _, l_aux_i = layer( x, encoder_out["encoder_out"] if encoder_out is not None else None, encoder_out["encoder_padding_mask"] if encoder_out is not None else None, incremental_state[idx] if incremental_state is not None else None, self_attn_mask=self_attn_mask, self_attn_padding_mask=self_attn_padding_mask, self_attn_rel_pos=self_attn_rel_pos_bias, cross_attn_rel_pos=cross_attn_rel_pos_bias, is_first_step=is_first_step, ) l_aux.append(l_aux_i) inner_states.append(x) if self.layer_norm is not None: x = self.layer_norm(x) if not features_only: x = self.output_layer(x) return x, { "inner_states": inner_states, "l_aux": l_aux, "attn": None, } def output_layer(self, features): return self.output_projection(features)
GeneSplice-main
GeneSplice/torchscale/torchscale/architecture/decoder.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] class EncoderConfig(object): def __init__(self, **kwargs): self.encoder_embed_dim = kwargs.pop("encoder_embed_dim", 768) self.encoder_attention_heads = kwargs.pop("encoder_attention_heads", 12) self.encoder_ffn_embed_dim = kwargs.pop("encoder_ffn_embed_dim", 3072) self.encoder_layers = kwargs.pop("encoder_layers", 12) self.encoder_normalize_before = kwargs.pop("encoder_normalize_before", True) self.normalize_output = kwargs.pop("normalize_output", True) self.activation_fn = kwargs.pop("activation_fn", "gelu") self.dropout = kwargs.pop("dropout", 0.0) self.drop_path_rate = kwargs.pop("drop_path_rate", 0.0) self.attention_dropout = kwargs.pop("attention_dropout", 0.0) self.activation_dropout = kwargs.pop("activation_dropout", 0.0) self.no_scale_embedding = kwargs.pop("no_scale_embedding", True) self.layernorm_embedding = kwargs.pop("layernorm_embedding", False) self.moe_freq = kwargs.pop("moe_freq", 0) self.moe_top1_expert = kwargs.pop("moe_top1_expert", False) self.moe_expert_count = kwargs.pop("moe_expert_count", 0) self.moe_gating_use_fp32 = kwargs.pop("moe_gating_use_fp32", True) self.moe_eval_capacity_token_fraction = kwargs.pop( "moe_eval_capacity_token_fraction", 0.25 ) self.moe_second_expert_policy = kwargs.pop("moe_second_expert_policy", "random") self.moe_normalize_gate_prob_before_dropping = kwargs.pop( "moe_normalize_gate_prob_before_dropping", False ) self.use_xmoe = kwargs.pop("use_xmoe", False) self.rel_pos_buckets = kwargs.pop("rel_pos_buckets", 0) self.max_rel_pos = kwargs.pop("max_rel_pos", 0) self.deepnorm = kwargs.pop("deepnorm", False) self.subln = kwargs.pop("subln", True) self.bert_init = kwargs.pop("bert_init", False) self.multiway = kwargs.pop("multiway", False) self.share_encoder_input_output_embed = kwargs.pop( "share_encoder_input_output_embed", False ) self.max_source_positions = kwargs.pop("max_source_positions", 1024) self.no_output_layer = kwargs.pop("no_output_layer", False) self.layernorm_eps = kwargs.pop("layernorm_eps", 1e-5) # Text self.vocab_size = kwargs.pop("vocab_size", -1) # Vision self.img_size = kwargs.pop("img_size", 224) self.patch_size = kwargs.pop("patch_size", 16) self.in_chans = kwargs.pop("in_chans", 3) # Fairscale self.checkpoint_activations = kwargs.pop("checkpoint_activations", False) self.fsdp = kwargs.pop("fsdp", False) self.ddp_rank = kwargs.pop("ddp_rank", 0) self.xpos_rel_pos = kwargs.pop("xpos_rel_pos", False) self.xpos_scale_base = kwargs.pop("xpos_scale_base", 512) if self.deepnorm: self.encoder_normalize_before = False self.subln = False if self.subln: self.encoder_normalize_before = True self.deepnorm = False if self.use_xmoe: self.moe_normalize_gate_prob_before_dropping = True self.moe_second_expert_policy = "random" assert self.moe_freq > 0 and self.moe_expert_count > 0 def override(self, args): for hp in self.__dict__.keys(): if getattr(args, hp, None) is not None: self.__dict__[hp] = getattr(args, hp, None) class DecoderConfig(object): def __init__(self, **kwargs): self.decoder_embed_dim = kwargs.pop("decoder_embed_dim", 768) self.decoder_attention_heads = kwargs.pop("decoder_attention_heads", 12) self.decoder_ffn_embed_dim = kwargs.pop("decoder_ffn_embed_dim", 3072) self.decoder_layers = kwargs.pop("decoder_layers", 12) self.decoder_normalize_before = kwargs.pop("decoder_normalize_before", True) self.activation_fn = kwargs.pop("activation_fn", "gelu") self.dropout = kwargs.pop("dropout", 0.0) self.drop_path_rate = kwargs.pop("drop_path_rate", 0.0) self.attention_dropout = kwargs.pop("attention_dropout", 0.0) self.activation_dropout = kwargs.pop("activation_dropout", 0.0) self.no_scale_embedding = kwargs.pop("no_scale_embedding", True) self.layernorm_embedding = kwargs.pop("layernorm_embedding", False) self.moe_freq = kwargs.pop("moe_freq", 0) self.moe_top1_expert = kwargs.pop("moe_top1_expert", False) self.moe_expert_count = kwargs.pop("moe_expert_count", 0) self.moe_gating_use_fp32 = kwargs.pop("moe_gating_use_fp32", True) self.moe_eval_capacity_token_fraction = kwargs.pop( "moe_eval_capacity_token_fraction", 0.25 ) self.moe_second_expert_policy = kwargs.pop("moe_second_expert_policy", "random") self.moe_normalize_gate_prob_before_dropping = kwargs.pop( "moe_normalize_gate_prob_before_dropping", False ) self.use_xmoe = kwargs.pop("use_xmoe", False) self.rel_pos_buckets = kwargs.pop("rel_pos_buckets", 0) self.max_rel_pos = kwargs.pop("max_rel_pos", 0) self.deepnorm = kwargs.pop("deepnorm", False) self.subln = kwargs.pop("subln", True) self.bert_init = kwargs.pop("bert_init", False) self.multiway = kwargs.pop("multiway", False) self.share_decoder_input_output_embed = kwargs.pop( "share_decoder_input_output_embed", False ) self.max_target_positions = kwargs.pop("max_target_positions", 1024) self.no_output_layer = kwargs.pop("no_output_layer", False) self.layernorm_eps = kwargs.pop("layernorm_eps", 1e-5) # Text self.vocab_size = kwargs.pop("vocab_size", -1) # Fairscale self.checkpoint_activations = kwargs.pop("checkpoint_activations", False) self.fsdp = kwargs.pop("fsdp", False) self.ddp_rank = kwargs.pop("ddp_rank", 0) self.xpos_rel_pos = kwargs.pop("xpos_rel_pos", False) self.xpos_scale_base = kwargs.pop("xpos_scale_base", 512) if self.deepnorm: self.decoder_normalize_before = False self.subln = False if self.subln: self.decoder_normalize_before = True self.deepnorm = False if self.use_xmoe: self.moe_normalize_gate_prob_before_dropping = True self.moe_second_expert_policy = "random" assert self.moe_freq > 0 and self.moe_expert_count > 0 def override(self, args): for hp in self.__dict__.keys(): if getattr(args, hp, None) is not None: self.__dict__[hp] = getattr(args, hp, None) class EncoderDecoderConfig(object): def __init__(self, **kwargs): self.encoder_embed_dim = kwargs.pop("encoder_embed_dim", 768) self.encoder_attention_heads = kwargs.pop("encoder_attention_heads", 12) self.encoder_ffn_embed_dim = kwargs.pop("encoder_ffn_embed_dim", 3072) self.encoder_layers = kwargs.pop("encoder_layers", 12) self.encoder_normalize_before = kwargs.pop("encoder_normalize_before", True) self.decoder_embed_dim = kwargs.pop("decoder_embed_dim", 768) self.decoder_attention_heads = kwargs.pop("decoder_attention_heads", 12) self.decoder_ffn_embed_dim = kwargs.pop("decoder_ffn_embed_dim", 3072) self.decoder_layers = kwargs.pop("decoder_layers", 12) self.decoder_normalize_before = kwargs.pop("decoder_normalize_before", True) self.activation_fn = kwargs.pop("activation_fn", "gelu") self.dropout = kwargs.pop("dropout", 0.0) self.drop_path_rate = kwargs.pop("drop_path_rate", 0.0) self.attention_dropout = kwargs.pop("attention_dropout", 0.0) self.activation_dropout = kwargs.pop("activation_dropout", 0.0) self.no_scale_embedding = kwargs.pop("no_scale_embedding", True) self.layernorm_embedding = kwargs.pop("layernorm_embedding", False) self.moe_freq = kwargs.pop("moe_freq", 0) self.moe_top1_expert = kwargs.pop("moe_top1_expert", False) self.moe_expert_count = kwargs.pop("moe_expert_count", 0) self.moe_gating_use_fp32 = kwargs.pop("moe_gating_use_fp32", True) self.moe_eval_capacity_token_fraction = kwargs.pop( "moe_eval_capacity_token_fraction", 0.25 ) self.moe_second_expert_policy = kwargs.pop("moe_second_expert_policy", "random") self.moe_normalize_gate_prob_before_dropping = kwargs.pop( "moe_normalize_gate_prob_before_dropping", False ) self.use_xmoe = kwargs.pop("use_xmoe", False) self.rel_pos_buckets = kwargs.pop("rel_pos_buckets", 0) self.max_rel_pos = kwargs.pop("max_rel_pos", 0) self.deepnorm = kwargs.pop("deepnorm", False) self.subln = kwargs.pop("subln", True) self.bert_init = kwargs.pop("bert_init", False) self.multiway = kwargs.pop("multiway", False) self.share_all_embeddings = kwargs.pop("share_all_embeddings", False) self.share_decoder_input_output_embed = kwargs.pop( "share_decoder_input_output_embed", False ) self.max_source_positions = kwargs.pop("max_source_positions", 1024) self.max_target_positions = kwargs.pop("max_target_positions", 1024) self.no_output_layer = kwargs.pop("no_output_layer", False) self.layernorm_eps = kwargs.pop("layernorm_eps", 1e-5) # Text self.vocab_size = kwargs.pop("vocab_size", -1) # Fairscale self.checkpoint_activations = kwargs.pop("checkpoint_activations", False) self.fsdp = kwargs.pop("fsdp", False) self.ddp_rank = kwargs.pop("ddp_rank", 0) self.xpos_rel_pos = kwargs.pop("xpos_rel_pos", False) self.xpos_scale_base = kwargs.pop("xpos_scale_base", 512) if self.deepnorm: self.encoder_normalize_before = False self.decoder_normalize_before = False self.subln = False if self.subln: self.encoder_normalize_before = True self.decoder_normalize_before = True self.deepnorm = False if self.use_xmoe: self.moe_normalize_gate_prob_before_dropping = True self.moe_second_expert_policy = "random" assert self.moe_freq > 0 and self.moe_expert_count > 0 def override(self, args): for hp in self.__dict__.keys(): if getattr(args, hp, None) is not None: self.__dict__[hp] = getattr(args, hp, None)
GeneSplice-main
GeneSplice/torchscale/torchscale/architecture/config.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import torch.nn as nn from torchscale.architecture.decoder import Decoder from torchscale.architecture.encoder import Encoder class EncoderDecoder(nn.Module): def __init__( self, args, encoder_embed_tokens=None, encoder_embed_positions=None, decoder_embed_tokens=None, decoder_embed_positions=None, output_projection=None, **kwargs ): super().__init__() self.args = args if args.share_all_embeddings: args.share_decoder_input_output_embed = True self.encoder = Encoder( args, encoder_embed_tokens, encoder_embed_positions, is_encoder_decoder=True, **kwargs ) if args.share_all_embeddings and decoder_embed_tokens is None: decoder_embed_tokens = self.encoder.embed_tokens self.decoder = Decoder( args, decoder_embed_tokens, decoder_embed_positions, output_projection, is_encoder_decoder=True, **kwargs ) def forward( self, src_tokens, prev_output_tokens, return_all_hiddens=False, features_only=False, **kwargs ): encoder_out = self.encoder(src_tokens, return_all_hiddens=return_all_hiddens) decoder_out = self.decoder( prev_output_tokens, encoder_out=encoder_out, features_only=features_only, return_all_hiddens=return_all_hiddens, ) return decoder_out
GeneSplice-main
GeneSplice/torchscale/torchscale/architecture/encoder_decoder.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/torchscale/architecture/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import math import numpy as np import torch import torch.nn as nn from fairscale.nn import checkpoint_wrapper, wrap try: from apex.normalization import FusedLayerNorm as LayerNorm except ModuleNotFoundError: from torch.nn import LayerNorm from torchscale.architecture.utils import init_bert_params from torchscale.component.droppath import DropPath from torchscale.component.feedforward_network import FeedForwardNetwork, make_experts from torchscale.component.multihead_attention import MultiheadAttention from torchscale.component.multiway_network import MultiwayWrapper, set_split_position from torchscale.component.relative_position_bias import RelativePositionBias from torchscale.component.xmoe.moe_layer import MOELayer from torchscale.component.xmoe.routing import Top1Gate, Top2Gate class EncoderLayer(nn.Module): def __init__(self, args, depth, is_moe_layer=False, is_encoder_decoder=False): super().__init__() self.args = args self.embed_dim = args.encoder_embed_dim self.self_attn = self.build_self_attention(self.embed_dim, args) self.self_attn_layer_norm = MultiwayWrapper(args, LayerNorm(self.embed_dim, eps=args.layernorm_eps)) self.dropout_module = torch.nn.Dropout(args.dropout) if args.drop_path_rate > 0: drop_path_prob = np.linspace(0, args.drop_path_rate, args.encoder_layers)[ depth ] self.drop_path = DropPath(drop_path_prob) else: self.drop_path = None self.normalize_before = args.encoder_normalize_before self.is_moe_layer = is_moe_layer self.ffn_dim = args.encoder_ffn_embed_dim if not self.is_moe_layer: self.ffn = MultiwayWrapper( args, self.build_ffn( self.embed_dim, self.args, ), ) else: assert not self.args.multiway if args.moe_top1_expert: gate = Top1Gate( self.embed_dim, args.moe_expert_count, use_fp32=args.moe_gating_use_fp32, moe_eval_capacity_token_fraction=args.moe_eval_capacity_token_fraction, use_xmoe=args.use_xmoe, ) else: gate = Top2Gate( self.embed_dim, args.moe_expert_count, args.moe_gating_use_fp32, args.moe_second_expert_policy, args.moe_normalize_gate_prob_before_dropping, args.moe_eval_capacity_token_fraction, use_xmoe=args.use_xmoe, ) experts = make_experts(args, self.embed_dim, self.ffn_dim) self.moe_layer = MOELayer(gate, experts, args) self.final_layer_norm = MultiwayWrapper(args, LayerNorm(self.embed_dim, eps=args.layernorm_eps)) if args.deepnorm: if is_encoder_decoder: self.alpha = ( math.pow( math.pow(args.encoder_layers, 4) * args.decoder_layers, 0.0625 ) * 0.81 ) else: self.alpha = math.pow(2.0 * args.encoder_layers, 0.25) else: self.alpha = 1.0 def build_ffn(self, embed_dim, args): return FeedForwardNetwork( embed_dim, self.ffn_dim, args.activation_fn, args.dropout, args.activation_dropout, args.layernorm_eps, args.subln, ) def build_self_attention(self, embed_dim, args): return MultiheadAttention( args, embed_dim, args.encoder_attention_heads, dropout=args.attention_dropout, self_attention=True, encoder_decoder_attention=False, subln=args.subln, ) def residual_connection(self, x, residual): return residual * self.alpha + x def forward(self, x, encoder_padding_mask, attn_mask=None, rel_pos=None, multiway_split_position=None, incremental_state=None): if multiway_split_position is not None: assert self.args.multiway self.apply(set_split_position(multiway_split_position)) if attn_mask is not None: attn_mask = attn_mask.masked_fill(attn_mask.to(torch.bool), -1e8) residual = x if self.normalize_before: x = self.self_attn_layer_norm(x) x, _ = self.self_attn( query=x, key=x, value=x, key_padding_mask=encoder_padding_mask, attn_mask=attn_mask, rel_pos=rel_pos, incremental_state=incremental_state, ) x = self.dropout_module(x) if self.drop_path is not None: x = self.drop_path(x) x = self.residual_connection(x, residual) if not self.normalize_before: x = self.self_attn_layer_norm(x) residual = x if self.normalize_before: x = self.final_layer_norm(x) if not self.is_moe_layer: x = self.ffn(x) l_aux = None else: x = x.transpose(0, 1) x, l_aux = self.moe_layer(x) x = x.transpose(0, 1) if self.drop_path is not None: x = self.drop_path(x) x = self.residual_connection(x, residual) if not self.normalize_before: x = self.final_layer_norm(x) return x, l_aux class Encoder(nn.Module): def __init__( self, args, embed_tokens=None, embed_positions=None, output_projection=None, is_encoder_decoder=False, **kwargs ): self.args = args super().__init__(**kwargs) self.dropout_module = torch.nn.Dropout(args.dropout) embed_dim = args.encoder_embed_dim self.embed_scale = 1.0 if args.no_scale_embedding else math.sqrt(embed_dim) self.embed_tokens = embed_tokens self.embed_positions = embed_positions if ( output_projection is None and not is_encoder_decoder and not args.no_output_layer and args.vocab_size > 0 ): self.output_projection = self.build_output_projection(args) else: self.output_projection = output_projection if args.layernorm_embedding: self.layernorm_embedding = MultiwayWrapper( args, LayerNorm(embed_dim, eps=args.layernorm_eps), dim=1 ) else: self.layernorm_embedding = None self.layers = nn.ModuleList([]) moe_freq = args.moe_freq for i in range(args.encoder_layers): is_moe_layer = moe_freq != 0 and (i + 1) % moe_freq == 0 self.layers.append( self.build_encoder_layer( args, depth=i, is_moe_layer=is_moe_layer, is_encoder_decoder=is_encoder_decoder, ) ) self.num_layers = len(self.layers) if args.encoder_normalize_before and args.normalize_output: self.layer_norm = MultiwayWrapper(args, LayerNorm(embed_dim, eps=args.layernorm_eps)) else: self.layer_norm = None if args.rel_pos_buckets > 0 and args.max_rel_pos > 0: self.relative_position = RelativePositionBias( num_buckets=args.rel_pos_buckets, max_distance=args.max_rel_pos, n_heads=args.encoder_attention_heads, ) else: self.relative_position = None if args.bert_init: self.apply(init_bert_params) if args.deepnorm: if is_encoder_decoder: init_scale = ( math.pow( math.pow(args.encoder_layers, 4) * args.decoder_layers, 0.0625 ) / 1.15 ) else: init_scale = math.pow(8.0 * args.encoder_layers, 0.25) for name, p in self.named_parameters(): if ( "fc1" in name or "fc2" in name or "out_proj" in name or "v_proj" in name ): p.data.div_(init_scale) if args.subln: if is_encoder_decoder: init_scale = math.sqrt( math.log(3 * args.decoder_layers) * math.log(2 * args.encoder_layers) / 3 ) else: init_scale = math.sqrt(math.log(args.encoder_layers * 2)) for name, p in self.named_parameters(): if ( "fc1" in name or "fc2" in name or "out_proj" in name or "v_proj" in name ): p.data.mul_(init_scale) def build_output_projection( self, args, ): if args.share_encoder_input_output_embed: assert args.encoder_embedding_type == "language" output_projection = torch.nn.Linear( self.embed_tokens.weight.shape[1], self.embed_tokens.weight.shape[0], bias=False, ) output_projection.weight = self.embed_tokens.weight else: output_projection = torch.nn.Linear( args.encoder_embed_dim, args.vocab_size, bias=False ) torch.nn.init.normal_( output_projection.weight, mean=0, std=args.encoder_embed_dim**-0.5 ) return output_projection def build_encoder_layer( self, args, depth, is_moe_layer=False, is_encoder_decoder=False ): layer = EncoderLayer( args, depth, is_moe_layer=is_moe_layer, is_encoder_decoder=is_encoder_decoder, ) if args.checkpoint_activations: layer = checkpoint_wrapper(layer) if args.fsdp: layer = wrap(layer) return layer def forward_embedding( self, src_tokens, token_embedding=None, positions=None, ): if token_embedding is None: token_embedding = self.embed_tokens(src_tokens) x = embed = self.embed_scale * token_embedding if self.embed_positions is not None: if src_tokens is not None: x = embed + self.embed_positions(src_tokens, positions=positions) else: x = embed + self.embed_positions(x, positions=positions) if self.layernorm_embedding is not None: x = self.layernorm_embedding(x) x = self.dropout_module(x) return x, embed def forward( self, src_tokens, encoder_padding_mask=None, attn_mask=None, return_all_hiddens=False, token_embeddings=None, multiway_split_position=None, features_only=False, incremental_state=None, positions=None, **kwargs ): assert src_tokens is not None or token_embeddings is not None if encoder_padding_mask is None: if src_tokens is not None: encoder_padding_mask = torch.zeros_like( src_tokens, device=src_tokens.device ).bool() else: encoder_padding_mask = torch.zeros( [token_embeddings.size(0), token_embeddings.size(1)], device=token_embeddings.device, ).bool() if multiway_split_position is not None: assert self.args.multiway self.apply(set_split_position(multiway_split_position)) x, encoder_embedding = self.forward_embedding(src_tokens, token_embeddings, positions) x = x * (1 - encoder_padding_mask.unsqueeze(-1).type_as(x)) encoder_states = [] if return_all_hiddens: encoder_states.append(x) rel_pos_bias = None if self.relative_position is not None: rel_pos_bias = self.relative_position( batch_size=x.size(0), qlen=x.size(1), klen=x.size(1) ) # incremental_state is not None during inference if we use the bidirectional encoder as a generator as in s2s-ft (https://arxiv.org/abs/2110.13640) l_aux = [] for idx, layer in enumerate(self.layers): x, l_aux_i = layer( x, encoder_padding_mask=encoder_padding_mask if incremental_state is None else None, attn_mask=attn_mask, rel_pos=rel_pos_bias, multiway_split_position=multiway_split_position, incremental_state=incremental_state[idx] if incremental_state is not None else None, ) if return_all_hiddens: assert encoder_states is not None encoder_states.append(x) l_aux.append(l_aux_i) if self.layer_norm is not None: x = self.layer_norm(x) if not features_only and self.output_projection is not None: x = self.output_projection(x) return { "encoder_out": x, "encoder_embedding": encoder_embedding, "encoder_padding_mask": encoder_padding_mask, "encoder_states": encoder_states, "l_aux": l_aux, }
GeneSplice-main
GeneSplice/torchscale/torchscale/architecture/encoder.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import torch.nn as nn from torchscale.component.multihead_attention import MultiheadAttention from torchscale.component.multiway_network import MultiwayNetwork def init_bert_params(module): def normal_(data): data.copy_(data.cpu().normal_(mean=0.0, std=0.02).to(data.device)) if isinstance(module, nn.Linear): normal_(module.weight.data) if module.bias is not None: module.bias.data.zero_() if isinstance(module, nn.Embedding): normal_(module.weight.data) if module.padding_idx is not None: module.weight.data[module.padding_idx].zero_() if isinstance(module, MultiheadAttention): if isinstance(module.q_proj, MultiwayNetwork): normal_(module.q_proj.A.weight.data) normal_(module.q_proj.B.weight.data) normal_(module.k_proj.A.weight.data) normal_(module.k_proj.B.weight.data) normal_(module.v_proj.A.weight.data) normal_(module.v_proj.B.weight.data) else: normal_(module.q_proj.weight.data) normal_(module.k_proj.weight.data) normal_(module.v_proj.weight.data)
GeneSplice-main
GeneSplice/torchscale/torchscale/architecture/utils.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import torch import torch.nn as nn from torchscale.architecture.encoder import Encoder from torchscale.component.embedding import ( PositionalEmbedding, TextEmbedding, VisionEmbedding, ) from torchscale.component.multiway_network import MutliwayEmbedding class BEiT3(nn.Module): def __init__(self, args, **kwargs): super().__init__() self.args = args assert args.multiway assert args.vocab_size > 0 assert not args.share_encoder_input_output_embed self.text_embed = TextEmbedding(args.vocab_size, args.encoder_embed_dim) self.vision_embed = VisionEmbedding( args.img_size, args.patch_size, args.in_chans, args.encoder_embed_dim, contain_mask_token=True, prepend_cls_token=True, ) # being consistent with Fairseq, which starts from 2 for position embedding embed_positions = MutliwayEmbedding( modules=[ PositionalEmbedding(self.vision_embed.num_position_embeddings() + 2, args.encoder_embed_dim), PositionalEmbedding(args.max_source_positions, args.encoder_embed_dim), ], dim=1, ) self.encoder = Encoder( args, embed_tokens=None, embed_positions=embed_positions, output_projection=None, is_encoder_decoder=False, ) def forward( self, textual_tokens=None, visual_tokens=None, text_padding_position=None, attn_mask=None, vision_masked_position=None, incremental_state=None, positions=None, ): assert textual_tokens is not None or visual_tokens is not None if textual_tokens is None: x = self.vision_embed(visual_tokens, vision_masked_position) encoder_padding_mask = None multiway_split_position = -1 elif visual_tokens is None: x = self.text_embed(textual_tokens) encoder_padding_mask = text_padding_position multiway_split_position = 0 else: x1 = self.vision_embed(visual_tokens, vision_masked_position) multiway_split_position = x1.size(1) x2 = self.text_embed(textual_tokens) x = torch.cat([x1, x2], dim=1) if text_padding_position is not None: encoder_padding_mask = torch.cat( [ torch.zeros(x1.shape[:-1]).to(x1.device).bool(), text_padding_position, ], dim=1, ) else: encoder_padding_mask = None encoder_out = self.encoder( src_tokens=None, encoder_padding_mask=encoder_padding_mask, attn_mask=attn_mask, token_embeddings=x, multiway_split_position=multiway_split_position, incremental_state=incremental_state, positions=positions, ) encoder_out["multiway_split_position"] = multiway_split_position return encoder_out
GeneSplice-main
GeneSplice/torchscale/torchscale/model/BEiT3.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/torchscale/model/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import pytest import torch from torchscale.architecture.config import DecoderConfig from torchscale.architecture.decoder import Decoder testcases = [ {}, {"vocab_size": 64000}, {"activation_fn": "relu"}, {"drop_path_rate": 0.1}, {"decoder_normalize_before": False}, {"no_scale_embedding": False}, {"layernorm_embedding": True}, {"rel_pos_buckets": 32, "max_rel_pos": 256}, {"deepnorm": True, "subln": False, "decoder_normalize_before": False}, {"bert_init": True}, {"multiway": True}, {"share_decoder_input_output_embed": True}, {"checkpoint_activations": True}, {"fsdp": True}, ] @pytest.mark.parametrize("args", testcases) def test_decoder(args): config = DecoderConfig(**args) model = Decoder(config) prev_output_tokens = torch.ones(2, 10) token_embeddings = torch.rand(2, 10, config.decoder_embed_dim) model( prev_output_tokens=prev_output_tokens, token_embeddings=token_embeddings, features_only=True, )
GeneSplice-main
GeneSplice/torchscale/tests/test_decoder.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import pytest import torch from torchscale.architecture.config import EncoderConfig from torchscale.architecture.encoder import Encoder testcases = [ {}, {"vocab_size": 64000}, {"activation_fn": "relu"}, {"drop_path_rate": 0.1}, {"encoder_normalize_before": False}, {"no_scale_embedding": False}, {"layernorm_embedding": True}, {"rel_pos_buckets": 32, "max_rel_pos": 256}, {"deepnorm": True, "subln": False, "encoder_normalize_before": False}, {"bert_init": True}, {"multiway": True}, {"share_encoder_input_output_embed": True}, {"checkpoint_activations": True}, {"fsdp": True}, ] @pytest.mark.parametrize("args", testcases) def test_encoder(args): config = EncoderConfig(**args) model = Encoder(config) token_embeddings = torch.rand(2, 10, config.encoder_embed_dim) model(src_tokens=None, token_embeddings=token_embeddings)
GeneSplice-main
GeneSplice/torchscale/tests/test_encoder.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/tests/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import pytest import torch from torchscale.architecture.config import EncoderDecoderConfig from torchscale.architecture.encoder_decoder import EncoderDecoder from torchscale.component.embedding import PositionalEmbedding, TextEmbedding testcases = [ {}, {"vocab_size": 64000}, {"activation_fn": "relu"}, {"drop_path_rate": 0.1}, {"encoder_normalize_before": False, "decoder_normalize_before": False}, {"no_scale_embedding": False}, {"layernorm_embedding": True}, {"rel_pos_buckets": 32, "max_rel_pos": 256}, { "deepnorm": True, "subln": False, "encoder_normalize_before": False, "decoder_normalize_before": False, }, {"bert_init": True}, {"multiway": True}, {"share_decoder_input_output_embed": True}, {"share_all_embeddings": True}, {"checkpoint_activations": True}, {"fsdp": True}, ] @pytest.mark.parametrize("args", testcases) def test_decoder(args): config = EncoderDecoderConfig(**args) model = EncoderDecoder( config, encoder_embed_tokens=TextEmbedding(64000, config.encoder_embed_dim), decoder_embed_tokens=TextEmbedding(64000, config.decoder_embed_dim), encoder_embed_positions=PositionalEmbedding( config.max_source_positions, config.encoder_embed_dim ), decoder_embed_positions=PositionalEmbedding( config.max_target_positions, config.decoder_embed_dim ), ) src_tokens = torch.ones(2, 20).long() prev_output_tokens = torch.ones(2, 10).long() model( src_tokens=src_tokens, prev_output_tokens=prev_output_tokens, features_only=True, )
GeneSplice-main
GeneSplice/torchscale/tests/test_encoder_decoder.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/examples/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # flake8: noqa import models import tasks import criterions from fairseq_cli.generate import cli_main if __name__ == "__main__": cli_main()
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/generate.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # flake8: noqa import models import tasks import criterions from fairseq_cli.interactive import cli_main if __name__ == "__main__": cli_main()
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/interactive.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # flake8: noqa import models import tasks import criterions from fairseq_cli.train import cli_main if __name__ == "__main__": cli_main()
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/train.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import json import logging import os from argparse import Namespace # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from dataclasses import dataclass, field import sentencepiece as spm from fairseq import utils from fairseq.data import Dictionary from fairseq.dataclass import ChoiceEnum, FairseqDataclass from fairseq.tasks import FairseqTask, register_task from omegaconf import II, MISSING from .data.mlm_loader import MLMLoader logger = logging.getLogger(__name__) SAMPLE_BREAK_MODE_CHOICES = ChoiceEnum(["none", "complete", "complete_doc", "eos"]) SHORTEN_METHOD_CHOICES = ChoiceEnum(["none", "truncate", "random_crop"]) @dataclass class PretrainingConfig(FairseqDataclass): data: str = field( default=MISSING, metadata={ "help": "colon separated path to data directories list, \ will be iterated upon during epochs in round-robin manner" }, ) sample_break_mode: SAMPLE_BREAK_MODE_CHOICES = field( default="complete", metadata={ "help": 'If omitted or "none", fills each sample with tokens-per-sample ' 'tokens. If set to "complete", splits samples only at the end ' "of sentence, but may include multiple sentences per sample. " '"complete_doc" is similar but respects doc boundaries. ' 'If set to "eos", includes only one sentence per sample.' }, ) tokens_per_sample: int = field( default=1024, metadata={"help": "max number of tokens per sample for LM dataset"}, ) mask_prob: float = field( default=0.15, metadata={"help": "probability of replacing a token with mask"}, ) leave_unmasked_prob: float = field( default=0.1, metadata={"help": "probability that a masked token is unmasked"}, ) random_token_prob: float = field( default=0.1, metadata={"help": "probability of replacing a token with a random token"}, ) freq_weighted_replacement: bool = field( default=False, metadata={"help": "sample random replacement words based on word frequencies"}, ) mask_whole_words: bool = field( default=False, metadata={"help": "mask whole words; you may also want to set --bpe"}, ) mask_multiple_length: int = field( default=1, metadata={"help": "repeat the mask indices multiple times"}, ) mask_stdev: float = field( default=0.0, metadata={"help": "stdev of the mask length"}, ) shorten_method: SHORTEN_METHOD_CHOICES = field( default="none", metadata={ "help": "if not none, shorten sequences that exceed --tokens-per-sample" }, ) shorten_data_split_list: str = field( default="", metadata={ "help": "comma-separated list of dataset splits to apply shortening to, " 'e.g., "train,valid" (default: all dataset splits)' }, ) seed: int = II("common.seed") span_length: float = field( default=3.0, metadata={"help": "average span length for masking"}, ) remove_source_sentinel: bool = field( default=False, metadata={"help": "remove the source sentinel for the span corruption task"}, ) remove_target_sentinel: bool = field( default=False, metadata={"help": "remove the target sentinel for the span corruption task"}, ) batch_read_ahead: int = field( default=100000, metadata={"help": "batch read ahead size for infinibatch"}, ) required_batch_size_multiple: int = II("dataset.required_batch_size_multiple") spm_model: str = field( default="", metadata={"help": "sentencepice model to tokenize the data"}, ) dict_file: str = field( default="", metadata={"help": ""}, ) pad_to_max_length: bool = field( default=False, ) @register_task("pretraining", dataclass=PretrainingConfig) class PLMTask(FairseqTask): def __init__(self, cfg, dictionary, tokenizer): super().__init__(cfg) self.cfg = cfg self.dictionary = dictionary self.tokenizer = tokenizer self.seed = cfg.seed self.mask_idx = dictionary.index("<mask>") @classmethod def setup_task(cls, cfg, **kwargs): paths = utils.split_paths(cfg.data) assert len(paths) > 0 if cfg.dict_file != "": dictionary = Dictionary.load(cfg.dict_file) else: dictionary = Dictionary.load(os.path.join(paths[0], "dict.txt")) # add mask token dictionary.add_symbol("<mask>") for i in range(100): dictionary.add_symbol(f"<mask_{i}>") dictionary.pad_to_multiple_(cfg.required_batch_size_multiple) logger.info("dictionary: {} types".format(len(dictionary))) # tokenizer = SentencepieceBPE(Namespace(sentencepiece_model=cfg.spm_model)) tokenizer = spm.SentencePieceProcessor() tokenizer.Load(cfg.spm_model) return cls(cfg, dictionary, tokenizer) def load_dataset(self, split, epoch=1, combine=False, **kwargs): self.datasets[split] = { "data": json.load(open(f"{self.cfg.data}/json/{split}.json")), "data_dir": self.cfg.data, "shuffle": True if split == "train" else False, } self.datasets[split] = Namespace(**self.datasets[split]) def dataset(self, split): if split not in self.datasets: raise KeyError("Dataset not loaded: " + split) return self.datasets[split] def get_batch_iterator( self, dataset, max_tokens=None, max_sentences=None, max_positions=None, ignore_invalid_inputs=False, required_batch_size_multiple=1, seed=1, num_shards=1, shard_id=0, num_workers=0, epoch=1, data_buffer_size=0, disable_iterator_cache=False, **kwargs, ): return MLMLoader( self.cfg, dataset, self.dictionary, self.tokenizer, max_tokens=max_tokens, max_sentences=max_sentences, max_positions=max_positions, ignore_invalid_inputs=ignore_invalid_inputs, required_batch_size_multiple=required_batch_size_multiple, seed=seed, num_shards=num_shards, shard_id=shard_id, ) @property def source_dictionary(self): return self.dictionary @property def target_dictionary(self): return self.dictionary
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/tasks/pretraining.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import argparse import importlib import os # register dataclass TASK_DATACLASS_REGISTRY = {} TASK_REGISTRY = {} TASK_CLASS_NAMES = set() # automatically import any Python files in the tasks/ directory tasks_dir = os.path.dirname(__file__) for file in os.listdir(tasks_dir): path = os.path.join(tasks_dir, file) if ( not file.startswith("_") and not file.startswith(".") and (file.endswith(".py") or os.path.isdir(path)) ): task_name = file[: file.find(".py")] if file.endswith(".py") else file module = importlib.import_module("tasks." + task_name) # expose `task_parser` for sphinx if task_name in TASK_REGISTRY: parser = argparse.ArgumentParser(add_help=False) group_task = parser.add_argument_group("Task name") # fmt: off group_task.add_argument('--task', metavar=task_name, help='Enable this task with: ``--task=' + task_name + '``') # fmt: on group_args = parser.add_argument_group("Additional command-line arguments") TASK_REGISTRY[task_name].add_args(group_args) globals()[task_name + "_parser"] = parser
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/tasks/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import torch from infinibatch.iterators import CheckpointableIterator from . import utils class BaseBatchGen(CheckpointableIterator): """ This is a base class for batch generators that use infinibatch """ def __init__(self): self._iter = None self.epoch = 1 self.next_epoch_idx = 1 self.sharded_checkpoint = True self.should_close_after_finished = True def _build_iter(self): """ Build infinibatch iterator and assign to self._iter """ raise NotImplementedError() def _move_to_tensor(self, batch): def to_tensor(x): return torch.tensor(x) return utils.apply_to_sample(to_tensor, batch) @property def iterator(self): if self._iter is None: raise NotImplementedError("_build_iter() must called first") return self._iter def __iter__(self): if self._iter is None: raise NotImplementedError("_build_iter() must called first") return self._iter def __next__(self): return next(self._iter) def setstate(self, value): self._iter.setstate(value) def getstate(self): return self._iter.getstate() def close(self): self._iter.close() def __len__(self) -> int: return 819200000 def next_epoch_itr( self, shuffle=True, fix_batches_to_gpus=False, set_dataset_epoch=True ): return self def end_of_epoch(self) -> bool: return False def state_dict(self): """Returns a dictionary containing a whole state of the iterator.""" return self.getstate() def load_state_dict(self, state_dict): """Copies the state of the iterator from the given *state_dict*.""" self.setstate(state_dict) @property def first_batch(self): return "DUMMY"
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/tasks/data/basic_loader.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/tasks/data/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import collections from random import Random from typing import Dict, Iterable, Optional import numpy as np from infinibatch import iterators def apply_to_sample(f, sample): if hasattr(sample, "__len__") and len(sample) == 0: return {} def _apply(x): if isinstance(x, np.ndarray): return f(x) elif isinstance(x, collections.OrderedDict): # OrderedDict has attributes that needs to be preserved od = collections.OrderedDict( (key, _apply(value)) for key, value in x.items() ) od.__dict__ = x.__dict__ return od elif isinstance(x, dict): return {key: _apply(value) for key, value in x.items()} elif isinstance(x, list): return [_apply(x) for x in x] elif isinstance(x, tuple): return tuple(_apply(x) for x in x) elif isinstance(x, set): return {_apply(x) for x in x} else: return x return _apply(sample) class NativeCheckpointableIterator(iterators.CheckpointableIterator): def __init__(self, iterable: Iterable): self._input_iterable = iterable self.setstate(None) def getstate(self) -> Dict: return {"num_items_yielded": self._num_items_yielded} def setstate(self, checkpoint: Optional[Dict]): self._iterator = iter(self._input_iterable) self._num_items_yielded = ( iterators._advance_iterator(self._iterator, checkpoint["num_items_yielded"]) if checkpoint is not None else 0 ) def __next__(self): item = next(self._iterator) self._num_items_yielded += 1 return item def close(self): pass class WeightIterator(object): def __init__(self, weights, seed): self.weights = weights self.seed = seed self.control_index = list(range(len(weights))) self.setstate(None) def __iter__(self): return self def getstate(self): return {"random_state": self._random_state} def setstate(self, checkpoint): self._random_state = checkpoint["random_state"] if checkpoint else None self._random = ( None # this will trigger the lazy initialization in self.__next__ ) def __next__(self): if self._random is None: self._random = Random(self.seed) if self._random_state is not None: self._random.setstate(self._random_state) idx = self._random.choices(self.control_index, self.weights)[0] self._random_state = self._random.getstate() return idx def close(self): pass
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/tasks/data/utils.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import copy import itertools import os import numpy as np from infinibatch import iterators from .basic_loader import BaseBatchGen from .utils import NativeCheckpointableIterator, WeightIterator class MLMLoader(BaseBatchGen): def __init__( self, args, dataset, dictionary, tokenizer, max_tokens=None, max_sentences=None, max_positions=None, ignore_invalid_inputs=False, required_batch_size_multiple=1, seed=1, num_shards=1, shard_id=0, ): super().__init__() self.args = args self.data = dataset.data self.data_dir = dataset.data_dir self.shuffle = dataset.shuffle self.dictionary = dictionary self.tokenizer = tokenizer self.max_tokens = max_tokens self.max_sentences = max_sentences self.max_positions = max_positions self.tokens_per_sample = args.tokens_per_sample self.sample_break_mode = args.sample_break_mode self.ignore_invalid_inputs = ignore_invalid_inputs self.required_batch_size_multiple = required_batch_size_multiple self.seed = str(seed) self.num_shards = num_shards self.shard_id = shard_id self.batch_read_ahead = args.batch_read_ahead self._build_iter() def _build_iter(self): tokenized_lines = self._multilingual_tokenize() self.padded_batches = self._batchify(tokenized_lines) prefetch_batches = iterators.PrefetchIterator( self.padded_batches, buffer_size=10000, buffer_in_main_process=True, log_empty_buffer_warning=True and self.shard_id == 0, ) prefetch_batches = iterators.MapIterator(prefetch_batches, self._move_to_tensor) self._iter = prefetch_batches def _multilingual_tokenize(self): multilingual_iters = [] weights = [] for data in self.data: multilingual_iters.append(self._tokenize(data)) if "weight" in data: weights.append(float(data["weight"])) else: weights.append(int(data["count"])) if len(multilingual_iters) == 1: return multilingual_iters[0] sampling_iterator = WeightIterator(weights) control_iterator = NativeCheckpointableIterator(sampling_iterator) tokenized_lines = iterators.MultiplexIterator( control_iterator, multilingual_iters ) return tokenized_lines def _tokenize(self, data): """ data: { 'source': list[Path], 'source_lang': str, 'count': int, 'weight': float, 'name': str, } """ dataset = list( zip( data["source"], itertools.repeat(data["source_lang"]), ) ) if self.shuffle: chunk_files = iterators.InfinitePermutationSourceIterator( dataset, seed=self.seed, shuffle=self.shuffle, num_instances=self.num_shards, instance_rank=self.shard_id, ) else: chunk_files = iterators.ChunkedSourceIterator( dataset, num_instances=self.num_shards, instance_rank=self.shard_id, ) tokenized_lines = iterators.SelectManyIterator( chunk_files, lambda files: self._read_from_files(*files) ) tokenized_lines = iterators.SamplingRandomMapIterator( tokenized_lines, self._prepare, self.seed ) return tokenized_lines def _batchify(self, lines): if self.max_sentences is not None: if self.batch_read_ahead > 0: lines = iterators.BlockwiseShuffleIterator( lines, self.batch_read_ahead, self.seed ) batches = iterators.FixedBatchIterator(lines, self.max_sentences) else: def dynamic_batch_size(sample): lengths = [len(x) for x in sample] batch_size = self.max_tokens // max(lengths) batch_size = ( batch_size // self.required_batch_size_multiple * self.required_batch_size_multiple ) return max(1, batch_size) batches = iterators.BucketedReadaheadBatchIterator( lines, read_ahead=self.batch_read_ahead, key=(lambda x: max(len(x[0]), len(x[1]))) if self.shuffle else None, batch_size=dynamic_batch_size, shuffle=self.shuffle, seed=self.seed, ) def collate(batch): batch_size = len(batch) mlm_source_max_length = max([len(x[0]) for x in batch]) mlm_target_max_length = max([len(x[1]) for x in batch]) s2s_source_max_length = max([len(x[2]) for x in batch]) s2s_target_max_length = max([len(x[3]) for x in batch]) if self.args.pad_to_max_length: mlm_source_max_length = self.args.tokens_per_sample mlm_target_max_length = self.args.tokens_per_sample mlm_source_ids = np.full( shape=(batch_size, mlm_source_max_length), dtype=np.int32, fill_value=self.dictionary.pad(), ) mlm_target_ids = np.full( shape=(batch_size, mlm_target_max_length), dtype=np.int32, fill_value=self.dictionary.pad(), ) s2s_source_ids = np.full( shape=(batch_size, s2s_source_max_length), dtype=np.int32, fill_value=self.dictionary.pad(), ) s2s_target_ids = np.full( shape=(batch_size, s2s_target_max_length - 1), dtype=np.int32, fill_value=self.dictionary.pad(), ) s2s_prev_input_ids = np.full( shape=(batch_size, s2s_target_max_length - 1), dtype=np.int32, fill_value=self.dictionary.pad(), ) for i, ( mlm_input_ids, mlm_label_ids, s2s_input_ids, s2s_label_ids, ) in enumerate(batch): mlm_source_ids[i, : len(mlm_input_ids)] = mlm_input_ids mlm_target_ids[i, : len(mlm_label_ids)] = mlm_label_ids s2s_source_ids[i, : len(s2s_input_ids)] = s2s_input_ids s2s_target_ids[i, : len(s2s_label_ids) - 1] = s2s_label_ids[1:] s2s_prev_input_ids[i, : len(s2s_label_ids) - 1] = s2s_label_ids[:-1] ret_batch = { "net_input": { "src_tokens": mlm_source_ids.astype(np.int64), }, "target": mlm_target_ids.astype(np.int64), "nsentences": batch_size, "ntokens": sum([len(x[0]) for x in batch]), } return ret_batch padded_batches = iterators.MapIterator(batches, collate) return padded_batches def _prepare(self, _random, doc): nonmasked_tokens, masked_tokens = self._mask_lm(_random, doc) nonnoise_spans, noise_spans = self._span_corruption(_random, doc) return nonmasked_tokens, masked_tokens, nonnoise_spans, noise_spans def _mask_lm(self, _random, doc): def mask_tokens(): return "<mask>" length = len(doc) mask_tokens_num = int(length * self.args.mask_prob) mask_tokens_num = min(max(mask_tokens_num, 1), length - 1) possible_mask_positions = _random.sample(range(length), k=mask_tokens_num) possible_mask_positions = sorted(possible_mask_positions) nonmasked_tokens = copy.deepcopy(doc) masked_tokens = [self.dictionary.pad() for _ in range(len(doc))] for position in possible_mask_positions: # masked_tokens.append(nonmasked_tokens[position]) masked_tokens[position] = nonmasked_tokens[position] nonmasked_tokens[position] = self.dictionary.indices[mask_tokens()] return nonmasked_tokens, masked_tokens def _span_corruption(self, _random, doc): def mask_tokens(i): return f"<mask_{i}>" length = len(doc) noise_tokens_num = int(length * self.args.mask_prob) noise_tokens_num = min(max(noise_tokens_num, 1), length - 1) noise_spans_num = int(noise_tokens_num / self.args.span_length) noise_spans_num = max(noise_spans_num, 1) nonnoise_tokens_num = length - noise_tokens_num if noise_spans_num == 1: noise_split_positions = [0, noise_tokens_num] else: possible_split_positions = list(range(1, noise_tokens_num)) _random.shuffle(possible_split_positions) noise_split_positions = sorted( possible_split_positions[: noise_spans_num - 1] ) noise_split_positions = [0] + noise_split_positions + [noise_tokens_num] possible_insert_positions = list(range(nonnoise_tokens_num)) _random.shuffle(possible_insert_positions) noise_insert_positions = sorted(possible_insert_positions[:noise_spans_num]) nonnoise_spans, noise_spans = [], [] last_end = 0 for i in range(noise_spans_num): start_pos = noise_insert_positions[i] + noise_split_positions[i] end_pos = noise_insert_positions[i] + noise_split_positions[i + 1] mask_id = self.dictionary.indices[mask_tokens(i)] if getattr(self.args, "remove_target_sentinel", False): noise_spans.append(doc[start_pos:end_pos]) else: noise_spans.append([mask_id] + doc[start_pos:end_pos]) if getattr(self.args, "remove_source_sentinel", False): nonnoise_spans.extend(doc[last_end:start_pos]) else: nonnoise_spans.extend(doc[last_end:start_pos] + [mask_id]) last_end = end_pos nonnoise_spans.extend(doc[last_end:]) noise_spans = sum(noise_spans, []) return nonnoise_spans, noise_spans def _read_from_files(self, source_file, source_lang): # data = [] file_path = os.path.join(self.data_dir, source_file) if not os.path.exists(file_path): print("| file {} not exists".format(file_path), flush=True) return iter([]) # skip bad file with open(file_path, "r", encoding="utf8") as f: lines = f.read().strip().split("\n") doc = [self.dictionary.bos()] for line in lines: if line == "": if self.sample_break_mode == "complete_doc": # data.append(doc) yield doc doc = [self.dictionary.bos()] continue tokenized_line = self.tokenizer.EncodeAsPieces(line) tokenized_id = [ self.dictionary.index(token) for token in tokenized_line ] + [self.dictionary.eos_index] if len(tokenized_id) > self.tokens_per_sample: continue if len(doc) + len(tokenized_id) > self.tokens_per_sample: # data.append(doc) yield doc doc = [self.dictionary.bos()] doc.extend(tokenized_id) if len(doc) > 1 and len(doc) <= self.tokens_per_sample: # data.append(doc) yield doc # return data
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/tasks/data/mlm_loader.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import math import warnings import torch import torch.distributed as dist from fairseq.utils import multi_tensor_l2norm_available, multi_tensor_total_norm @torch.no_grad() def clip_grad_norm_( params, max_norm, moe_expert_count, aggregate_norm_fn=None ) -> torch.Tensor: def grad_exists(p): return p is not None and getattr(p, "grad", None) is not None if isinstance(params, torch.Tensor): params = [params] params = list(params) params = list(filter(grad_exists, params)) grads, expert_grads, base_expert_grads, sharded_grads = [], [], [], [] denom = math.sqrt(max(dist.get_global_world_size(), moe_expert_count)) for p in params: if hasattr(p, "expert"): expert_grads.append(p.grad.detach() / denom) elif hasattr(p, "base_expert"): base_expert_grads.append(p.grad.detach()) elif hasattr(p, "_is_sharded"): sharded_grads.append(p.grad.detach()) else: grads.append(p.grad.detach()) if len(grads) == 0: if len(params) > 0: total_norm = params[0].new_tensor(0.0) else: total_norm = torch.tensor(0.0) elif len(grads) == 1: total_norm = torch.norm(grads[0], p=2, dtype=torch.float32) else: if multi_tensor_l2norm_available: total_norm = multi_tensor_total_norm(grads) else: if torch.cuda.is_available(): warnings.warn( "amp_C fused kernels unavailable, disabling multi_tensor_l2norm; " "you may get better performance by installing NVIDIA's apex library" ) device = torch.cuda.current_device() elif grads[0].device.type == "xla": device = grads[0].device else: device = torch.device("cpu") total_norm = torch.norm( torch.stack( [torch.norm(g, p=2, dtype=torch.float32).to(device) for g in grads] ) ) # calculate split_norm and all_reduce with other workers norms = [total_norm] for split_grads in [expert_grads, sharded_grads]: if len(split_grads) == 0: continue split_norm = torch.norm( torch.stack([torch.norm(g, p=2, dtype=torch.float32) for g in split_grads]) ) if dist.is_initialized(): split_norm.pow_(2) dist.all_reduce(split_norm) split_norm.sqrt_() norms.append(split_norm) if len(norms) > 1: total_norm = torch.norm(torch.stack(norms)) if aggregate_norm_fn is not None: total_norm = aggregate_norm_fn(total_norm) if max_norm > 0: max_norm = float(max_norm) clip_coef = (max_norm / (total_norm + 1e-6)).clamp_(max=1) for g in grads + expert_grads + sharded_grads + base_expert_grads: g.mul_(clip_coef) return total_norm
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/utils/sparse_clip.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details]
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/utils/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import logging from dataclasses import dataclass, field from typing import Optional import torch from fairseq import distributed_utils, utils from fairseq.dataclass import ChoiceEnum, FairseqDataclass from fairseq.models import ( FairseqIncrementalDecoder, FairseqLanguageModel, register_model, register_model_architecture, ) from fairseq.models.transformer import DEFAULT_MIN_PARAMS_TO_WRAP, Embedding from fairseq.modules import PositionalEmbedding from omegaconf import II from torchscale.architecture.config import DecoderConfig from torchscale.architecture.decoder import Decoder DEFAULT_MAX_TARGET_POSITIONS = 1024 logger = logging.getLogger(__name__) @dataclass class LanguageConfig(FairseqDataclass): activation_fn: ChoiceEnum(utils.get_available_activation_fns()) = field( default="relu", metadata={"help": "activation function to use"} ) dropout: float = field(default=0.1, metadata={"help": "dropout probability"}) attention_dropout: float = field( default=0.0, metadata={"help": "dropout probability for attention weights"} ) activation_dropout: float = field( default=0.0, metadata={"help": "dropout probability after activation in FFN."} ) relu_dropout: float = field( default=0.0, metadata={"help": "dropout probability after activation in FFN."} ) decoder_embed_dim: int = field( default=512, metadata={"help": "decoder embedding dimension"} ) decoder_output_dim: int = field( default=512, metadata={"help": "decoder output dimension"} ) decoder_input_dim: int = field( default=512, metadata={"help": "decoder input dimension"} ) decoder_ffn_embed_dim: int = field( default=2048, metadata={"help": "decoder embedding dimension for FFN"} ) decoder_layers: int = field(default=6, metadata={"help": "num decoder layers"}) decoder_attention_heads: int = field( default=8, metadata={"help": "num decoder attention heads"} ) decoder_normalize_before: bool = field( default=False, metadata={"help": "apply layernorm before each decoder block"} ) no_token_positional_embeddings: bool = field( default=False, metadata={ "help": "if set, disables positional embeddings (outside self attention)" }, ) share_decoder_input_output_embed: bool = field( default=False, metadata={"help": "share decoder input and output embeddings"} ) decoder_learned_pos: bool = field( default=False, metadata={"help": "use learned positional embeddings in the decoder"}, ) layernorm_embedding: bool = field( default=False, metadata={"help": "add layernorm to embedding"} ) no_scale_embedding: bool = field( default=False, metadata={"help": "if True, dont scale embeddings"} ) checkpoint_activations: bool = field( default=False, metadata={"help": "checkpoint activations at each layer"} ) offload_activations: bool = field( default=False, metadata={"help": "move checkpointed activations to CPU after they are used."}, ) # config for Fully Sharded Data Parallel (FSDP) training min_params_to_wrap: int = field( default=DEFAULT_MIN_PARAMS_TO_WRAP, metadata={ "help": ( "minimum number of params for a layer to be wrapped with FSDP() when " "training with --ddp-backend=fully_sharded. Smaller values will " "improve memory efficiency, but may make torch.distributed " "communication less efficient due to smaller input sizes. This option " "is set to 0 (i.e., always wrap) when --checkpoint-activations or " "--offload-activations are passed." ) }, ) moe_freq: int = field( default=0, metadata={"help": "Frequency at which we insert MoE Transformer layers"}, ) moe_expert_count: int = field( default=0, metadata={"help": "Number of experts in each MoE Layer"} ) moe_gating_use_fp32: bool = field( default=False, metadata={"help": "Use FP32 computations in MoE top2 gating function"}, ) moe_second_expert_policy: str = field( default="sampling", metadata={"help": "policy for second expert, options: all/sampling/random"}, ) moe_normalize_gate_prob_before_dropping: bool = field( default=False, metadata={ "help": "whether to normalize gate probs before or after dropping experts for capacity and randomization" }, ) moe_expert_ffn_dim: Optional[int] = field( default=None, metadata={"help": "MoE expert FFN dimension"} ) moe_top1_expert: Optional[bool] = field( default=False, metadata={"help": "Use top1 gate instead of top2"} ) moe_eval_capacity_token_fraction: Optional[float] = field( default=0.25, metadata={ "help": ( "Default: 0.25, Fraction of tokens as capacity during validation, " "if set to negative, use same as training. range: (0.0, 1.0]." ) }, ) moe_normalize_expert_grad: Optional[str] = field( default="world_size", metadata={ "help": "Divide expert gradients by (1) 'world_size' (2) 'sqrt_world_size'" }, ) record_a2a_perf_stats: Optional[bool] = field( default=False, metadata={"help": "records all to all perf stats during distributed training"}, ) dummy_a2a: Optional[bool] = field( default=False, metadata={ "help": "By passes all to all during distributed training by returning the input buffer as output" }, ) moe_batch_prioritized_routing: Optional[bool] = field( default=False, metadata={ "help": "if true orders token by the gate prob before capacity dropping." }, ) use_xmoe: Optional[bool] = field( default=False, ) # options from other parts of the config add_bos_token: bool = II("task.add_bos_token") tokens_per_sample: int = II("task.tokens_per_sample") max_target_positions: Optional[int] = II("task.max_target_positions") tpu: bool = II("common.tpu") memory_efficient_fp16: bool = II("common.memory_efficient_fp16") fp16: bool = II("common.fp16") fp16_no_flatten_grads: bool = II("common.fp16_no_flatten_grads") ddp_backend: str = II("distributed_training.ddp_backend") world_size: int = II("distributed_training.distributed_world_size") distributed_rank: int = II("distributed_training.distributed_rank") ddp_rank: int = II("distributed_training.distributed_rank") deepnorm: Optional[bool] = field( default=False, ) subln: Optional[bool] = field( default=False, ) rel_pos_buckets: Optional[int] = field( default=0, ) max_rel_pos: Optional[int] = field( default=0, ) xpos_rel_pos: Optional[bool] = field( default=False, ) xpos_scale_base: Optional[int] = field( default=512, ) @register_model("lm", dataclass=LanguageConfig) class LanguageModel(FairseqLanguageModel): def __init__(self, args, decoder): self.args = args super().__init__(decoder) @classmethod def build_model(cls, args, task): if getattr(args, "max_target_positions", None) is None: args.max_target_positions = getattr( args, "tokens_per_sample", DEFAULT_MAX_TARGET_POSITIONS ) embed_tokens = cls.build_embedding( args, task.source_dictionary, args.decoder_embed_dim ) embed_positions = ( PositionalEmbedding( args.max_target_positions, args.decoder_embed_dim, task.dictionary.pad(), learned=args.decoder_learned_pos, ) if not args.no_token_positional_embeddings else None ) if args.share_decoder_input_output_embed: output_projection = torch.nn.Linear( embed_tokens.weight.shape[1], embed_tokens.weight.shape[0], bias=False, ) output_projection.weight = embed_tokens.weight else: output_projection = torch.nn.Linear( args.decoder_embed_dim, len(task.dictionary), bias=False ) torch.nn.init.normal_( output_projection.weight, mean=0, std=args.decoder_embed_dim**-0.5 ) if getattr(args, "moe_freq", 0) > 0 and ( getattr(args, "fp16", False) and not getattr(args, "memory_efficient_fp16", False) and getattr(args, "ddp_backend", None) != "fully_sharded" ): assert ( args.fp16_no_flatten_grads ), "If training moe models, set --fp16-no-flatten-grads to calculate correct gradnorm" args.ddp_rank = distributed_utils.get_data_parallel_rank() config = DecoderConfig() config.override(args) decoder = LMDecoder( config, embed_tokens, embed_positions, output_projection, is_encoder_decoder=False, dictionary=task.dictionary, ) return cls(args, decoder) @classmethod def build_embedding(cls, args, dictionary, embed_dim, path=None): return Embedding(len(dictionary), embed_dim, dictionary.pad()) class LMDecoder(Decoder, FairseqIncrementalDecoder): def forward(self, src_tokens, **kwargs): self_attn_padding_mask = src_tokens.eq(self.dictionary.pad()) return super().forward(src_tokens, self_attn_padding_mask, **kwargs) def max_positions(self): return self.embed_positions.max_positions def reorder_incremental_state_scripting( self, incremental_state, new_order, ): for module in incremental_state: for key in incremental_state[module]: result = incremental_state[module][key].index_select(0, new_order) incremental_state[module][key] = result @register_model_architecture("lm", "lm_base") def base_lm_architecture(args): # backward compatibility for older model checkpoints if hasattr(args, "no_tie_adaptive_proj"): # previous models defined --no-tie-adaptive-proj, so use the existence of # that option to determine if this is an "old" model checkpoint args.no_decoder_final_norm = True # old models always set this to True if args.no_tie_adaptive_proj is False: args.tie_adaptive_proj = True if hasattr(args, "decoder_final_norm"): args.no_decoder_final_norm = not args.decoder_final_norm args.dropout = getattr(args, "dropout", 0.1) args.attention_dropout = getattr(args, "attention_dropout", 0.0) args.decoder_embed_dim = getattr(args, "decoder_embed_dim", 512) args.decoder_ffn_embed_dim = getattr(args, "decoder_ffn_embed_dim", 2048) args.decoder_layers = getattr(args, "decoder_layers", 6) args.decoder_attention_heads = getattr(args, "decoder_attention_heads", 8) args.adaptive_softmax_cutoff = getattr(args, "adaptive_softmax_cutoff", None) args.adaptive_softmax_dropout = getattr(args, "adaptive_softmax_dropout", 0) args.adaptive_softmax_factor = getattr(args, "adaptive_softmax_factor", 4) args.decoder_learned_pos = getattr(args, "decoder_learned_pos", False) args.activation_fn = getattr(args, "activation_fn", "relu") args.decoder_layerdrop = getattr(args, "decoder_layerdrop", 0) args.decoder_layers_to_keep = getattr(args, "decoder_layers_to_keep", None) args.base_layers = getattr(args, "base_layers", 0) args.base_sublayers = getattr(args, "base_sublayers", 1) args.base_shuffle = getattr(args, "base_shuffle", False) args.add_bos_token = getattr(args, "add_bos_token", False) args.no_token_positional_embeddings = getattr( args, "no_token_positional_embeddings", False ) args.share_decoder_input_output_embed = getattr( args, "share_decoder_input_output_embed", False ) args.character_embeddings = getattr(args, "character_embeddings", False) args.decoder_output_dim = getattr( args, "decoder_output_dim", args.decoder_embed_dim ) args.decoder_input_dim = getattr(args, "decoder_input_dim", args.decoder_embed_dim) # Model training is not stable without this args.decoder_normalize_before = True args.no_decoder_final_norm = getattr(args, "no_decoder_final_norm", False) args.adaptive_input = getattr(args, "adaptive_input", False) args.adaptive_input_factor = getattr(args, "adaptive_input_factor", 4) args.adaptive_input_cutoff = getattr(args, "adaptive_input_cutoff", None) args.tie_adaptive_weights = getattr(args, "tie_adaptive_weights", False) args.tie_adaptive_proj = getattr(args, "tie_adaptive_proj", False) args.no_scale_embedding = getattr(args, "no_scale_embedding", False) args.layernorm_embedding = getattr(args, "layernorm_embedding", False) args.checkpoint_activations = getattr(args, "checkpoint_activations", False) args.offload_activations = getattr(args, "offload_activations", False) if args.offload_activations: args.checkpoint_activations = True
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/models/language_modeling.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import argparse import importlib import os MODEL_REGISTRY = {} MODEL_DATACLASS_REGISTRY = {} ARCH_MODEL_REGISTRY = {} ARCH_MODEL_NAME_REGISTRY = {} ARCH_MODEL_INV_REGISTRY = {} ARCH_CONFIG_REGISTRY = {} # automatically import any Python files in the models/ directory models_dir = os.path.dirname(__file__) for file in os.listdir(models_dir): path = os.path.join(models_dir, file) if ( not file.startswith("_") and not file.startswith(".") and (file.endswith(".py") or os.path.isdir(path)) ): model_name = file[: file.find(".py")] if file.endswith(".py") else file module = importlib.import_module("models." + model_name) # extra `model_parser` for sphinx if model_name in MODEL_REGISTRY: parser = argparse.ArgumentParser(add_help=False) group_archs = parser.add_argument_group("Named architectures") group_archs.add_argument( "--arch", choices=ARCH_MODEL_INV_REGISTRY[model_name] ) group_args = parser.add_argument_group("Additional command-line arguments") MODEL_REGISTRY[model_name].add_args(group_args) globals()[model_name + "_parser"] = parser
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/models/__init__.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import logging from typing import Dict, List, Optional, Tuple import torch from fairseq import distributed_utils, utils from fairseq.distributed import fsdp_wrap from fairseq.models import ( FairseqEncoder, FairseqEncoderDecoderModel, register_model, register_model_architecture, ) from fairseq.models.transformer import Embedding from fairseq.modules import PositionalEmbedding from torch import Tensor from torchscale.architecture.config import DecoderConfig, EncoderConfig from torchscale.architecture.encoder import Encoder from .language_modeling import LMDecoder as MTDecoder logger = logging.getLogger(__name__) DEFAULT_MAX_SOURCE_POSITIONS = 1024 DEFAULT_MAX_TARGET_POSITIONS = 1024 DEFAULT_MIN_PARAMS_TO_WRAP = int(1e8) @register_model("mt") class TranslationModel(FairseqEncoderDecoderModel): def __init__(self, args, encoder, decoder): super().__init__(encoder, decoder) self.args = args @staticmethod def add_args(parser): """Add model-specific arguments to the parser.""" # fmt: off parser.add_argument('--activation-fn', choices=utils.get_available_activation_fns(), help='activation function to use') parser.add_argument('--dropout', type=float, metavar='D', help='dropout probability') parser.add_argument('--attention-dropout', type=float, metavar='D', help='dropout probability for attention weights') parser.add_argument('--activation-dropout', '--relu-dropout', type=float, metavar='D', help='dropout probability after activation in FFN.') parser.add_argument('--encoder-embed-path', type=str, metavar='STR', help='path to pre-trained encoder embedding') parser.add_argument('--encoder-embed-dim', type=int, metavar='N', help='encoder embedding dimension') parser.add_argument('--encoder-ffn-embed-dim', type=int, metavar='N', help='encoder embedding dimension for FFN') parser.add_argument('--encoder-layers', type=int, metavar='N', help='num encoder layers') parser.add_argument('--encoder-attention-heads', type=int, metavar='N', help='num encoder attention heads') parser.add_argument('--encoder-normalize-before', action='store_true', help='apply layernorm before each encoder block') parser.add_argument('--encoder-learned-pos', action='store_true', help='use learned positional embeddings in the encoder') parser.add_argument('--decoder-embed-path', type=str, metavar='STR', help='path to pre-trained decoder embedding') parser.add_argument('--decoder-embed-dim', type=int, metavar='N', help='decoder embedding dimension') parser.add_argument('--decoder-ffn-embed-dim', type=int, metavar='N', help='decoder embedding dimension for FFN') parser.add_argument('--decoder-layers', type=int, metavar='N', help='num decoder layers') parser.add_argument('--decoder-attention-heads', type=int, metavar='N', help='num decoder attention heads') parser.add_argument('--decoder-learned-pos', action='store_true', help='use learned positional embeddings in the decoder') parser.add_argument('--decoder-normalize-before', action='store_true', help='apply layernorm before each decoder block') parser.add_argument('--decoder-output-dim', type=int, metavar='N', help='decoder output dimension (extra linear layer ' 'if different from decoder embed dim') parser.add_argument('--share-decoder-input-output-embed', action='store_true', help='share decoder input and output embeddings') parser.add_argument('--share-all-embeddings', action='store_true', help='share encoder, decoder and output embeddings' ' (requires shared dictionary and embed dim)') parser.add_argument('--no-token-positional-embeddings', default=False, action='store_true', help='if set, disables positional embeddings (outside self attention)') parser.add_argument('--adaptive-softmax-cutoff', metavar='EXPR', help='comma separated list of adaptive softmax cutoff points. ' 'Must be used with adaptive_loss criterion'), parser.add_argument('--adaptive-softmax-dropout', type=float, metavar='D', help='sets adaptive softmax dropout for the tail projections') parser.add_argument('--layernorm-embedding', action='store_true', help='add layernorm to embedding') parser.add_argument('--no-scale-embedding', action='store_true', help='if True, dont scale embeddings') parser.add_argument('--checkpoint-activations', action='store_true', help='checkpoint activations at each layer, which saves GPU ' 'memory usage at the cost of some additional compute') parser.add_argument('--offload-activations', action='store_true', help='checkpoint activations at each layer, then save to gpu. Sets --checkpoint-activations.') # args for "Cross+Self-Attention for Transformer Models" (Peitz et al., 2019) parser.add_argument('--no-cross-attention', default=False, action='store_true', help='do not perform cross-attention') parser.add_argument('--cross-self-attention', default=False, action='store_true', help='perform cross+self-attention') # args for "Reducing Transformer Depth on Demand with Structured Dropout" (Fan et al., 2019) parser.add_argument('--encoder-layerdrop', type=float, metavar='D', default=0, help='LayerDrop probability for encoder') parser.add_argument('--decoder-layerdrop', type=float, metavar='D', default=0, help='LayerDrop probability for decoder') parser.add_argument('--encoder-layers-to-keep', default=None, help='which layers to *keep* when pruning as a comma-separated list') parser.add_argument('--decoder-layers-to-keep', default=None, help='which layers to *keep* when pruning as a comma-separated list') # args for Training with Quantization Noise for Extreme Model Compression ({Fan*, Stock*} et al., 2020) parser.add_argument('--quant-noise-pq', type=float, metavar='D', default=0, help='iterative PQ quantization noise at training time') parser.add_argument('--quant-noise-pq-block-size', type=int, metavar='D', default=8, help='block size of quantization noise at training time') parser.add_argument('--quant-noise-scalar', type=float, metavar='D', default=0, help='scalar quantization noise and scalar quantization at training time') # args for Fully Sharded Data Parallel (FSDP) training parser.add_argument( '--min-params-to-wrap', type=int, metavar='D', default=DEFAULT_MIN_PARAMS_TO_WRAP, help=( 'minimum number of params for a layer to be wrapped with FSDP() when ' 'training with --ddp-backend=fully_sharded. Smaller values will ' 'improve memory efficiency, but may make torch.distributed ' 'communication less efficient due to smaller input sizes. This option ' 'is set to 0 (i.e., always wrap) when --checkpoint-activations or ' '--offload-activations are passed.' ) ) # args for mixture-of-expert layers parser.add_argument('--moe-freq', type=int, metavar='D', default=0, help='Frequency at which we insert MoE Transformer layers') parser.add_argument('--encoder-moe-freq', type=int, metavar='D', default=0, help='Frequency at which we insert MoE Transformer encoder layers') parser.add_argument('--decoder-moe-freq', type=int, metavar='D', default=0, help='Frequency at which we insert MoE Transformer decoder layers') parser.add_argument('--moe-expert-count', type=int, metavar='D', default=0, help='Number of experts in each MoE Layer') parser.add_argument('--moe-gating-use-fp32', default=False, action='store_true', help="Use FP32 computations in MoE top2 gating function") parser.add_argument('--moe-second-expert-policy', type=str, default='sampling', help="policy for second expert, options: all/sampling/random") parser.add_argument( '--moe-normalize-gate-prob-before-dropping', default=False, action='store_true', help=( "whether to normalize gate probs before or after dropping experts " "for capacity and randomization" ) ) parser.add_argument('--moe-expert-ffn-dim', type=int, default=0, help="MoE Expert FFN dimension") parser.add_argument('--moe-top1-expert', default=False, action='store_true', help="Use top1 gate instead of top2") parser.add_argument( '--moe-eval-capacity-token-fraction', type=float, default=0.25, help=( "Fraction of tokens as capacity during validation" "if set to negative, use same as training. range: (0.0, 1.0]." ) ) parser.add_argument('--moe-normalize-expert-grad', type=str, default='world_size', help="Divide expert gradients by (1) 'world_size' (2) 'sqrt_world_size'") parser.add_argument('--use-moe-pad-mask', default=False, action='store_true', help="Don't route padding tokens to any expert") parser.add_argument('--use-xmoe', default=False, action='store_true', help="Enable X-Moe") parser.add_argument('--freeze-moe', default=False, action='store_true', help="Freeze MoE Params") parser.add_argument('--deepnorm', default=False, action='store_true', help="Enable DeepNorm") parser.add_argument('--subln', default=False, action='store_true', help="Enable SubLN") parser.add_argument('--pretrained-dense-mt-model-path', type=str, default='') # args for pseudo-MoE layers parser.add_argument('--alternate-ffn-embed-dim', type=int, default=0, help="FFN embed dim of alternate pseudo-MoE blocks") parser.add_argument('--rel-pos-buckets', type=int, default=0, help='') parser.add_argument('--max-rel-pos', type=int, default=0, help='') # fmt: on @classmethod def build_model(cls, args, task): """Build a new model instance.""" # make sure all arguments are present in older models base_architecture(args) if getattr(args, "max_source_positions", None) is None: args.max_source_positions = DEFAULT_MAX_SOURCE_POSITIONS if getattr(args, "max_target_positions", None) is None: args.max_target_positions = DEFAULT_MAX_TARGET_POSITIONS args.ddp_rank = distributed_utils.get_data_parallel_rank() src_dict, tgt_dict = task.source_dictionary, task.target_dictionary if args.share_all_embeddings: if src_dict != tgt_dict: raise ValueError("--share-all-embeddings requires a joined dictionary") if args.encoder_embed_dim != args.decoder_embed_dim: raise ValueError( "--share-all-embeddings requires --encoder-embed-dim to match --decoder-embed-dim" ) if args.decoder_embed_path and ( args.decoder_embed_path != args.encoder_embed_path ): raise ValueError( "--share-all-embeddings not compatible with --decoder-embed-path" ) encoder_embed_tokens = cls.build_embedding( args, src_dict, args.encoder_embed_dim, args.encoder_embed_path ) decoder_embed_tokens = encoder_embed_tokens args.share_decoder_input_output_embed = True else: encoder_embed_tokens = cls.build_embedding( args, src_dict, args.encoder_embed_dim, args.encoder_embed_path ) decoder_embed_tokens = cls.build_embedding( args, tgt_dict, args.decoder_embed_dim, args.decoder_embed_path ) if getattr(args, "offload_activations", False): args.checkpoint_activations = True # offloading implies checkpointing encoder_embed_positions = ( PositionalEmbedding( args.max_source_positions, args.encoder_embed_dim, src_dict.pad(), learned=args.encoder_learned_pos, ) if not args.no_token_positional_embeddings else None ) decoder_embed_positions = ( PositionalEmbedding( args.max_target_positions, args.decoder_embed_dim, tgt_dict.pad(), learned=args.decoder_learned_pos, ) if not args.no_token_positional_embeddings else None ) if args.share_decoder_input_output_embed: output_projection = torch.nn.Linear( decoder_embed_tokens.weight.shape[1], decoder_embed_tokens.weight.shape[0], bias=False, ) output_projection.weight = decoder_embed_tokens.weight else: output_projection = torch.nn.Linear( args.decoder_embed_dim, len(tgt_dict), bias=False ) torch.nn.init.normal_( output_projection.weight, mean=0, std=args.decoder_embed_dim**-0.5 ) encoder = cls.build_encoder( args, encoder_embed_tokens, encoder_embed_positions, src_dict, ) decoder = cls.build_decoder( args, decoder_embed_tokens, decoder_embed_positions, output_projection, tgt_dict, ) if not args.share_all_embeddings: min_params_to_wrap = getattr( args, "min_params_to_wrap", DEFAULT_MIN_PARAMS_TO_WRAP ) # fsdp_wrap is a no-op when --ddp-backend != fully_sharded encoder = fsdp_wrap(encoder, min_num_params=min_params_to_wrap) decoder = fsdp_wrap(decoder, min_num_params=min_params_to_wrap) return cls(args, encoder, decoder) @classmethod def build_embedding(cls, args, dictionary, embed_dim, path=None): num_embeddings = len(dictionary) padding_idx = dictionary.pad() emb = Embedding(num_embeddings, embed_dim, padding_idx) # if provided, load from preloaded dictionaries if path: embed_dict = utils.parse_embedding(path) utils.load_embedding(embed_dict, dictionary, emb) return emb @classmethod def build_encoder(cls, args, embed_tokens, embed_positions, dictionary): config = EncoderConfig() config.override(args) return MTEncoder( config, embed_tokens, embed_positions, is_encoder_decoder=True, dictionary=dictionary, ) @classmethod def build_decoder( cls, args, embed_tokens, embed_positions, output_projection, dictionary ): config = DecoderConfig() config.override(args) return MTDecoder( config, embed_tokens, embed_positions, output_projection, is_encoder_decoder=True, dictionary=dictionary, ) def forward( self, src_tokens, src_lengths, prev_output_tokens, return_all_hiddens: bool = False, features_only: bool = False, **kwargs ): encoder_out = self.encoder(src_tokens, return_all_hiddens=return_all_hiddens) decoder_out = self.decoder( prev_output_tokens, encoder_out=encoder_out, features_only=features_only, return_all_hiddens=return_all_hiddens, ) return decoder_out def get_normalized_probs( self, net_output: Tuple[Tensor, Optional[Dict[str, List[Optional[Tensor]]]]], log_probs: bool, sample: Optional[Dict[str, Tensor]] = None, ): """Get normalized probabilities (or log probs) from a net's output.""" return self.get_normalized_probs_scriptable(net_output, log_probs, sample) class MTEncoder(Encoder, FairseqEncoder): def forward(self, src_tokens, **kwargs): self_attn_padding_mask = src_tokens.eq(self.dictionary.pad()) return super().forward( src_tokens=src_tokens, encoder_padding_mask=self_attn_padding_mask, **kwargs ) def reorder_encoder_out(self, encoder_out, new_order): new_encoder_out = encoder_out["encoder_out"].index_select(0, new_order) new_encoder_embedding = encoder_out["encoder_embedding"].index_select( 0, new_order ) new_encoder_padding_mask = encoder_out["encoder_padding_mask"].index_select( 0, new_order ) encoder_states = encoder_out["encoder_states"] if len(encoder_states) > 0: for idx, state in enumerate(encoder_states): encoder_states[idx] = state.index_select(0, new_order) return { "encoder_out": new_encoder_out, # T x B x C "encoder_padding_mask": new_encoder_padding_mask, "encoder_embedding": new_encoder_embedding, # B x T x C "encoder_states": encoder_states, # List[T x B x C] } def max_positions(self): return self.embed_positions.max_positions @register_model_architecture("mt", "mt_base") def base_architecture(args): args.encoder_embed_path = getattr(args, "encoder_embed_path", None) args.encoder_embed_dim = getattr(args, "encoder_embed_dim", 512) args.encoder_ffn_embed_dim = getattr(args, "encoder_ffn_embed_dim", 2048) args.encoder_layers = getattr(args, "encoder_layers", 6) args.encoder_attention_heads = getattr(args, "encoder_attention_heads", 8) args.encoder_normalize_before = getattr(args, "encoder_normalize_before", False) args.encoder_learned_pos = getattr(args, "encoder_learned_pos", False) args.decoder_embed_path = getattr(args, "decoder_embed_path", None) args.decoder_embed_dim = getattr(args, "decoder_embed_dim", args.encoder_embed_dim) args.decoder_ffn_embed_dim = getattr( args, "decoder_ffn_embed_dim", args.encoder_ffn_embed_dim ) args.decoder_layers = getattr(args, "decoder_layers", 6) args.decoder_attention_heads = getattr(args, "decoder_attention_heads", 8) args.decoder_normalize_before = getattr(args, "decoder_normalize_before", False) args.decoder_learned_pos = getattr(args, "decoder_learned_pos", False) args.attention_dropout = getattr(args, "attention_dropout", 0.0) args.activation_dropout = getattr(args, "activation_dropout", 0.0) args.activation_fn = getattr(args, "activation_fn", "relu") args.dropout = getattr(args, "dropout", 0.1) args.adaptive_softmax_cutoff = getattr(args, "adaptive_softmax_cutoff", None) args.adaptive_softmax_dropout = getattr(args, "adaptive_softmax_dropout", 0) args.share_decoder_input_output_embed = getattr( args, "share_decoder_input_output_embed", False ) args.share_all_embeddings = getattr(args, "share_all_embeddings", False) args.no_token_positional_embeddings = getattr( args, "no_token_positional_embeddings", False ) args.adaptive_input = getattr(args, "adaptive_input", False) args.no_cross_attention = getattr(args, "no_cross_attention", False) args.cross_self_attention = getattr(args, "cross_self_attention", False) args.decoder_output_dim = getattr( args, "decoder_output_dim", args.decoder_embed_dim ) args.decoder_input_dim = getattr(args, "decoder_input_dim", args.decoder_embed_dim) args.no_scale_embedding = getattr(args, "no_scale_embedding", False) args.layernorm_embedding = getattr(args, "layernorm_embedding", False) args.tie_adaptive_weights = getattr(args, "tie_adaptive_weights", False) args.checkpoint_activations = getattr(args, "checkpoint_activations", False) args.offload_activations = getattr(args, "offload_activations", False) if args.offload_activations: args.checkpoint_activations = True args.encoder_layers_to_keep = getattr(args, "encoder_layers_to_keep", None) args.decoder_layers_to_keep = getattr(args, "decoder_layers_to_keep", None) args.encoder_layerdrop = getattr(args, "encoder_layerdrop", 0) args.decoder_layerdrop = getattr(args, "decoder_layerdrop", 0) args.quant_noise_pq = getattr(args, "quant_noise_pq", 0) args.quant_noise_pq_block_size = getattr(args, "quant_noise_pq_block_size", 8) args.quant_noise_scalar = getattr(args, "quant_noise_scalar", 0) args.is_moe = getattr(args, "is_moe", False) args.selected_expert_count = getattr(args, "selected_expert_count", 2)
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/models/machine_translation.py
# Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] import logging from dataclasses import dataclass, field from typing import Optional import torch import torch.nn as nn import torch.nn.functional as F from fairseq import utils from fairseq.dataclass import ChoiceEnum, FairseqDataclass from fairseq.models import BaseFairseqModel, register_model, register_model_architecture from fairseq.models.squad import SQuADHead from fairseq.models.transformer import DEFAULT_MIN_PARAMS_TO_WRAP, Embedding from fairseq.modules import PositionalEmbedding from omegaconf import II try: from apex.normalization import FusedLayerNorm as LayerNorm except ModuleNotFoundError: from torch.nn import LayerNorm from torchscale.architecture.config import EncoderConfig from .machine_translation import MTEncoder as Encoder DEFAULT_MAX_SOURCE_POSITIONS = 1024 logger = logging.getLogger(__name__) @dataclass class BertConfig(FairseqDataclass): activation_fn: ChoiceEnum(utils.get_available_activation_fns()) = field( default="relu", metadata={"help": "activation function to use"} ) dropout: float = field(default=0.1, metadata={"help": "dropout probability"}) attention_dropout: float = field( default=0.0, metadata={"help": "dropout probability for attention weights"} ) activation_dropout: float = field( default=0.0, metadata={"help": "dropout probability after activation in FFN."} ) encoder_embed_dim: int = field( default=512, metadata={"help": "encoder embedding dimension"} ) encoder_output_dim: int = field( default=512, metadata={"help": "encoder output dimension"} ) encoder_input_dim: int = field( default=512, metadata={"help": "encoder input dimension"} ) encoder_ffn_embed_dim: int = field( default=2048, metadata={"help": "encoder embedding dimension for FFN"} ) encoder_layers: int = field(default=6, metadata={"help": "num encoder layers"}) encoder_attention_heads: int = field( default=8, metadata={"help": "num encoder attention heads"} ) encoder_normalize_before: bool = field( default=False, metadata={"help": "apply layernorm before each encoder block"} ) no_encoder_final_norm: bool = field( default=False, metadata={"help": "don't add an extra layernorm after the last encoder block"}, ) no_token_positional_embeddings: bool = field( default=False, metadata={ "help": "if set, disables positional embeddings (outside self attention)" }, ) share_encoder_input_output_embed: bool = field( default=False, metadata={"help": "share encoder input and output embeddings"} ) encoder_learned_pos: bool = field( default=False, metadata={"help": "use learned positional embeddings in the encoder"}, ) layernorm_embedding: bool = field( default=False, metadata={"help": "add layernorm to embedding"} ) no_scale_embedding: bool = field( default=False, metadata={"help": "if True, dont scale embeddings"} ) checkpoint_activations: bool = field( default=False, metadata={"help": "checkpoint activations at each layer"} ) offload_activations: bool = field( default=False, metadata={"help": "move checkpointed activations to CPU after they are used."}, ) # config for "Reducing Transformer Depth on Demand with Structured Dropout" (Fan et al., 2019) encoder_layerdrop: float = field( default=0.0, metadata={"help": "LayerDrop probability for encoder"} ) encoder_layers_to_keep: Optional[str] = field( default=None, metadata={ "help": "which layers to *keep* when pruning as a comma-separated list" }, ) # config for Fully Sharded Data Parallel (FSDP) training min_params_to_wrap: int = field( default=DEFAULT_MIN_PARAMS_TO_WRAP, metadata={ "help": ( "minimum number of params for a layer to be wrapped with FSDP() when " "training with --ddp-backend=fully_sharded. Smaller values will " "improve memory efficiency, but may make torch.distributed " "communication less efficient due to smaller input sizes. This option " "is set to 0 (i.e., always wrap) when --checkpoint-activations or " "--offload-activations are passed." ) }, ) max_source_positions: int = field( default=1024, metadata={"help": "max source positions"} ) pooler_activation_fn: ChoiceEnum(utils.get_available_activation_fns()) = field( default="relu", metadata={"help": "activation function to use for pooler layer"} ) pooler_dropout: float = field( default=0.0, metadata={"help": "dropout probability in the masked_lm pooler layers"}, ) # options from other parts of the config # add_bos_token: bool = II("task.add_bos_token") # tokens_per_sample: int = II("task.tokens_per_sample") tpu: bool = II("common.tpu") rel_pos_buckets: int = field(default=0, metadata={"help": ""}) max_rel_pos: int = field(default=0, metadata={"help": ""}) use_xmoe: Optional[bool] = field( default=False, ) moe_freq: int = field( default=0, metadata={"help": "Frequency at which we insert MoE Transformer layers"}, ) moe_expert_count: int = field( default=0, metadata={"help": "Number of experts in each MoE Layer"} ) moe_gating_use_fp32: bool = field( default=False, metadata={"help": "Use FP32 computations in MoE top2 gating function"}, ) moe_second_expert_policy: str = field( default="sampling", metadata={"help": "policy for second expert, options: all/sampling/random"}, ) moe_normalize_gate_prob_before_dropping: bool = field( default=False, metadata={ "help": "whether to normalize gate probs before or after dropping experts for capacity and randomization" }, ) moe_expert_ffn_dim: Optional[int] = field( default=None, metadata={"help": "MoE expert FFN dimension"} ) moe_top1_expert: Optional[bool] = field( default=False, metadata={"help": "Use top1 gate instead of top2"} ) moe_eval_capacity_token_fraction: Optional[float] = field( default=0.25, metadata={ "help": ( "Default: 0.25, Fraction of tokens as capacity during validation, " "if set to negative, use same as training. range: (0.0, 1.0]." ) }, ) moe_normalize_expert_grad: Optional[str] = field( default="world_size", metadata={ "help": "Divide expert gradients by (1) 'world_size' (2) 'sqrt_world_size'" }, ) record_a2a_perf_stats: Optional[bool] = field( default=False, metadata={"help": "records all to all perf stats during distributed training"}, ) dummy_a2a: Optional[bool] = field( default=False, metadata={ "help": "By passes all to all during distributed training by returning the input buffer as output" }, ) moe_batch_prioritized_routing: Optional[bool] = field( default=False, metadata={ "help": "if true orders token by the gate prob before capacity dropping." }, ) ddp_rank: int = II("distributed_training.distributed_rank") deepnorm: Optional[bool] = field( default=False, ) subln: Optional[bool] = field( default=False, ) @register_model("mlm", dataclass=BertConfig) class BertModel(BaseFairseqModel): def __init__(self, args, encoder): super().__init__() self.args = args self.encoder = encoder self.padding_idx = self.encoder.embed_tokens.padding_idx self.classification_heads = nn.ModuleDict() @classmethod def build_model(cls, args, task): """Build a new model instance.""" args.max_source_positions = getattr( args, "max_source_positions", DEFAULT_MAX_SOURCE_POSITIONS ) embed_tokens = cls.build_embedding( args, task.dictionary, args.encoder_embed_dim ) embed_positions = ( PositionalEmbedding( args.max_source_positions, args.encoder_embed_dim, task.dictionary.pad(), learned=args.encoder_learned_pos, ) if not args.no_token_positional_embeddings else None ) lm_head = cls.build_lm_head( args, args.encoder_embed_dim, len(task.dictionary), args.activation_fn, weight=embed_tokens.weight, ) config = EncoderConfig() config.override(args) encoder = Encoder( config, embed_tokens=embed_tokens, embed_positions=embed_positions, output_projection=lm_head, is_encoder_decoder=False, dictionary=task.dictionary, ) return cls(args, encoder) @classmethod def build_embedding(cls, args, dictionary, embed_dim, path=None): embed_tokens = Embedding(len(dictionary), embed_dim, dictionary.pad()) return embed_tokens @classmethod def build_lm_head(cls, args, embed_dim, output_dim, activation_fn, weight): return LMHead(embed_dim, output_dim, activation_fn, weight) def output_layer(self, features, masked_tokens=None): return self.encoder.output_projection(features, masked_tokens=masked_tokens) def register_classification_head( self, name, num_classes=None, inner_dim=None, **kwargs ): """Register a classification head.""" if name in self.classification_heads: prev_num_classes = self.classification_heads[name].out_proj.out_features prev_inner_dim = self.classification_heads[name].dense.out_features if num_classes != prev_num_classes or inner_dim != prev_inner_dim: logger.warning( 're-registering head "{}" with num_classes {} (prev: {}) ' "and inner_dim {} (prev: {})".format( name, num_classes, prev_num_classes, inner_dim, prev_inner_dim ) ) self.classification_heads[name] = ClassificationHead( self.args.encoder_embed_dim, inner_dim or self.args.encoder_embed_dim, num_classes, self.args.pooler_activation_fn, self.args.pooler_dropout, ) def register_question_answering_head(self, name, num_classes=None): self.classification_heads[name] = SQuADHead( self.args.encoder_embed_dim, ) def upgrade_state_dict_named(self, state_dict, name): prefix = name + "." if name != "" else "" # upgrade children modules super().upgrade_state_dict_named(state_dict, name) # Handle new classification heads present in the state dict. current_head_names = ( [] if not hasattr(self, "classification_heads") else self.classification_heads.keys() ) keys_to_delete = [] for k in state_dict.keys(): if not k.startswith(prefix + "classification_heads."): continue head_name = k[len(prefix + "classification_heads.") :].split(".")[0] # noqa: E203 num_classes = state_dict[ prefix + "classification_heads." + head_name + ".out_proj.weight" ].size(0) inner_dim = state_dict[ prefix + "classification_heads." + head_name + ".dense.weight" ].size(0) if getattr(self.args, "load_checkpoint_heads", False): if head_name not in current_head_names: self.register_classification_head(head_name, num_classes, inner_dim) else: if head_name not in current_head_names: logger.warning( "deleting classification head ({}) from checkpoint " "not present in current model: {}".format(head_name, k) ) keys_to_delete.append(k) elif ( num_classes != self.classification_heads[head_name].out_proj.out_features or inner_dim != self.classification_heads[head_name].dense.out_features ): logger.warning( "deleting classification head ({}) from checkpoint " "with different dimensions than current model: {}".format( head_name, k ) ) keys_to_delete.append(k) for k in keys_to_delete: del state_dict[k] # Copy any newly-added classification heads into the state dict # with their current weights. if hasattr(self, "classification_heads"): cur_state = self.classification_heads.state_dict() for k, v in cur_state.items(): if prefix + "classification_heads." + k not in state_dict: logger.info("Overwriting " + prefix + "classification_heads." + k) state_dict[prefix + "classification_heads." + k] = v def get_normalized_probs_scriptable( self, net_output, log_probs, sample = None, ): logits = net_output[0] if log_probs: return utils.log_softmax(logits, dim=-1) else: return utils.softmax(logits, dim=-1) def forward( self, src_tokens=None, features_only=False, return_all_hiddens=False, classification_head_name=None, masked_tokens=None, **kwargs ): encoder_out = self.encoder( src_tokens, features_only=True, return_all_hiddens=return_all_hiddens ) x, extra = encoder_out["encoder_out"], encoder_out if classification_head_name is not None: x = self.classification_heads[classification_head_name](x) elif not features_only: x = self.output_layer(x, masked_tokens=masked_tokens) return x, extra class ClassificationHead(nn.Module): """Head for sentence-level classification tasks.""" def __init__( self, input_dim, inner_dim, num_classes, activation_fn, pooler_dropout, ): super().__init__() self.dense = nn.Linear(input_dim, inner_dim) self.activation_fn = utils.get_activation_fn(activation_fn) self.dropout = nn.Dropout(p=pooler_dropout) self.out_proj = nn.Linear(inner_dim, num_classes) def forward(self, features, **kwargs): x = features[:, 0, :] # take <s> token (equiv. to [CLS]) x = self.dropout(x) x = self.dense(x) x = self.activation_fn(x.float()).type_as(x) x = self.dropout(x) x = self.out_proj(x) return x class LMHead(nn.Module): """Head for masked language modeling.""" def __init__(self, embed_dim, output_dim, activation_fn, weight=None): super().__init__() self.dense = nn.Linear(embed_dim, embed_dim) self.activation_fn = utils.get_activation_fn(activation_fn) self.layer_norm = LayerNorm(embed_dim) if weight is None: weight = nn.Linear(embed_dim, output_dim, bias=False).weight self.weight = weight self.bias = nn.Parameter(torch.zeros(output_dim)) def forward(self, features, masked_tokens=None, **kwargs): # Only project the masked tokens while training, # saves both memory and computation if masked_tokens is not None: features = features[masked_tokens, :] x = self.dense(features) x = self.activation_fn(x.float()).type_as(x) x = self.layer_norm(x) # project back to size of vocabulary with bias x = F.linear(x, self.weight) + self.bias return x @register_model_architecture("mlm", "mlm_base") def base_unilm_architecture(args): if hasattr(args, "encoder_final_norm"): args.no_encoder_final_norm = not args.encoder_final_norm args.dropout = getattr(args, "dropout", 0.1) args.attention_dropout = getattr(args, "attention_dropout", 0.0) args.activation_dropout = getattr(args, "activation_dropout", 0.0) args.pooler_dropout = getattr(args, "pooler_dropout", 0.0) args.encoder_embed_dim = getattr(args, "encoder_embed_dim", 768) args.encoder_ffn_embed_dim = getattr(args, "encoder_ffn_embed_dim", 3072) args.encoder_layers = getattr(args, "encoder_layers", 12) args.encoder_attention_heads = getattr(args, "encoder_attention_heads", 12) args.encoder_learned_pos = getattr(args, "encoder_learned_pos", True) args.activation_fn = getattr(args, "activation_fn", "gelu") args.pooler_activation_fn = getattr(args, "pooler_activation_fn", "tanh") args.encoder_layerdrop = getattr(args, "encoder_layerdrop", 0) args.encoder_layers_to_keep = getattr(args, "encoder_layers_to_keep", None) # args.add_bos_token = getattr(args, "add_bos_token", False) args.no_token_positional_embeddings = getattr( args, "no_token_positional_embeddings", False ) args.share_encoder_input_output_embed = getattr( args, "share_encoder_input_output_embed", True ) args.encoder_output_dim = getattr( args, "encoder_output_dim", args.encoder_embed_dim ) args.encoder_input_dim = getattr(args, "encoder_input_dim", args.encoder_embed_dim) # Model training is not stable without this args.encoder_normalize_before = getattr(args, "encoder_normalize_before", False) args.no_encoder_final_norm = getattr(args, "no_encoder_final_norm", False) args.no_scale_embedding = getattr(args, "no_scale_embedding", True) args.layernorm_embedding = getattr(args, "layernorm_embedding", True) args.checkpoint_activations = getattr(args, "checkpoint_activations", False) args.offload_activations = getattr(args, "offload_activations", False) if args.offload_activations: args.checkpoint_activations = True
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/models/bert.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import math import torch import torch.nn.functional as F from fairseq import metrics, utils from fairseq.criterions import MoECriterion, register_criterion, MoECriterionConfig @register_criterion("masked_lm_moe_cross_entropy", dataclass=MoECriterionConfig) class MaskedLMMoECrossEntropyCriterion(MoECriterion): def compute_inner_loss(self, model, sample, reduce=True): masked_tokens = sample["target"].ne(self.padding_idx) sample_size = masked_tokens.int().sum() masked_tokens = torch.where( masked_tokens.any(), masked_tokens, masked_tokens.new([True]), ) net_output = model(**sample["net_input"], masked_tokens=masked_tokens) lprobs = model.get_normalized_probs(net_output, log_probs=True) lprobs = lprobs.view(-1, lprobs.size(-1)) target = model.get_targets(sample, net_output) if masked_tokens is not None: target = target[masked_tokens] nll_loss = F.nll_loss( lprobs, target.view(-1), ignore_index=self.padding_idx, reduction="sum" if reduce else "none", ) logging_output = { "inner_loss": nll_loss.data, "ntokens": sample["ntokens"], "nsentences": sample["target"].size(0), "sample_size": sample_size, } return net_output, nll_loss, sample_size, logging_output @staticmethod def reduce_metrics(logging_outputs) -> None: """Aggregate logging outputs from data parallel training.""" MaskedLMMoECrossEntropyCriterion.reduce_moe_metrics(logging_outputs) loss_sum = sum(log.get("inner_loss", 0) for log in logging_outputs) ntokens = sum(log.get("ntokens", 0) for log in logging_outputs) sample_size = sum(log.get("sample_size", 0) for log in logging_outputs) # we divide by log(2) to convert the loss from base e to base 2 metrics.log_scalar( "inner_loss", loss_sum / sample_size / math.log(2), sample_size, round=3 ) if sample_size != ntokens: metrics.log_scalar( "nll_loss", loss_sum / ntokens / math.log(2), ntokens, round=3 ) metrics.log_derived( "ppl", lambda meters: utils.get_perplexity(meters["nll_loss"].avg) ) else: metrics.log_derived( "ppl", lambda meters: utils.get_perplexity(meters["inner_loss"].avg) )
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/criterions/masked_lm_moe.py
import importlib import os # automatically import any Python files in the criterions/ directory for file in sorted(os.listdir(os.path.dirname(__file__))): if file.endswith(".py") and not file.startswith("_"): file_name = file[: file.find(".py")] importlib.import_module("criterions." + file_name)
GeneSplice-main
GeneSplice/torchscale/examples/fairseq/criterions/__init__.py
import sys import warnings import os from packaging.version import parse, Version from setuptools import setup, find_packages import subprocess import torch from torch.utils.cpp_extension import ( BuildExtension, CppExtension, CUDAExtension, CUDA_HOME, load, ) # ninja build does not work unless include_dirs are abs path this_dir = os.path.dirname(os.path.abspath(__file__)) def get_cuda_bare_metal_version(cuda_dir): raw_output = subprocess.check_output([cuda_dir + "/bin/nvcc", "-V"], universal_newlines=True) output = raw_output.split() release_idx = output.index("release") + 1 bare_metal_version = parse(output[release_idx].split(",")[0]) return raw_output, bare_metal_version def check_cuda_torch_binary_vs_bare_metal(cuda_dir): raw_output, bare_metal_version = get_cuda_bare_metal_version(cuda_dir) torch_binary_version = parse(torch.version.cuda) print("\nCompiling cuda extensions with") print(raw_output + "from " + cuda_dir + "/bin\n") if (bare_metal_version != torch_binary_version): raise RuntimeError( "Cuda extensions are being compiled with a version of Cuda that does " "not match the version used to compile Pytorch binaries. " "Pytorch binaries were compiled with Cuda {}.\n".format(torch.version.cuda) + "In some cases, a minor-version mismatch will not cause later errors: " "https://github.com/NVIDIA/apex/pull/323#discussion_r287021798. " "You can try commenting out this check (at your own risk)." ) def raise_if_cuda_home_none(global_option: str) -> None: if CUDA_HOME is not None: return raise RuntimeError( f"{global_option} was requested, but nvcc was not found. Are you sure your environment has nvcc available? " "If you're installing within a container from https://hub.docker.com/r/pytorch/pytorch, " "only images whose names contain 'devel' will provide nvcc." ) def check_cudnn_version_and_warn(global_option: str, required_cudnn_version: int) -> bool: cudnn_available = torch.backends.cudnn.is_available() cudnn_version = torch.backends.cudnn.version() if cudnn_available else None if not (cudnn_available and (cudnn_version >= required_cudnn_version)): warnings.warn( f"Skip `{global_option}` as it requires cuDNN {required_cudnn_version} or later, " f"but {'cuDNN is not available' if not cudnn_available else cudnn_version}" ) return False return True if not torch.cuda.is_available(): # https://github.com/NVIDIA/apex/issues/486 # Extension builds after https://github.com/pytorch/pytorch/pull/23408 attempt to query torch.cuda.get_device_capability(), # which will fail if you are compiling in an environment without visible GPUs (e.g. during an nvidia-docker build command). print( "\nWarning: Torch did not find available GPUs on this system.\n", "If your intention is to cross-compile, this is not an error.\n" "By default, Apex will cross-compile for Pascal (compute capabilities 6.0, 6.1, 6.2),\n" "Volta (compute capability 7.0), Turing (compute capability 7.5),\n" "and, if the CUDA version is >= 11.0, Ampere (compute capability 8.0).\n" "If you wish to cross-compile for a single specific architecture,\n" 'export TORCH_CUDA_ARCH_LIST="compute capability" before running setup.py.\n', ) if os.environ.get("TORCH_CUDA_ARCH_LIST", None) is None and CUDA_HOME is not None: _, bare_metal_version = get_cuda_bare_metal_version(CUDA_HOME) if bare_metal_version >= Version("11.8"): os.environ["TORCH_CUDA_ARCH_LIST"] = "6.0;6.1;6.2;7.0;7.5;8.0;8.6;9.0" elif bare_metal_version >= Version("11.1"): os.environ["TORCH_CUDA_ARCH_LIST"] = "6.0;6.1;6.2;7.0;7.5;8.0;8.6" elif bare_metal_version == Version("11.0"): os.environ["TORCH_CUDA_ARCH_LIST"] = "6.0;6.1;6.2;7.0;7.5;8.0" else: os.environ["TORCH_CUDA_ARCH_LIST"] = "6.0;6.1;6.2;7.0;7.5" print("\n\ntorch.__version__ = {}\n\n".format(torch.__version__)) TORCH_MAJOR = int(torch.__version__.split(".")[0]) TORCH_MINOR = int(torch.__version__.split(".")[1]) if TORCH_MAJOR == 0 and TORCH_MINOR < 4: raise RuntimeError( "Apex requires Pytorch 0.4 or newer.\nThe latest stable release can be obtained from https://pytorch.org/" ) cmdclass = {} ext_modules = [] extras = {} if "--cpp_ext" in sys.argv or "--cuda_ext" in sys.argv: if TORCH_MAJOR == 0: raise RuntimeError( "--cpp_ext requires Pytorch 1.0 or later, " "found torch.__version__ = {}".format(torch.__version__) ) if "--cpp_ext" in sys.argv: sys.argv.remove("--cpp_ext") ext_modules.append(CppExtension("apex_C", ["csrc/flatten_unflatten.cpp"])) # Set up macros for forward/backward compatibility hack around # https://github.com/pytorch/pytorch/commit/4404762d7dd955383acee92e6f06b48144a0742e # and # https://github.com/NVIDIA/apex/issues/456 # https://github.com/pytorch/pytorch/commit/eb7b39e02f7d75c26d8a795ea8c7fd911334da7e#diff-4632522f237f1e4e728cb824300403ac version_ge_1_1 = [] if (TORCH_MAJOR > 1) or (TORCH_MAJOR == 1 and TORCH_MINOR > 0): version_ge_1_1 = ["-DVERSION_GE_1_1"] version_ge_1_3 = [] if (TORCH_MAJOR > 1) or (TORCH_MAJOR == 1 and TORCH_MINOR > 2): version_ge_1_3 = ["-DVERSION_GE_1_3"] version_ge_1_5 = [] if (TORCH_MAJOR > 1) or (TORCH_MAJOR == 1 and TORCH_MINOR > 4): version_ge_1_5 = ["-DVERSION_GE_1_5"] version_dependent_macros = version_ge_1_1 + version_ge_1_3 + version_ge_1_5 _, bare_metal_version = get_cuda_bare_metal_version(CUDA_HOME) if "--distributed_adam" in sys.argv: sys.argv.remove("--distributed_adam") raise_if_cuda_home_none("--distributed_adam") ext_modules.append( CUDAExtension( name="distributed_adam_cuda", sources=[ "apex/contrib/csrc/optimizers/multi_tensor_distopt_adam.cpp", "apex/contrib/csrc/optimizers/multi_tensor_distopt_adam_kernel.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3", "--use_fast_math"] + version_dependent_macros, }, ) ) if "--distributed_lamb" in sys.argv: sys.argv.remove("--distributed_lamb") raise_if_cuda_home_none("--distributed_lamb") ext_modules.append( CUDAExtension( name="distributed_lamb_cuda", sources=[ "apex/contrib/csrc/optimizers/multi_tensor_distopt_lamb.cpp", "apex/contrib/csrc/optimizers/multi_tensor_distopt_lamb_kernel.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3", "--use_fast_math"] + version_dependent_macros, }, ) ) if "--cuda_ext" in sys.argv: sys.argv.remove("--cuda_ext") raise_if_cuda_home_none("--cuda_ext") check_cuda_torch_binary_vs_bare_metal(CUDA_HOME) ext_modules.append( CUDAExtension( name="amp_C", sources=[ "csrc/amp_C_frontend.cpp", "csrc/multi_tensor_sgd_kernel.cu", "csrc/multi_tensor_scale_kernel.cu", "csrc/multi_tensor_axpby_kernel.cu", "csrc/multi_tensor_l2norm_kernel.cu", "csrc/multi_tensor_l2norm_kernel_mp.cu", "csrc/multi_tensor_l2norm_scale_kernel.cu", "csrc/multi_tensor_lamb_stage_1.cu", "csrc/multi_tensor_lamb_stage_2.cu", "csrc/multi_tensor_adam.cu", "csrc/multi_tensor_adagrad.cu", "csrc/multi_tensor_novograd.cu", "csrc/multi_tensor_lamb.cu", "csrc/multi_tensor_lamb_mp.cu", ], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": [ "-lineinfo", "-O3", # '--resource-usage', "--use_fast_math", ] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="syncbn", sources=["csrc/syncbn.cpp", "csrc/welford.cu"], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3"] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="fused_layer_norm_cuda", sources=["csrc/layer_norm_cuda.cpp", "csrc/layer_norm_cuda_kernel.cu"], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-maxrregcount=50", "-O3", "--use_fast_math"] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="mlp_cuda", sources=["csrc/mlp.cpp", "csrc/mlp_cuda.cu"], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3"] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="fused_dense_cuda", sources=["csrc/fused_dense.cpp", "csrc/fused_dense_cuda.cu"], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3"] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="scaled_upper_triang_masked_softmax_cuda", sources=[ "csrc/megatron/scaled_upper_triang_masked_softmax.cpp", "csrc/megatron/scaled_upper_triang_masked_softmax_cuda.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "--expt-relaxed-constexpr", "--expt-extended-lambda", ] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="generic_scaled_masked_softmax_cuda", sources=[ "csrc/megatron/generic_scaled_masked_softmax.cpp", "csrc/megatron/generic_scaled_masked_softmax_cuda.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "--expt-relaxed-constexpr", "--expt-extended-lambda", ] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="scaled_masked_softmax_cuda", sources=["csrc/megatron/scaled_masked_softmax.cpp", "csrc/megatron/scaled_masked_softmax_cuda.cu"], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "--expt-relaxed-constexpr", "--expt-extended-lambda", ] + version_dependent_macros, }, ) ) ext_modules.append( CUDAExtension( name="scaled_softmax_cuda", sources=["csrc/megatron/scaled_softmax.cpp", "csrc/megatron/scaled_softmax_cuda.cu"], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "--expt-relaxed-constexpr", "--expt-extended-lambda", ] + version_dependent_macros, }, ) ) if bare_metal_version >= Version("11.0"): cc_flag = [] cc_flag.append("-gencode") cc_flag.append("arch=compute_70,code=sm_70") cc_flag.append("-gencode") cc_flag.append("arch=compute_80,code=sm_80") if bare_metal_version >= Version("11.1"): cc_flag.append("-gencode") cc_flag.append("arch=compute_86,code=sm_86") if bare_metal_version >= Version("11.8"): cc_flag.append("-gencode") cc_flag.append("arch=compute_90,code=sm_90") ext_modules.append( CUDAExtension( name="fused_weight_gradient_mlp_cuda", include_dirs=[os.path.join(this_dir, "csrc")], sources=[ "csrc/megatron/fused_weight_gradient_dense.cpp", "csrc/megatron/fused_weight_gradient_dense_cuda.cu", "csrc/megatron/fused_weight_gradient_dense_16bit_prec_cuda.cu", ], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "--expt-relaxed-constexpr", "--expt-extended-lambda", "--use_fast_math", ] + version_dependent_macros + cc_flag, }, ) ) if "--permutation_search" in sys.argv: sys.argv.remove("--permutation_search") if CUDA_HOME is None: raise RuntimeError("--permutation_search was requested, but nvcc was not found. Are you sure your environment has nvcc available? If you're installing within a container from https://hub.docker.com/r/pytorch/pytorch, only images whose names contain 'devel' will provide nvcc.") else: cc_flag = ['-Xcompiler', '-fPIC', '-shared'] ext_modules.append( CUDAExtension(name='permutation_search_cuda', sources=['apex/contrib/sparsity/permutation_search_kernels/CUDA_kernels/permutation_search_kernels.cu'], include_dirs=[os.path.join(this_dir, 'apex', 'contrib', 'sparsity', 'permutation_search_kernels', 'CUDA_kernels')], extra_compile_args={'cxx': ['-O3'] + version_dependent_macros, 'nvcc':['-O3'] + version_dependent_macros + cc_flag})) if "--bnp" in sys.argv: sys.argv.remove("--bnp") raise_if_cuda_home_none("--bnp") ext_modules.append( CUDAExtension( name="bnp", sources=[ "apex/contrib/csrc/groupbn/batch_norm.cu", "apex/contrib/csrc/groupbn/ipc.cu", "apex/contrib/csrc/groupbn/interface.cpp", "apex/contrib/csrc/groupbn/batch_norm_add_relu.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": [] + version_dependent_macros, "nvcc": [ "-DCUDA_HAS_FP16=1", "-D__CUDA_NO_HALF_OPERATORS__", "-D__CUDA_NO_HALF_CONVERSIONS__", "-D__CUDA_NO_HALF2_OPERATORS__", ] + version_dependent_macros, }, ) ) if "--xentropy" in sys.argv: sys.argv.remove("--xentropy") raise_if_cuda_home_none("--xentropy") ext_modules.append( CUDAExtension( name="xentropy_cuda", sources=["apex/contrib/csrc/xentropy/interface.cpp", "apex/contrib/csrc/xentropy/xentropy_kernel.cu"], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3"] + version_dependent_macros, }, ) ) if "--focal_loss" in sys.argv: sys.argv.remove("--focal_loss") raise_if_cuda_home_none("--focal_loss") ext_modules.append( CUDAExtension( name='focal_loss_cuda', sources=[ 'apex/contrib/csrc/focal_loss/focal_loss_cuda.cpp', 'apex/contrib/csrc/focal_loss/focal_loss_cuda_kernel.cu', ], include_dirs=[os.path.join(this_dir, 'csrc')], extra_compile_args={ 'cxx': ['-O3'] + version_dependent_macros, 'nvcc':['-O3', '--use_fast_math', '--ftz=false'] + version_dependent_macros, }, ) ) if "--index_mul_2d" in sys.argv: sys.argv.remove("--index_mul_2d") raise_if_cuda_home_none("--index_mul_2d") ext_modules.append( CUDAExtension( name='fused_index_mul_2d', sources=[ 'apex/contrib/csrc/index_mul_2d/index_mul_2d_cuda.cpp', 'apex/contrib/csrc/index_mul_2d/index_mul_2d_cuda_kernel.cu', ], include_dirs=[os.path.join(this_dir, 'csrc')], extra_compile_args={ 'cxx': ['-O3'] + version_dependent_macros, 'nvcc':['-O3', '--use_fast_math', '--ftz=false'] + version_dependent_macros, }, ) ) if "--deprecated_fused_adam" in sys.argv: sys.argv.remove("--deprecated_fused_adam") raise_if_cuda_home_none("--deprecated_fused_adam") ext_modules.append( CUDAExtension( name="fused_adam_cuda", sources=[ "apex/contrib/csrc/optimizers/fused_adam_cuda.cpp", "apex/contrib/csrc/optimizers/fused_adam_cuda_kernel.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3", "--use_fast_math"] + version_dependent_macros, }, ) ) if "--deprecated_fused_lamb" in sys.argv: sys.argv.remove("--deprecated_fused_lamb") raise_if_cuda_home_none("--deprecated_fused_lamb") ext_modules.append( CUDAExtension( name="fused_lamb_cuda", sources=[ "apex/contrib/csrc/optimizers/fused_lamb_cuda.cpp", "apex/contrib/csrc/optimizers/fused_lamb_cuda_kernel.cu", "csrc/multi_tensor_l2norm_kernel.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3", "--use_fast_math"] + version_dependent_macros, }, ) ) # Check, if ATen/CUDAGeneratorImpl.h is found, otherwise use ATen/cuda/CUDAGeneratorImpl.h # See https://github.com/pytorch/pytorch/pull/70650 generator_flag = [] torch_dir = torch.__path__[0] if os.path.exists(os.path.join(torch_dir, "include", "ATen", "CUDAGeneratorImpl.h")): generator_flag = ["-DOLD_GENERATOR_PATH"] if "--fast_layer_norm" in sys.argv: sys.argv.remove("--fast_layer_norm") raise_if_cuda_home_none("--fast_layer_norm") cc_flag = [] cc_flag.append("-gencode") cc_flag.append("arch=compute_70,code=sm_70") if bare_metal_version >= Version("11.0"): cc_flag.append("-gencode") cc_flag.append("arch=compute_80,code=sm_80") if bare_metal_version >= Version("11.8"): cc_flag.append("-gencode") cc_flag.append("arch=compute_90,code=sm_90") ext_modules.append( CUDAExtension( name="fast_layer_norm", sources=[ "apex/contrib/csrc/layer_norm/ln_api.cpp", "apex/contrib/csrc/layer_norm/ln_fwd_cuda_kernel.cu", "apex/contrib/csrc/layer_norm/ln_bwd_semi_cuda_kernel.cu", ], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros + generator_flag, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "-U__CUDA_NO_BFLOAT16_OPERATORS__", "-U__CUDA_NO_BFLOAT16_CONVERSIONS__", "-U__CUDA_NO_BFLOAT162_OPERATORS__", "-U__CUDA_NO_BFLOAT162_CONVERSIONS__", "-I./apex/contrib/csrc/layer_norm/", "--expt-relaxed-constexpr", "--expt-extended-lambda", "--use_fast_math", ] + version_dependent_macros + generator_flag + cc_flag, }, include_dirs=[os.path.join(this_dir, "apex/contrib/csrc/layer_norm")], ) ) if "--fmha" in sys.argv: sys.argv.remove("--fmha") raise_if_cuda_home_none("--fmha") if bare_metal_version < Version("11.0"): raise RuntimeError("--fmha only supported on sm_80 and sm_90 GPUs") cc_flag = [] cc_flag.append("-gencode") cc_flag.append("arch=compute_80,code=sm_80") if bare_metal_version >= Version("11.8"): cc_flag.append("-gencode") cc_flag.append("arch=compute_90,code=sm_90") ext_modules.append( CUDAExtension( name="fmhalib", sources=[ "apex/contrib/csrc/fmha/fmha_api.cpp", "apex/contrib/csrc/fmha/src/fmha_fill.cu", "apex/contrib/csrc/fmha/src/fmha_noloop_reduce.cu", "apex/contrib/csrc/fmha/src/fmha_fprop_fp16_128_64_kernel.sm80.cu", "apex/contrib/csrc/fmha/src/fmha_fprop_fp16_256_64_kernel.sm80.cu", "apex/contrib/csrc/fmha/src/fmha_fprop_fp16_384_64_kernel.sm80.cu", "apex/contrib/csrc/fmha/src/fmha_fprop_fp16_512_64_kernel.sm80.cu", "apex/contrib/csrc/fmha/src/fmha_dgrad_fp16_128_64_kernel.sm80.cu", "apex/contrib/csrc/fmha/src/fmha_dgrad_fp16_256_64_kernel.sm80.cu", "apex/contrib/csrc/fmha/src/fmha_dgrad_fp16_384_64_kernel.sm80.cu", "apex/contrib/csrc/fmha/src/fmha_dgrad_fp16_512_64_kernel.sm80.cu", ], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros + generator_flag, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "--expt-relaxed-constexpr", "--expt-extended-lambda", "--use_fast_math", ] + version_dependent_macros + generator_flag + cc_flag, }, include_dirs=[ os.path.join(this_dir, "apex/contrib/csrc"), os.path.join(this_dir, "apex/contrib/csrc/fmha/src"), ], ) ) if "--fast_multihead_attn" in sys.argv: sys.argv.remove("--fast_multihead_attn") raise_if_cuda_home_none("--fast_multihead_attn") cc_flag = [] cc_flag.append("-gencode") cc_flag.append("arch=compute_70,code=sm_70") if bare_metal_version >= Version("11.0"): cc_flag.append("-gencode") cc_flag.append("arch=compute_80,code=sm_80") if bare_metal_version >= Version("11.1"): cc_flag.append("-gencode") cc_flag.append("arch=compute_86,code=sm_86") if bare_metal_version >= Version("11.8"): cc_flag.append("-gencode") cc_flag.append("arch=compute_90,code=sm_90") subprocess.run(["git", "submodule", "update", "--init", "apex/contrib/csrc/multihead_attn/cutlass"]) ext_modules.append( CUDAExtension( name="fast_multihead_attn", sources=[ "apex/contrib/csrc/multihead_attn/multihead_attn_frontend.cpp", "apex/contrib/csrc/multihead_attn/additive_masked_softmax_dropout_cuda.cu", "apex/contrib/csrc/multihead_attn/masked_softmax_dropout_cuda.cu", "apex/contrib/csrc/multihead_attn/encdec_multihead_attn_cuda.cu", "apex/contrib/csrc/multihead_attn/encdec_multihead_attn_norm_add_cuda.cu", "apex/contrib/csrc/multihead_attn/self_multihead_attn_cuda.cu", "apex/contrib/csrc/multihead_attn/self_multihead_attn_bias_additive_mask_cuda.cu", "apex/contrib/csrc/multihead_attn/self_multihead_attn_bias_cuda.cu", "apex/contrib/csrc/multihead_attn/self_multihead_attn_norm_add_cuda.cu", ], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros + generator_flag, "nvcc": [ "-O3", "-U__CUDA_NO_HALF_OPERATORS__", "-U__CUDA_NO_HALF_CONVERSIONS__", "--expt-relaxed-constexpr", "--expt-extended-lambda", "--use_fast_math", ] + version_dependent_macros + generator_flag + cc_flag, }, include_dirs=[ os.path.join(this_dir, "apex/contrib/csrc/multihead_attn/cutlass/include/"), os.path.join(this_dir, "apex/contrib/csrc/multihead_attn/cutlass/tools/util/include") ], ) ) if "--transducer" in sys.argv: sys.argv.remove("--transducer") raise_if_cuda_home_none("--transducer") ext_modules.append( CUDAExtension( name="transducer_joint_cuda", sources=[ "apex/contrib/csrc/transducer/transducer_joint.cpp", "apex/contrib/csrc/transducer/transducer_joint_kernel.cu", ], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros + generator_flag, "nvcc": ["-O3"] + version_dependent_macros + generator_flag, }, include_dirs=[os.path.join(this_dir, "csrc"), os.path.join(this_dir, "apex/contrib/csrc/multihead_attn")], ) ) ext_modules.append( CUDAExtension( name="transducer_loss_cuda", sources=[ "apex/contrib/csrc/transducer/transducer_loss.cpp", "apex/contrib/csrc/transducer/transducer_loss_kernel.cu", ], include_dirs=[os.path.join(this_dir, "csrc")], extra_compile_args={ "cxx": ["-O3"] + version_dependent_macros, "nvcc": ["-O3"] + version_dependent_macros, }, ) ) if "--cudnn_gbn" in sys.argv: sys.argv.remove("--cudnn_gbn") raise_if_cuda_home_none("--cudnn_gbn") if check_cudnn_version_and_warn("--cudnn_gbn", 8500): subprocess.run(["git", "submodule", "update", "--init", "apex/contrib/csrc/cudnn-frontend/"]) ext_modules.append( CUDAExtension( name="cudnn_gbn_lib", sources=[ "apex/contrib/csrc/cudnn_gbn/norm_sample.cpp", "apex/contrib/csrc/cudnn_gbn/cudnn_gbn.cpp", ], include_dirs=[os.path.join(this_dir, "apex/contrib/csrc/cudnn-frontend/include")], extra_compile_args={"cxx": ["-O3", "-g"] + version_dependent_macros + generator_flag}, ) ) if "--peer_memory" in sys.argv: sys.argv.remove("--peer_memory") raise_if_cuda_home_none("--peer_memory") ext_modules.append( CUDAExtension( name="peer_memory_cuda", sources=[ "apex/contrib/csrc/peer_memory/peer_memory_cuda.cu", "apex/contrib/csrc/peer_memory/peer_memory.cpp", ], extra_compile_args={"cxx": ["-O3"] + version_dependent_macros + generator_flag}, ) ) # NOTE: Requires NCCL >= 2.10.3 if "--nccl_p2p" in sys.argv: sys.argv.remove("--nccl_p2p") raise_if_cuda_home_none("--nccl_p2p") # Check NCCL version. _nccl_version_getter = load( name="_nccl_version_getter", sources=["apex/contrib/csrc/nccl_p2p/nccl_version.cpp", "apex/contrib/csrc/nccl_p2p/nccl_version_check.cu"], ) _available_nccl_version = _nccl_version_getter.get_nccl_version() if _available_nccl_version >= (2, 10): ext_modules.append( CUDAExtension( name="nccl_p2p_cuda", sources=[ "apex/contrib/csrc/nccl_p2p/nccl_p2p_cuda.cu", "apex/contrib/csrc/nccl_p2p/nccl_p2p.cpp", ], extra_compile_args={"cxx": ["-O3"] + version_dependent_macros + generator_flag}, ) ) else: warnings.warn( f"Skip `--nccl_p2p` as it requires NCCL 2.10.3 or later, but {_available_nccl_version[0]}.{_available_nccl_version[1]}" ) # note (mkozuki): Now `--fast_bottleneck` option (i.e. apex/contrib/bottleneck) depends on `--peer_memory` and `--nccl_p2p`. if "--fast_bottleneck" in sys.argv: sys.argv.remove("--fast_bottleneck") raise_if_cuda_home_none("--fast_bottleneck") if check_cudnn_version_and_warn("--fast_bottleneck", 8400): subprocess.run(["git", "submodule", "update", "--init", "apex/contrib/csrc/cudnn-frontend/"]) ext_modules.append( CUDAExtension( name="fast_bottleneck", sources=["apex/contrib/csrc/bottleneck/bottleneck.cpp"], include_dirs=[os.path.join(this_dir, "apex/contrib/csrc/cudnn-frontend/include")], extra_compile_args={"cxx": ["-O3"] + version_dependent_macros + generator_flag}, ) ) if "--fused_conv_bias_relu" in sys.argv: sys.argv.remove("--fused_conv_bias_relu") raise_if_cuda_home_none("--fused_conv_bias_relu") if check_cudnn_version_and_warn("--fused_conv_bias_relu", 8400): subprocess.run(["git", "submodule", "update", "--init", "apex/contrib/csrc/cudnn-frontend/"]) ext_modules.append( CUDAExtension( name="fused_conv_bias_relu", sources=["apex/contrib/csrc/conv_bias_relu/conv_bias_relu.cpp"], include_dirs=[os.path.join(this_dir, "apex/contrib/csrc/cudnn-frontend/include")], extra_compile_args={"cxx": ["-O3"] + version_dependent_macros + generator_flag}, ) ) setup( name="apex", version="0.1", packages=find_packages( exclude=("build", "csrc", "include", "tests", "dist", "docs", "tests", "examples", "apex.egg-info",) ), install_requires=["packaging>20.6"], description="PyTorch Extensions written by NVIDIA", ext_modules=ext_modules, cmdclass={"build_ext": BuildExtension} if ext_modules else {}, extras_require=extras, )
GeneSplice-main
GeneSplice/apex/setup.py
import logging import warnings # May help avoid undefined symbol errors https://pytorch.org/cppdocs/notes/faq.html#undefined-symbol-errors-from-pytorch-aten import torch __all__ = ["amp", "fp16_utils", "optimizers", "normalization", "transformer"] if torch.distributed.is_available(): from . import parallel __all__.append("parallel") from . import amp from . import fp16_utils # For optimizers and normalization there is no Python fallback. # Absence of cuda backend is a hard error. # I would like the errors from importing fused_adam_cuda or fused_layer_norm_cuda # to be triggered lazily, because if someone has installed with --cpp_ext and --cuda_ext # so they expect those backends to be available, but for some reason they actually aren't # available (for example because they built improperly in a way that isn't revealed until # load time) the error message is timely and visible. from . import optimizers from . import normalization from . import transformer # Logging utilities for apex.transformer module class RankInfoFormatter(logging.Formatter): def format(self, record): from apex.transformer.parallel_state import get_rank_info record.rank_info = get_rank_info() return super().format(record) _library_root_logger = logging.getLogger(__name__) handler = logging.StreamHandler() handler.setFormatter(RankInfoFormatter("%(asctime)s - PID:%(process)d - rank:%(rank_info)s - %(filename)s:%(lineno)d - %(levelname)s - %(message)s", "%y-%m-%d %H:%M:%S")) _library_root_logger.addHandler(handler) _library_root_logger.propagate = False def check_cudnn_version_and_warn(global_option: str, required_cudnn_version: int) -> bool: cudnn_available = torch.backends.cudnn.is_available() cudnn_version = torch.backends.cudnn.version() if cudnn_available else None if not (cudnn_available and (cudnn_version >= required_cudnn_version)): warnings.warn( f"`{global_option}` depends on cuDNN {required_cudnn_version} or later, " f"but {'cuDNN is not available' if not cudnn_available else cudnn_version}" ) return False return True class DeprecatedFeatureWarning(FutureWarning): pass def deprecated_warning(msg: str) -> None: if ( not torch.distributed.is_available or not torch.distributed.is_initialized() or (torch.distributed.is_initialized() and torch.distributed.get_rank() == 0) ): warnings.warn(msg, DeprecatedFeatureWarning)
GeneSplice-main
GeneSplice/apex/apex/__init__.py
from typing import Optional, Sequence import torch __all__ = ["_cast_if_autocast_enabled"] def _get_autocast_dtypes() -> Sequence[torch.dtype]: if torch.cuda.is_bf16_supported(): return [torch.half, torch.bfloat16] return [torch.half] def _get_current_dtype(dtype: Optional[torch.dtype] = None) -> torch.dtype: if not torch.is_autocast_enabled(): return torch.float or dtype else: return torch.get_autocast_gpu_dtype() def _cast_if_autocast_enabled(*args): if not torch.is_autocast_enabled(): return args else: return torch.cuda.amp.autocast_mode._cast(args, torch.get_autocast_gpu_dtype())
GeneSplice-main
GeneSplice/apex/apex/_autocast_utils.py
import torch from torch.nn.modules.batchnorm import _BatchNorm from torch.nn import functional as F import syncbn from .optimized_sync_batchnorm_kernel import SyncBatchnormFunction class SyncBatchNorm(_BatchNorm): """ synchronized batch normalization module extented from `torch.nn.BatchNormNd` with the added stats reduction across multiple processes. :class:`apex.parallel.SyncBatchNorm` is designed to work with `DistributedDataParallel`. When running in training mode, the layer reduces stats across all processes to increase the effective batchsize for normalization layer. This is useful in applications where batch size is small on a given process that would diminish converged accuracy of the model. The model uses collective communication package from `torch.distributed`. When running in evaluation mode, the layer falls back to `torch.nn.functional.batch_norm` Args: num_features: :math:`C` from an expected input of size :math:`(N, C, L)` or :math:`L` from input of size :math:`(N, L)` eps: a value added to the denominator for numerical stability. Default: 1e-5 momentum: the value used for the running_mean and running_var computation. Can be set to ``None`` for cumulative moving average (i.e. simple average). Default: 0.1 affine: a boolean value that when set to ``True``, this module has learnable affine parameters. Default: ``True`` track_running_stats: a boolean value that when set to ``True``, this module tracks the running mean and variance, and when set to ``False``, this module does not track such statistics and always uses batch statistics in both training and eval modes. Default: ``True`` process_group: pass in a process group within which the stats of the mini-batch is being synchronized. ``None`` for using default process group channel_last: a boolean value that when set to ``True``, this module take the last dimension of the input tensor to be the channel dimension. Default: False Examples:: >>> # channel first tensor >>> sbn = apex.parallel.SyncBatchNorm(100).cuda() >>> inp = torch.randn(10, 100, 14, 14).cuda() >>> out = sbn(inp) >>> inp = torch.randn(3, 100, 20).cuda() >>> out = sbn(inp) >>> # channel last tensor >>> sbn = apex.parallel.SyncBatchNorm(100, channel_last=True).cuda() >>> inp = torch.randn(10, 14, 14, 100).cuda() """ def __init__(self, num_features, eps=1e-5, momentum=0.1, affine=True, track_running_stats=True, process_group=None, channel_last=False, fuse_relu=False): super(SyncBatchNorm, self).__init__(num_features, eps=eps, momentum=momentum, affine=affine, track_running_stats=track_running_stats) self.process_group = process_group self.channel_last = channel_last self.fuse_relu = fuse_relu def _specify_process_group(self, process_group): self.process_group = process_group def _specify_channel_last(self, channel_last): self.channel_last = channel_last def forward(self, input, z = None): # if input.dim() == 2, we switch to channel_last for efficient memory accessing channel_last = self.channel_last if input.dim() != 2 else True if not self.training and self.track_running_stats and not channel_last and not self.fuse_relu and z == None: # fall back to pytorch implementation for inference return F.batch_norm(input, self.running_mean, self.running_var, self.weight, self.bias, False, 0.0, self.eps) else: exponential_average_factor = 0.0 if self.training and self.track_running_stats: self.num_batches_tracked += 1 if self.momentum is None: exponential_average_factor = 1.0 / float(self.num_batches_tracked) else: exponential_average_factor = self.momentum return SyncBatchnormFunction.apply(input, z, self.weight, self.bias, self.running_mean, self.running_var, self.eps, self.training or not self.track_running_stats, exponential_average_factor, self.process_group, channel_last, self.fuse_relu)
GeneSplice-main
GeneSplice/apex/apex/parallel/optimized_sync_batchnorm.py
import torch from torch.autograd.function import Function from apex.parallel import ReduceOp class SyncBatchnormFunction(Function): @staticmethod def forward(ctx, input, weight, bias, running_mean, running_variance, eps, process_group, world_size): torch.cuda.nvtx.range_push("sync_BN_fw") # transpose it to channel last to support broadcasting for input with different rank c_last_input = input.transpose(1, -1).contiguous().clone() ctx.save_for_backward(c_last_input, weight, bias, running_mean, running_variance) ctx.eps = eps ctx.process_group = process_group ctx.world_size = world_size c_last_input = (c_last_input - running_mean) / \ torch.sqrt(running_variance + eps) if weight is not None: c_last_input = c_last_input * weight if bias is not None: c_last_input = c_last_input + bias torch.cuda.nvtx.range_pop() return c_last_input.transpose(1, -1).contiguous().clone() @staticmethod def backward(ctx, grad_output): torch.cuda.nvtx.range_push("sync_BN_bw") # mini batch mean & var are calculated by forward path. # mu = 1./N*np.sum(h, axis = 0) # var = 1./N*np.sum((h-mu)**2, axis = 0) c_last_input, weight, bias, running_mean, running_variance = ctx.saved_tensors eps = ctx.eps process_group = ctx.process_group world_size = ctx.world_size grad_input = grad_weight = grad_bias = None num_features = running_mean.size()[0] # transpose it to channel last to support broadcasting for input with different rank torch.cuda.nvtx.range_push("carilli field") c_last_grad = grad_output.transpose(1, -1).contiguous() # squash non-channel dimension so we can easily calculate mean c_grad = c_last_grad.view(-1, num_features).contiguous() torch.cuda.nvtx.range_pop() # calculate grad_input if ctx.needs_input_grad[0]: # dh = gamma * (var + eps)**(-1. / 2.) * (dy - np.mean(dy, axis=0) # - (h - mu) * (var + eps)**(-1.0) * np.mean(dy * (h - mu), axis=0)) mean_dy = c_grad.mean(0) mean_dy_xmu = (c_last_grad * (c_last_input - running_mean)).view(-1, num_features).mean(0) if torch.distributed.is_initialized(): torch.distributed.all_reduce( mean_dy, ReduceOp.SUM, process_group) mean_dy = mean_dy / world_size torch.distributed.all_reduce( mean_dy_xmu, ReduceOp.SUM, process_group) mean_dy_xmu = mean_dy_xmu / world_size c_last_grad_input = (c_last_grad - mean_dy - (c_last_input - running_mean) / ( running_variance + eps) * mean_dy_xmu) / torch.sqrt(running_variance + eps) if weight is not None: c_last_grad_input.mul_(weight) grad_input = c_last_grad_input.transpose(1, -1).contiguous() # calculate grad_weight grad_weight = None if weight is not None and ctx.needs_input_grad[1]: # dgamma = np.sum((h - mu) * (var + eps)**(-1. / 2.) * dy, axis=0) grad_weight = ((c_last_input - running_mean) / torch.sqrt( running_variance + eps) * c_last_grad).view(-1, num_features).sum(0) # calculate grad_bias grad_bias = None if bias is not None and ctx.needs_input_grad[2]: # dbeta = np.sum(dy, axis=0) grad_bias = c_grad.sum(0) torch.cuda.nvtx.range_pop() return grad_input, grad_weight, grad_bias, None, None, None, None, None
GeneSplice-main
GeneSplice/apex/apex/parallel/sync_batchnorm_kernel.py
import torch if hasattr(torch.distributed, 'ReduceOp'): ReduceOp = torch.distributed.ReduceOp elif hasattr(torch.distributed, 'reduce_op'): ReduceOp = torch.distributed.reduce_op else: ReduceOp = torch.distributed.deprecated.reduce_op from .distributed import DistributedDataParallel, Reducer # This is tricky because I'd like SyncBatchNorm to be exposed the same way # for both the cuda-enabled and python-fallback versions, and I don't want # to suppress the error information. try: import syncbn from .optimized_sync_batchnorm import SyncBatchNorm except ImportError as err: from .sync_batchnorm import SyncBatchNorm SyncBatchNorm.syncbn_import_error = err def convert_syncbn_model(module, process_group=None, channel_last=False): ''' Recursively traverse module and its children to replace all instances of ``torch.nn.modules.batchnorm._BatchNorm`` with :class:`apex.parallel.SyncBatchNorm`. All ``torch.nn.BatchNorm*N*d`` wrap around ``torch.nn.modules.batchnorm._BatchNorm``, so this function lets you easily switch to use sync BN. Args: module (torch.nn.Module): input module Example:: >>> # model is an instance of torch.nn.Module >>> import apex >>> sync_bn_model = apex.parallel.convert_syncbn_model(model) ''' from apex import deprecated_warning deprecated_warning("apex.parallel.convert_syncbn_model is deprecated and will be removed by the end of February 2023. Use `torch.nn.SyncBatchNorm.convert_sync_batchnorm`.") mod = module if isinstance(module, torch.nn.modules.instancenorm._InstanceNorm): return module if isinstance(module, torch.nn.modules.batchnorm._BatchNorm): mod = SyncBatchNorm(module.num_features, module.eps, module.momentum, module.affine, module.track_running_stats, process_group, channel_last=channel_last) mod.running_mean = module.running_mean mod.running_var = module.running_var mod.num_batches_tracked = module.num_batches_tracked if module.affine: mod.weight.data = module.weight.data.clone().detach() mod.bias.data = module.bias.data.clone().detach() for name, child in module.named_children(): mod.add_module(name, convert_syncbn_model(child, process_group=process_group, channel_last=channel_last)) # TODO(jie) should I delete model explicitly? del module return mod def create_syncbn_process_group(group_size): ''' Creates process groups to be used for syncbn of a give ``group_size`` and returns process group that current GPU participates in. ``group_size`` must divide the total number of GPUs (world_size). ``group_size`` of 0 would be considered as =world_size. In this case ``None`` will be returned. ``group_size`` of 1 would be equivalent to using non-sync bn, but will still carry the overhead. Args: group_size (int): number of GPU's to collaborate for sync bn Example:: >>> # model is an instance of torch.nn.Module >>> import apex >>> group = apex.parallel.create_syncbn_process_group(group_size) ''' if group_size==0: return None world_size = torch.distributed.get_world_size() assert(world_size >= group_size) assert(world_size % group_size == 0) group=None for group_num in (range(world_size//group_size)): group_ids = range(group_num*group_size, (group_num+1)*group_size) cur_group = torch.distributed.new_group(ranks=group_ids) if (torch.distributed.get_rank()//group_size == group_num): group = cur_group #can not drop out and return here, every process must go through creation of all subgroups assert(group is not None) return group
GeneSplice-main
GeneSplice/apex/apex/parallel/__init__.py
import torch from torch.nn.modules.batchnorm import _BatchNorm from torch.nn import functional as F from .sync_batchnorm_kernel import SyncBatchnormFunction from apex.parallel import ReduceOp class SyncBatchNorm(_BatchNorm): """ synchronized batch normalization module extented from ``torch.nn.BatchNormNd`` with the added stats reduction across multiple processes. :class:`apex.parallel.SyncBatchNorm` is designed to work with ``DistributedDataParallel``. When running in training mode, the layer reduces stats across all processes to increase the effective batchsize for normalization layer. This is useful in applications where batch size is small on a given process that would diminish converged accuracy of the model. The model uses collective communication package from ``torch.distributed``. When running in evaluation mode, the layer falls back to ``torch.nn.functional.batch_norm``. Args: num_features: :math:`C` from an expected input of size :math:`(N, C, L)` or :math:`L` from input of size :math:`(N, L)` eps: a value added to the denominator for numerical stability. Default: 1e-5 momentum: the value used for the running_mean and running_var computation. Can be set to ``None`` for cumulative moving average (i.e. simple average). Default: 0.1 affine: a boolean value that when set to ``True``, this module has learnable affine parameters. Default: ``True`` track_running_stats: a boolean value that when set to ``True``, this module tracks the running mean and variance, and when set to ``False``, this module does not track such statistics and always uses batch statistics in both training and eval modes. Default: ``True`` Example:: >>> sbn = apex.parallel.SyncBatchNorm(100).cuda() >>> inp = torch.randn(10, 100, 14, 14).cuda() >>> out = sbn(inp) >>> inp = torch.randn(3, 100, 20).cuda() >>> out = sbn(inp) """ warned = False def __init__(self, num_features, eps=1e-5, momentum=0.1, affine=True, track_running_stats=True, process_group=None, channel_last=False): from apex import deprecated_warning deprecated_warning("apex.parallel.SyncBatchNorm is deprecated and will be removed by the end of February 2023. Use `torch.nn.SyncBatchNorm`.") if channel_last == True: raise AttributeError("channel_last is not supported by primitive SyncBatchNorm implementation. Try install apex with `--cuda_ext` if channel_last is desired.") if not SyncBatchNorm.warned: if hasattr(self, "syncbn_import_error"): print("Warning: using Python fallback for SyncBatchNorm, possibly because apex was installed without --cuda_ext. The exception raised when attempting to import the cuda backend was: ", self.syncbn_import_error) else: print("Warning: using Python fallback for SyncBatchNorm") SyncBatchNorm.warned = True super(SyncBatchNorm, self).__init__(num_features, eps=eps, momentum=momentum, affine=affine, track_running_stats=track_running_stats) self.process_group = process_group def _specify_process_group(self, process_group): self.process_group = process_group def forward(self, input): torch.cuda.nvtx.range_push("sync_bn_fw_with_mean_var") mean = None var = None cast = None out = None # casting to handle mismatch input type to layer type if self.running_mean is not None: if self.running_mean.dtype != input.dtype: input = input.to(self.running_mean.dtype) cast = input.dtype elif self.weight is not None: if self.weight.dtype != input.dtype: input = input.to(self.weight.dtype) cast = input.dtype if not self.training and self.track_running_stats: # fall back to pytorch implementation for inference torch.cuda.nvtx.range_pop() out = F.batch_norm(input, self.running_mean, self.running_var, self.weight, self.bias, False, 0.0, self.eps) else: process_group = self.process_group world_size = 1 if not self.process_group: process_group = torch.distributed.group.WORLD self.num_batches_tracked += 1 with torch.no_grad(): channel_first_input = input.transpose(0, 1).contiguous() squashed_input_tensor_view = channel_first_input.view( channel_first_input.size(0), -1) # total number of data points for each variance entry. Used to calculate unbiased variance estimate m = None local_m = float(squashed_input_tensor_view.size()[1]) local_mean = torch.mean(squashed_input_tensor_view, 1) local_sqr_mean = torch.pow( squashed_input_tensor_view, 2).mean(1) if torch.distributed.is_initialized(): world_size = torch.distributed.get_world_size(process_group) torch.distributed.all_reduce( local_mean, ReduceOp.SUM, process_group) mean = local_mean / world_size torch.distributed.all_reduce( local_sqr_mean, ReduceOp.SUM, process_group) sqr_mean = local_sqr_mean / world_size m = local_m * world_size else: m = local_m mean = local_mean sqr_mean = local_sqr_mean # var(x) = E (( x - mean_x ) ** 2) # = 1 / N * sum ( x - mean_x ) ** 2 # = 1 / N * sum (x**2) - mean_x**2 var = sqr_mean - mean.pow(2) if self.running_mean is not None: self.running_mean = self.momentum * mean + \ (1 - self.momentum) * self.running_mean if self.running_var is not None: # as noted by the paper, we used unbiased variance estimate of the mini-batch # Var[x] = m / (m-1) * Eb (sample_variance) self.running_var = m / \ (m-1) * self.momentum * var + \ (1 - self.momentum) * self.running_var torch.cuda.nvtx.range_pop() out = SyncBatchnormFunction.apply(input, self.weight, self.bias, mean, var, self.eps, process_group, world_size) return out.to(cast)
GeneSplice-main
GeneSplice/apex/apex/parallel/sync_batchnorm.py
from collections import OrderedDict import copy import importlib from itertools import chain import torch import torch.distributed as dist from torch.nn.modules import Module from torch.autograd import Variable from ..multi_tensor_apply import multi_tensor_applier imported_flatten_impl = False def import_flatten_impl(): global flatten_impl, unflatten_impl, imported_flatten_impl try: import apex_C flatten_impl = apex_C.flatten unflatten_impl = apex_C.unflatten except ImportError: print("Warning: apex was installed without --cpp_ext. Falling back to Python flatten and unflatten.") flatten_impl = torch._utils._flatten_dense_tensors unflatten_impl = torch._utils._unflatten_dense_tensors imported_flatten_impl = True def flatten(bucket): if not imported_flatten_impl: import_flatten_impl() return flatten_impl(bucket) def unflatten(coalesced, bucket): if not imported_flatten_impl: import_flatten_impl() return unflatten_impl(coalesced, bucket) # apply_dist_call requires that tensors in 'bucket' are all the same type. def apply_flat_dist_call(bucket, call, extra_args=None): coalesced = flatten(bucket) if extra_args is not None: call(coalesced, *extra_args) else: call(coalesced) if call is dist.all_reduce: coalesced /= dist.get_world_size() for buf, synced in zip(bucket, unflatten(coalesced, bucket)): buf.copy_(synced) def split_half_float_double(tensors): dtypes = ["torch.cuda.HalfTensor", "torch.cuda.FloatTensor", "torch.cuda.DoubleTensor"] buckets = [] for i, dtype in enumerate(dtypes): bucket = [t for t in tensors if t.type() == dtype] if bucket: buckets.append(bucket) return buckets def split_by_type(tensors): buckets = OrderedDict() for tensor in tensors: tp = tensor.type() if tp not in buckets: buckets[tp] = [] buckets[tp].append(tensor) return buckets # flat_dist_call organizes 'tensors' by type. def flat_dist_call(tensors, call, extra_args=None): buckets = split_by_type(tensors) for tp in buckets: bucket = buckets[tp] apply_flat_dist_call(bucket, call, extra_args) def extract_tensors(maybe_tensor, tensor_list): if torch.is_tensor(maybe_tensor): tensor_list.append(maybe_tensor) else: try: for item in maybe_tensor: extract_tensors(item, tensor_list) except TypeError: return class Reducer(object): """ :class:`apex.parallel.Reducer` is a simple class that helps allreduce a module's parameters across processes. :class:`Reducer` is intended to give the user additional control: Unlike :class:`DistributedDataParallel`, :class:`Reducer` will not automatically allreduce parameters during ``backward()``. Instead, :class:`Reducer` waits for the user to call ``<reducer_instance>.reduce()`` manually. This enables, for example, delaying the allreduce to be carried out every several iterations instead of every single iteration. Like :class:`DistributedDataParallel`, :class:`Reducer` averages any tensors it allreduces over the number of participating processes. :class:`Reducer` is designed to work with the upstream launch utility script ``torch.distributed.launch`` with ``--nproc_per_node <= number of gpus per node``. When used with this launcher, :class:`Reducer` assumes 1:1 mapping of processes to GPUs. It also assumes that your script calls ``torch.cuda.set_device(args.rank)`` before creating the model. Args: module_or_grads_list: Either a network definition (module) being run in multi-gpu/distributed mode, or an iterable of gradients to be reduced. If a module is passed in, the Reducer constructor will sync the parameters across processes (broadcasting from rank 0) to make sure they're all initialized with the same values. If a list of gradients (that came from some module) is passed in, the user is responsible for manually syncing that module's parameters at the beginning of training. """ def __init__(self, module_or_grads_list): if isinstance(module_or_grads_list, Module): self.module = module_or_grads_list flat_dist_call([param.data for param in self.module.parameters()], dist.broadcast, (0,) ) else: self.module = None self.grads = [] extract_tensors(module_or_grads_list, self.grads) def reduce(self): if self.module: grads = [param.grad.data for param in self.module.parameters() if param.grad is not None] flat_dist_call(grads, dist.all_reduce) else: flat_dist_call(self.grads, dist.all_reduce) class DistributedDataParallel(Module): """ :class:`apex.parallel.DistributedDataParallel` is a module wrapper that enables easy multiprocess distributed data parallel training, similar to ``torch.nn.parallel.DistributedDataParallel``. Parameters are broadcast across participating processes on initialization, and gradients are allreduced and averaged over processes during ``backward()``. :class:`DistributedDataParallel` is optimized for use with NCCL. It achieves high performance by overlapping communication with computation during ``backward()`` and bucketing smaller gradient transfers to reduce the total number of transfers required. :class:`DistributedDataParallel` is designed to work with the upstream launch utility script ``torch.distributed.launch`` with ``--nproc_per_node <= number of gpus per node``. When used with this launcher, :class:`DistributedDataParallel` assumes 1:1 mapping of processes to GPUs. It also assumes that your script calls ``torch.cuda.set_device(args.rank)`` before creating the model. https://github.com/NVIDIA/apex/tree/master/examples/simple/distributed shows detailed usage. https://github.com/NVIDIA/apex/tree/master/examples/imagenet shows another example that combines :class:`DistributedDataParallel` with mixed precision training. Args: module: Network definition to be run in multi-gpu/distributed mode. message_size (int, default=1e7): Minimum number of elements in a communication bucket. delay_allreduce (bool, default=False): Delay all communication to the end of the backward pass. This disables overlapping communication with computation. allreduce_trigger_params (list, optional, default=None): If supplied, should contain a list of parameters drawn from the model. Allreduces will be kicked off whenever one of these parameters receives its gradient (as opposed to when a bucket of size message_size is full). At the end of backward(), a cleanup allreduce to catch any remaining gradients will also be performed automatically. If allreduce_trigger_params is supplied, the message_size argument will be ignored. allreduce_always_fp32 (bool, default=False): Convert any FP16 gradients to FP32 before allreducing. This can improve stability for widely scaled-out runs. gradient_average (bool, default=True): Option to toggle whether or not DDP averages the allreduced gradients over processes. For proper scaling, the default value of True is recommended. gradient_predivide_factor (float, default=1.0): Allows perfoming the average of gradients over processes partially before and partially after the allreduce. Before allreduce: ``grads.mul_(1.0/gradient_predivide_factor)``. After allreduce: ``grads.mul_(gradient_predivide_factor/world size)``. This can reduce the stress on the dynamic range of FP16 allreduces for widely scaled-out runs. .. warning:: If ``gradient_average=False``, the pre-allreduce division (``grads.mul_(1.0/gradient_predivide_factor)``) will still be applied, but the post-allreduce gradient averaging (``grads.mul_(gradient_predivide_factor/world size)``) will be omitted. """ def __init__(self, module, message_size=10000000, delay_allreduce=False, shared_param=None, allreduce_trigger_params=None, retain_allreduce_buffers=False, allreduce_always_fp32=False, num_allreduce_streams=1, allreduce_communicators=None, gradient_average=True, gradient_predivide_factor=1.0, gradient_average_split_factor=None, prof=False): super(DistributedDataParallel, self).__init__() from apex import deprecated_warning deprecated_warning("apex.parallel.DistributedDataParallel is deprecated and will be removed by the end of February 2023.") # Backward/forward compatibility around # https://github.com/pytorch/pytorch/commit/540ef9b1fc5506369a48491af8a285a686689b36 and # https://github.com/pytorch/pytorch/commit/044d00516ccd6572c0d6ab6d54587155b02a3b86 if hasattr(dist, "get_backend"): self._backend = dist.get_backend() if hasattr(dist, "DistBackend"): self.backend_enum_holder = dist.DistBackend else: self.backend_enum_holder = dist.Backend else: self._backend = dist._backend self.backend_enum_holder = dist.dist_backend self.warn_on_half = True if self._backend == self.backend_enum_holder.GLOO else False self.prof = prof self.allreduce_different_streams = (num_allreduce_streams > 1) self.num_allreduce_streams = num_allreduce_streams self.allreduce_communicators = allreduce_communicators if self.allreduce_communicators: assert len(allreduce_communicators[0]) == num_allreduce_streams assert len(allreduce_communicators[0]) == len(allreduce_communicators[1]) assert self.allreduce_different_streams if self.allreduce_different_streams and delay_allreduce: raise ValueError("self.allreduce_different_streams may only be used if delay_allreduce=False.") if shared_param is not None: raise ValueError("shared_param is no longer supported as an option. It was misleadingly named from the start. It turns out overlapping communication with computation should work fine with shared parameters. If you still wish to delay communication to the end of the backward pass, use delay_allreduce=True|False instead.") self.world_size = float(dist.get_world_size()) self.retain_allreduce_buffers = retain_allreduce_buffers self.allreduce_always_fp32 = allreduce_always_fp32 self.gradient_average = gradient_average self.gradient_predivide_factor = gradient_predivide_factor self.custom_allreduce_triggers = False if allreduce_trigger_params is not None: if delay_allreduce: raise ValueError("Setting allreduce_trigger_params is only valid if delay_allreduce=False.") self.custom_allreduce_triggers = True self.allreduce_trigger_params = set([id(param) for param in allreduce_trigger_params]) self.delay_allreduce = delay_allreduce self.message_size = message_size self.main_stream = torch.cuda.current_stream() self.bucket_streams = [] self.bucket_events = [] self.module = module self._disable_allreduce = False if self._backend == self.backend_enum_holder.NCCL: for param in self.module.parameters(): assert param.is_cuda, "NCCL backend only supports model parameters to be on GPU." self.active_params = [] self.param_type_to_tmp_i = {"torch.cuda.HalfTensor" : 0, "torch.cuda.FloatTensor" : 1, "torch.cuda.DoubleTensor" : 2} if multi_tensor_applier.available: # TODO: I really need to centralize the C++ backed imports import amp_C self.multi_tensor_scale = amp_C.multi_tensor_scale self._overflow_buf = torch.cuda.IntTensor([0]) self.create_hooks() flat_dist_call([param.data for param in self.module.parameters()], dist.broadcast, (0,) ) def __setstate__(self, state): super(DistributedDataParallel, self).__setstate__(state) if self.allreduce_different_streams and delay_allreduce: raise ValueError("self.allreduce_different_streams may only be used if delay_allreduce=False.") if self.delay_allreduce: self.needs_refresh = True self.bucket_streams = [] self.bucket_events = [] def __getstate__(self): attrs = copy.copy(self.__dict__) if self._backend != self.backend_enum_holder.NCCL: del attrs['self.bucket_streams'] del attrs['self.bucket_events'] return attrs def enable_allreduce(self): self._disable_allreduce = False def disable_allreduce(self): self._disable_allreduce = True # Broadcast rank 0's bucket structure across all processes, and have all processes # regenerate their bucket structures to match. def sync_bucket_structure(self): # Append leftover buckets for tmp_bucket in self.tmp_buckets: if len(tmp_bucket) > 0: self.active_i_buckets.append(tmp_bucket) self.num_buckets = len(self.active_i_buckets) self.bucket_sizes = [len(bucket) for bucket in self.active_i_buckets] info_tensor = torch.cuda.IntTensor([self.num_buckets] + self.bucket_sizes + list(chain(*self.active_i_buckets))) dist.broadcast(info_tensor, 0) info = [int(entry) for entry in info_tensor] self.num_buckets = info[0] self.bucket_sizes = info[1:self.num_buckets + 1] self.buckets = [[None for _ in range(self.bucket_sizes[i])] for i in range(self.num_buckets)] # Technically, active_i_buckets' work is done. But the information is still useful to # keep around. Therefore, refresh active_i_buckets based on rank 0 as well. self.active_i_buckets = [[None for _ in range(self.bucket_sizes[i])] for i in range(self.num_buckets)] flattened_buckets = info[self.num_buckets + 1:] flat_i = 0 for bucket_idx in range(self.num_buckets): for bucket_loc in range(self.bucket_sizes[bucket_idx]): param_i = flattened_buckets[flat_i] self.active_i_buckets[bucket_idx][bucket_loc] = param_i self.param_id_to_bucket[id(self.active_params[param_i])] = (bucket_idx, bucket_loc) flat_i += 1 def create_hooks(self): # Fallback hook that's only called at the end of backward. # Used if you deliberately want to delay allreduces to the end, or to refresh the # bucket structure that will be used to overlap communication with computation in later # iterations. def allreduce_params(): # Bucket record refresh if not self.delay_allreduce: if self.needs_refresh: self.sync_bucket_structure() self.needs_refresh = False self.allreduce_fallback() def overlapping_backward_epilogue(): for stream, event in zip(self.bucket_streams, self.bucket_events): stream.record_event(event) torch.cuda.current_stream().wait_event(event) # Sanity checks that all the buckets were kicked off if self.next_bucket != self.num_buckets: raise RuntimeError("In epilogue, next_bucket ({}) != num_buckets ({}). ".format( self.next_bucket, self.num_buckets), "This probably indicates some buckets were not allreduced.") for actual, expected in zip(self.buckets_ready_size, self.bucket_sizes): if actual != expected: raise RuntimeError("Some param buckets were not allreduced.") self.grad_accs = [] for param in self.module.parameters(): if param.requires_grad: def wrapper(param): param_tmp = param.expand_as(param) grad_acc = param_tmp.grad_fn.next_functions[0][0] def allreduce_hook(*unused): if self.prof: torch.cuda.nvtx.range_push("allreduce_hook") if not self._disable_allreduce: if self.delay_allreduce or self.needs_refresh: # TODO: How do we want to handle multiple backward passes between # each forward, e.g., backward passes with retain_graph=True? # needs_refresh and callback_queued are both vulnerable states. if not self.delay_allreduce and self.needs_refresh: # Use the backward pass to build the bucket structure on the fly. active_i = self.param_id_to_active_i[id(param)] # Float, half, and double tensors are grouped into buckets separately. current_type = self.param_type_to_tmp_i[param.type()] self.tmp_buckets[current_type].append(active_i) ship_tmp_bucket = False if self.custom_allreduce_triggers: if id(param) in self.allreduce_trigger_params: ship_tmp_bucket = True else: self.tmp_numels[current_type] += param.numel() if self.tmp_numels[current_type] >= self.message_size: ship_tmp_bucket = True # To consider: If custom_allreduce_triggers are in use, ship all # tmp_buckets, not just tmp_buckets[current_type]. if ship_tmp_bucket: self.active_i_buckets.append(self.tmp_buckets[current_type]) self.tmp_buckets[current_type] = [] self.tmp_numels[current_type] = 0 if not self.callback_queued: Variable._execution_engine.queue_callback(allreduce_params) self.callback_queued = True else: if not self.callback_queued: Variable._execution_engine.queue_callback(overlapping_backward_epilogue) self.callback_queued = True self.comm_ready_buckets(param) if self.prof: torch.cuda.nvtx.range_pop() grad_acc.register_hook(allreduce_hook) self.grad_accs.append(grad_acc) wrapper(param) def _stream_this_bucket(self, bucket_idx): if self.allreduce_different_streams: return self.bucket_streams[bucket_idx%self.num_allreduce_streams] else: return self.bucket_streams[0] def _event_this_bucket(self, bucket_idx): if self.allreduce_different_streams: return self.bucket_events[bucket_idx%self.num_allreduce_streams] else: return self.bucket_events[0] def allreduce_bucket(self, bucket, bucket_idx, force_default_stream): tensor = flatten(bucket) if force_default_stream: bucket_stream = self.main_stream else: bucket_stream = self._stream_this_bucket(bucket_idx) bucket_event = self._event_this_bucket(bucket_idx) torch.cuda.current_stream().record_event(bucket_event) bucket_stream.wait_event(bucket_event) with torch.cuda.stream(bucket_stream): # self.main_stream.wait_stream(torch.cuda.current_stream()) # torch.cuda.synchronize() tensor_to_allreduce = tensor if self.allreduce_always_fp32: tensor_to_allreduce = tensor.float() if self.gradient_predivide_factor != 1.0: tensor_to_allreduce.mul_(1./self.gradient_predivide_factor) if self.allreduce_different_streams and not force_default_stream: dist.all_reduce(tensor_to_allreduce, group=self.bucket_pgs[bucket_idx%self.num_allreduce_streams]) else: dist.all_reduce(tensor_to_allreduce) if self.gradient_average: tensor_to_allreduce.mul_(self.gradient_predivide_factor/self.world_size) if self.allreduce_always_fp32 and tensor is not tensor_to_allreduce: tensor.copy_(tensor_to_allreduce) if not self.retain_allreduce_buffers: if multi_tensor_applier.available: multi_tensor_applier( self.multi_tensor_scale, self._overflow_buf, [unflatten(tensor, bucket), bucket], 1.0) else: for buf, synced in zip(bucket, unflatten(tensor, bucket)): buf.copy_(synced) # I think we actually do need this here. After allreduce_bucket returns, tensor will # eventually go out of scope and die, at which point it could otherwise be freed for # further reuse by the main stream while the allreduce/div/unflatten are underway in bucket_stream. tensor.record_stream(bucket_stream) return tensor def allreduce_maybe_retain(self, bucket, bucket_idx, force_default_stream=False): allreduced = self.allreduce_bucket(bucket, bucket_idx, force_default_stream) if self.retain_allreduce_buffers: if self.allreduce_buffers[bucket_idx] is not None: raise RuntimeError("The backward pass is attempting to replace an already-filled " "allreduce buffer. This is almost certainly an error.") self.allreduce_buffers[bucket_idx] = allreduced for view, grad in zip(unflatten(allreduced, bucket), bucket): grad.data = view # for buf, synced in zip(bucket, unflatten(allreduced, bucket)): # buf.copy_(synced) def allreduce_fallback(self): for stream, event in zip(self.bucket_streams, self.bucket_events): stream.record_event(event) torch.cuda.current_stream().wait_event(event) if self.retain_allreduce_buffers: grads = [param.grad for param in self.module.parameters() if param.grad is not None] else: grads = [param.grad.data for param in self.module.parameters() if param.grad is not None] split_buckets = split_half_float_double(grads) # If retain_allreduce_buffers is True and delay_allreduce is False, # this will only be done during the first backward pass, ignored by the # training script, and overwritten in the next forward pass. So it's harmless. if self.retain_allreduce_buffers: self.allreduce_buffers = [None for _ in range(len(split_buckets))] for i, bucket in enumerate(split_buckets): allreduced = self.allreduce_maybe_retain(bucket, i, force_default_stream=True) def comm_ready_buckets(self, param): # Need to do this in every hook for compatibility with Ruberry's streaming backward PR. # self.reduction_stream.wait_stream(torch.cuda.current_stream()) if self.prof: torch.cuda.nvtx.range_push("comm_ready_buckets") bucket_idx, bucket_loc = self.param_id_to_bucket[id(param)] if self.buckets[bucket_idx][bucket_loc] is not None: raise RuntimeError("The backward pass is attempting to replace an already-filled " "bucket slot. This is almost certainly an error.") if self.retain_allreduce_buffers: self.buckets[bucket_idx][bucket_loc] = param.grad else: self.buckets[bucket_idx][bucket_loc] = param.grad.data self.buckets_ready_size[bucket_idx] += 1 if self.buckets_ready_size[bucket_idx] == self.bucket_sizes[bucket_idx]: if bucket_idx == self.next_bucket: self.allreduce_maybe_retain(self.buckets[bucket_idx], bucket_idx) self.next_bucket += 1 # Reversing upstream's logic here, because we constructed our buckets based on # the order things were received during backward. if len(self.ready_buckets_not_reduced) > 0: sorted_todo = sorted(self.ready_buckets_not_reduced) for i in sorted_todo: # Nothing can be reduced now if i > self.next_bucket: break elif i == self.next_bucket: self.allreduce_maybe_retain(self.buckets[i], i) self.ready_buckets_not_reduced.remove(i) self.next_bucket += 1 else: raise ValueError("i should always be >= next_bucket") else: self.ready_buckets_not_reduced.add(bucket_idx) if self.prof: torch.cuda.nvtx.range_pop() def forward(self, *inputs, **kwargs): result = self.module(*inputs, **kwargs) if self.prof: torch.cuda.nvtx.range_push("forward pass DDP logic") if not self._disable_allreduce: if not self.delay_allreduce: param_list = [param for param in self.module.parameters() if param.requires_grad] # Conditions under which to refresh self.record # Forward has the authority to set needs_refresh to True, but only allreduce_params # in backward has the authority to set needs_refresh to False. # Parentheses are not necessary for correct order of operations, but make the intent clearer. if ((not self.active_params) or (len(param_list) != len(self.active_params)) or any([param1 is not param2 for param1, param2 in zip(param_list, self.active_params)])): self.needs_refresh = True if self.needs_refresh: self.active_i_buckets = [] self.buckets = [] self.tmp_buckets = [[], [], []] # [running half, float, double buckets] self.tmp_numels = [0, 0, 0] self.bucket_sizes = [] self.param_id_to_active_i = {id(param) : i for i, param in enumerate(param_list)} self.param_id_to_bucket = {} self.bucket_pgs = [] self.bucket_streams = [] self.bucket_events = [] else: # self.buckets = [[None for _ in range(self.bucket_sizes[i])] # for i in range(self.num_buckets)] if not self.buckets: self.buckets = [[None for _ in range(self.bucket_sizes[i])] for i in range(self.num_buckets)] else: assert len(self.buckets) == self.num_buckets, "len(buckets) = {}, expected {}".format( len(self.buckets), self.num_buckets) for b, bucket in enumerate(self.buckets): assert len(bucket) == self.bucket_sizes[b], "len(buckets[{}]) = {}, expected {})".format( b, len(buckets[b]), self.bucket_sizes[b]) for i in range(len(bucket)): bucket[i] = None if self.allreduce_communicators: self.bucket_pgs = self.allreduce_communicators[0] self.bucket_streams = self.allreduce_communicators[1] self.bucket_events = [torch.cuda.Event(enable_timing=False, blocking=False) for _ in range(self.num_allreduce_streams)] else: if self.allreduce_different_streams: if not self.bucket_pgs: self.bucket_pgs = [dist.new_group() for _ in range(self.num_allreduce_streams)] for i, bg in enumerate(self.bucket_pgs): print("rank {} created group {} with backend {}".format( dist.get_rank(), i, dist.get_backend(bg))) if self.allreduce_different_streams: if not self.bucket_streams: self.bucket_streams = [torch.cuda.Stream() for _ in range(self.num_allreduce_streams)] self.bucket_events = [torch.cuda.Event(enable_timing=False, blocking=False) for _ in range(self.num_allreduce_streams)] else: if not self.bucket_streams: self.bucket_streams = [torch.cuda.Stream()] self.bucket_events = [torch.cuda.Event(enable_timing=False, blocking=False)] self.buckets_ready_size = [0 for i in range(self.num_buckets)] if(self.retain_allreduce_buffers): self.allreduce_buffers = [None for _ in range(self.num_buckets)] self.next_bucket = 0 self.ready_buckets_not_reduced = set() self.active_params = param_list self.callback_queued = False if self.prof: torch.cuda.nvtx.range_pop() return result
GeneSplice-main
GeneSplice/apex/apex/parallel/distributed.py
import torch from torch.autograd.function import Function import syncbn from apex.parallel import ReduceOp class SyncBatchnormFunction(Function): @staticmethod def forward(ctx, input, z, weight, bias, running_mean, running_variance, eps, track_running_stats = True, momentum = 1.0, process_group = None, channel_last = False, fuse_relu = False): input = input.contiguous() world_size = 0 mean = None var_biased = None inv_std = None var = None out = None count = None if track_running_stats: if channel_last: count = int(input.numel()/input.size(-1)) mean, var_biased = syncbn.welford_mean_var_c_last(input) num_channels = input.size(-1) else: count = int(input.numel()/input.size(1)) mean, var_biased = syncbn.welford_mean_var(input) num_channels = input.size(1) if torch.distributed.is_initialized(): if not process_group: process_group = torch.distributed.group.WORLD device = mean.device world_size = torch.distributed.get_world_size(process_group) count_t = torch.empty(1, dtype=mean.dtype, device=mean.device).fill_(count) combined = torch.cat([mean.view(-1), var_biased.view(-1), count_t], dim=0) combined_list = [torch.empty_like(combined) for k in range(world_size)] torch.distributed.all_gather(combined_list, combined, process_group) combined = torch.stack(combined_list, dim=0) mean_all, invstd_all, count_all = torch.split(combined, num_channels, dim=1) count_all = count_all.view(-1) mean, var, inv_std = syncbn.welford_parallel(mean_all, invstd_all, count_all.to(torch.int32), eps) else: device = mean.device count_all = torch.cuda.IntTensor([count], device=device) inv_std = 1.0 / torch.sqrt(var_biased + eps) var = var_biased * (count) / (count-1) if count == 1 and world_size < 2: raise ValueError('Expected more than 1 value per channel when training, got input size{}'.format(input.size())) r_m_inc = mean if running_mean.dtype != torch.float16 else mean.half() r_v_inc = var if running_variance.dtype != torch.float16 else var.half() running_mean.data = running_mean.data * (1-momentum) + momentum*r_m_inc running_variance.data = running_variance.data * (1-momentum) + momentum*r_v_inc else: mean = running_mean.data inv_std = 1.0 / torch.sqrt(running_variance.data + eps) ctx.save_for_backward(input, weight, mean, inv_std, z, bias, count_all.to(torch.int32)) ctx.process_group = process_group ctx.channel_last = channel_last ctx.world_size = world_size ctx.fuse_relu = fuse_relu if channel_last: out = syncbn.batchnorm_forward_c_last(input, z, mean, inv_std, weight, bias, fuse_relu) else: out = syncbn.batchnorm_forward(input, mean, inv_std, weight, bias) return out @staticmethod def backward(ctx, grad_output): grad_output = grad_output.contiguous() # mini batch mean & var are calculated by forward path. # mu = 1./N*np.sum(h, axis = 0) # var = 1./N*np.sum((h-mu)**2, axis = 0) saved_input, weight, mean, inv_std, z, bias, count = ctx.saved_tensors process_group = ctx.process_group channel_last = ctx.channel_last world_size = ctx.world_size fuse_relu = ctx.fuse_relu grad_input = grad_z = grad_weight = grad_bias = None if fuse_relu: grad_output = syncbn.relu_bw_c_last(grad_output, saved_input, z, mean, inv_std, weight, bias) if isinstance(z, torch.Tensor) and ctx.needs_input_grad[1]: grad_z = grad_output.clone() # TODO: update kernel to not pre_divide by item_num if channel_last: sum_dy, sum_dy_xmu, grad_weight, grad_bias = syncbn.reduce_bn_c_last(grad_output, saved_input, mean, inv_std, weight) else: sum_dy, sum_dy_xmu, grad_weight, grad_bias = syncbn.reduce_bn(grad_output, saved_input, mean, inv_std, weight) # calculate grad_input if ctx.needs_input_grad[0]: if torch.distributed.is_initialized(): num_channels = sum_dy.shape[0] combined = torch.cat([sum_dy, sum_dy_xmu], dim=0) torch.distributed.all_reduce( combined, torch.distributed.ReduceOp.SUM, process_group, async_op=False) sum_dy, sum_dy_xmu = torch.split(combined, num_channels) if channel_last: grad_input = syncbn.batchnorm_backward_c_last(grad_output, saved_input, mean, inv_std, weight, sum_dy, sum_dy_xmu, count) else: grad_input = syncbn.batchnorm_backward(grad_output, saved_input, mean, inv_std, weight, sum_dy, sum_dy_xmu, count) if weight is None or not ctx.needs_input_grad[2]: grad_weight = None if weight is None or not ctx.needs_input_grad[3]: grad_bias = None return grad_input, grad_z, grad_weight, grad_bias, None, None, None, None, None, None, None, None
GeneSplice-main
GeneSplice/apex/apex/parallel/optimized_sync_batchnorm_kernel.py
import torch from torch import nn from torch.nn.parameter import Parameter class LARC(object): """ :class:`LARC` is a pytorch implementation of both the scaling and clipping variants of LARC, in which the ratio between gradient and parameter magnitudes is used to calculate an adaptive local learning rate for each individual parameter. The algorithm is designed to improve convergence of large batch training. See https://arxiv.org/abs/1708.03888 for calculation of the local learning rate. In practice it modifies the gradients of parameters as a proxy for modifying the learning rate of the parameters. This design allows it to be used as a wrapper around any torch.optim Optimizer. ``` model = ... optim = torch.optim.Adam(model.parameters(), lr=...) optim = LARC(optim) ``` It can even be used in conjunction with apex.fp16_utils.FP16_optimizer. ``` model = ... optim = torch.optim.Adam(model.parameters(), lr=...) optim = LARC(optim) optim = apex.fp16_utils.FP16_Optimizer(optim) ``` Args: optimizer: Pytorch optimizer to wrap and modify learning rate for. trust_coefficient: Trust coefficient for calculating the lr. See https://arxiv.org/abs/1708.03888 clip: Decides between clipping or scaling mode of LARC. If `clip=True` the learning rate is set to `min(optimizer_lr, local_lr)` for each parameter. If `clip=False` the learning rate is set to `local_lr*optimizer_lr`. eps: epsilon kludge to help with numerical stability while calculating adaptive_lr """ def __init__(self, optimizer, trust_coefficient=0.02, clip=True, eps=1e-8): self.optim = optimizer self.trust_coefficient = trust_coefficient self.eps = eps self.clip = clip def __getstate__(self): return self.optim.__getstate__() def __setstate__(self, state): self.optim.__setstate__(state) @property def state(self): return self.optim.state def __repr__(self): return self.optim.__repr__() @property def param_groups(self): return self.optim.param_groups @param_groups.setter def param_groups(self, value): self.optim.param_groups = value def state_dict(self): return self.optim.state_dict() def load_state_dict(self, state_dict): self.optim.load_state_dict(state_dict) def zero_grad(self): self.optim.zero_grad() def add_param_group(self, param_group): self.optim.add_param_group( param_group) def step(self): with torch.no_grad(): weight_decays = [] for group in self.optim.param_groups: # absorb weight decay control from optimizer weight_decay = group['weight_decay'] if 'weight_decay' in group else 0 weight_decays.append(weight_decay) group['weight_decay'] = 0 for p in group['params']: if p.grad is None: continue param_norm = torch.norm(p.data) grad_norm = torch.norm(p.grad.data) if param_norm != 0 and grad_norm != 0: # calculate adaptive lr + weight decay adaptive_lr = self.trust_coefficient * (param_norm) / (grad_norm + param_norm * weight_decay + self.eps) # clip learning rate for LARC if self.clip: # calculation of adaptive_lr so that when multiplied by lr it equals `min(adaptive_lr, lr)` adaptive_lr = min(adaptive_lr/group['lr'], 1) p.grad.data += weight_decay * p.data p.grad.data *= adaptive_lr self.optim.step() # return weight decay control to optimizer for i, group in enumerate(self.optim.param_groups): group['weight_decay'] = weight_decays[i]
GeneSplice-main
GeneSplice/apex/apex/parallel/LARC.py
import torch import sys import subprocess def docstring_hack(): """ Multiproc file which will launch a set of processes locally for multi-gpu usage: python -m apex.parallel.multiproc main.py ... """ pass argslist = list(sys.argv)[1:] world_size = torch.cuda.device_count() if '--world-size' in argslist: world_size = int(argslist[argslist.index('--world-size')+1]) else: argslist.append('--world-size') argslist.append(str(world_size)) workers = [] for i in range(world_size): if '--rank' in argslist: argslist[argslist.index('--rank')+1] = str(i) else: argslist.append('--rank') argslist.append(str(i)) stdout = None if i == 0 else open("GPU_"+str(i)+".log", "w") print(argslist) p = subprocess.Popen([str(sys.executable)]+argslist, stdout=stdout) workers.append(p) for p in workers: p.wait()
GeneSplice-main
GeneSplice/apex/apex/parallel/multiproc.py