python_code
stringlengths
0
456k
from typing import Dict, List, Tuple from torch.fx.node import Node from .trace_indice import TraceIndice from .utils import ( NodeMgr, find_chunk_all_input_nodes, find_chunk_compute_input_and_output_nodes, find_tensor_shape_node, flat_list, get_node_name, get_node_shape, is_non_compute_node, ) class TraceFlow(object): def __init__(self, trace_indice: TraceIndice, node_mgr: NodeMgr) -> None: self.trace_indice = trace_indice self.node_mgr = node_mgr def check_index_source(self, start_dim, start_node, start_idx, end_dim, end_node): """ Check 2 given index: one index should be source of the other Args: start_idx(int): start node chunk dim start_node(node): start node end_idx(int): end node chunk dim end_node(node): end node Returns: bool: True if check pass """ # we use start_node_idx instead of real chunk index start_node_idx = self.node_mgr.find_node_idx(start_node) end_node_trace = self.trace_indice._find_trace_from_node(end_node) end_node_trace_source = end_node_trace["source"][end_dim] sorted_source = sorted(end_node_trace_source.items(), key=lambda d: d[0], reverse=True) for node_idx, node_dim in sorted_source: if node_idx == start_node_idx and start_dim in node_dim: return True # it means we meet a node outside the loop, and the node is not input node if node_idx < start_node_idx: return False return False def check_index_compute(self, start_idx, end_dim, end_node, end_idx): """ Check 2 given index: check they haven't been computed in the source trace. Args: start_idx(int): start node chunk dim start_node(node): start node end_idx(int): end node chunk dim end_node(node): end node Returns: bool: True if check pass """ end_node_trace = self.trace_indice._find_trace_from_node(end_node) end_node_compute = end_node_trace["compute"][end_dim] if any(start_idx <= i <= end_idx for i in end_node_compute): return False return True def _assgin_single_node_flow( self, arg_node: Node, start_idx: int, end_idx: int, cur_node: Node, cur_node_dim: int, cur_node_compute: Dict, cur_node_source: Dict, cur_node_fix_dim: List, all_node_info: Dict, next_node_list: List, ) -> bool: """ Given the current node and one of its arg node, this function finds out arg node's chunk dim and fix dim Args: arg_node (Node): input node start_idx (int): chunk region start end_idx (int): chunk region end cur_node_dim (int): current node chunk dim cur_node_compute (Dict): current node compute dict cur_node_source (Dict): current node source dict cur_node_fix_dim (List): current node fix dim all_node_info (Dict): all node chunk info in the chunk region next_node_list (List) Returns: bool: True if this node can be added to the flow, vice versa. """ arg_idx = self.node_mgr.find_node_idx(arg_node) # arg in chunk range or be inputs if not (start_idx <= arg_idx < end_idx): return True # get fix dim arg_fix_dim = [] if cur_node_dim is not None: for i in cur_node_fix_dim: fix_dim_source = cur_node_source[i] if arg_idx in fix_dim_source: arg_fix_dim.append(fix_dim_source[arg_idx][0]) if arg_node in all_node_info: arg_fix_dim = list(set(all_node_info[arg_node]["fix_dim"] + arg_fix_dim)) # find arg dim if cur_node_dim is not None: # dim is computed if arg_idx in cur_node_compute[cur_node_dim]: return False if arg_idx not in cur_node_source[cur_node_dim]: arg_dim = None else: arg_dim = cur_node_source[cur_node_dim][arg_idx][0] # chunk dim cannot be in fix dims if arg_dim in arg_fix_dim: return False # chunk dim should be None if shape size is 1 if get_node_shape(arg_node)[arg_dim] == 1: arg_dim = None # chunk shape should equal cur node elif get_node_shape(arg_node)[arg_dim] != 1: if cur_node_dim is not None and get_node_shape(cur_node)[cur_node_dim] != 1: if get_node_shape(arg_node)[arg_dim] != get_node_shape(cur_node)[cur_node_dim]: return False else: arg_dim = None # add arg rest dim as fix dim arg_fix_dim = list(range(len(get_node_shape(arg_node)))) if arg_dim is not None: arg_fix_dim.remove(arg_dim) # if already in node_info, arg dim must be same if arg_node in all_node_info: if all_node_info[arg_node]["chunk_dim"] != arg_dim: return False all_node_info[arg_node]["fix_dim"] = arg_fix_dim # else add it to list else: all_node_info[arg_node] = {"chunk_dim": arg_dim, "fix_dim": arg_fix_dim} next_node_list.append(arg_node) return True def _get_all_node_info(self, end_dim, start_idx, end_idx): cur_node_list = [self.node_mgr.get_node_by_idx(end_idx)] # start from the last node all_node_info = {cur_node_list[0]: {"chunk_dim": end_dim, "fix_dim": []}} while len(cur_node_list) > 0: next_node_list = [] for cur_node in cur_node_list: # get cur node info cur_node_chunk_dim = all_node_info[cur_node]["chunk_dim"] cur_node_fix_dim = all_node_info[cur_node]["fix_dim"] if cur_node_chunk_dim is not None: cur_node_compute = self.trace_indice._find_compute_trace_from_node(cur_node) cur_node_source = self.trace_indice._find_source_trace_from_node(cur_node) else: cur_node_compute = cur_node_source = None # get all valid args arg_list = [] for arg in cur_node.all_input_nodes: if type(arg) != type(cur_node): continue if is_non_compute_node(arg): continue if get_node_shape(arg) is None: continue arg_list.append(arg) flow_flag = self._assgin_single_node_flow( arg, start_idx, end_idx, cur_node, cur_node_chunk_dim, cur_node_compute, cur_node_source, cur_node_fix_dim, all_node_info, next_node_list, ) if flow_flag == False: return None cur_node_list = next_node_list return all_node_info def _get_input_nodes_dim(self, inputs: List[Node], start_idx: int, end_idx: int, all_node_info: Dict) -> Tuple: """ Get chunk dim for every input node for their every entry, remove unchunked nodes Args: inputs (List[Node]): input nodes all_node_info (Dict): describe all node's chunk dim and fix dim start_idx (int): chunk start idx end_idx (int): chunk end idx Returns: inputs (List(Node)): new inputs inputs_dim (List): chunk dim for inputs """ inputs_dim = [] remove_inputs = [] for input_node in inputs: input_dict = {} input_node_idx = self.node_mgr.find_node_idx(input_node) for user in input_node.users.keys(): # skip non compute if is_non_compute_node(user): continue # untraced node, mostly non compute if user not in all_node_info: continue user_idx = self.node_mgr.find_node_idx(user) if start_idx <= user_idx <= end_idx: chunk_dim = all_node_info[user]["chunk_dim"] if chunk_dim is not None: user_source = self.trace_indice._find_source_trace_from_node(user)[chunk_dim] if input_node_idx in user_source: if get_node_shape(input_node)[user_source[input_node_idx][0]] == 1: input_dict[user_idx] = [None] else: input_dict[user_idx] = user_source[input_node_idx] else: return None, None if len(input_dict) == 0: remove_inputs.append(input_node) else: inputs_dim.append(input_dict) # remove unchunked inputs for i in remove_inputs: if i in inputs: inputs.remove(i) return inputs, inputs_dim def _get_prepose_nodes(self, all_node_info: Dict, start_idx: int, end_idx: int, chunk_info) -> List[Node]: """ get all useless nodes in chunk region and prepose them Args: all_node_info (Dict): describe all node's chunk dim and fix dim start_idx (int): chunk start idx end_idx (int): chunk end idx Returns: List[Node]: all nodes to be preposed """ # get all possible prepose nodes maybe_prepose_nodes = [] for node, node_info in all_node_info.items(): if node_info["chunk_dim"] is None: maybe_prepose_nodes.append(node) for node in self.node_mgr.get_node_slice_by_idx(start_idx, end_idx): if node not in all_node_info and node not in chunk_info["outputs"]: maybe_prepose_nodes.append(node) maybe_prepose_nodes.sort( key=lambda x: self.node_mgr.find_node_idx(x), reverse=True, ) # from last node to first node prepose_nodes = [] # set every node as root, search its args, if all legal, turn root and args as prepose nodes while len(maybe_prepose_nodes) > 0: tmp_cur_prepose_nodes = [maybe_prepose_nodes[0]] tmp_cur_related_prepose_nodes = [] prepose_flag = True # loop cur node's all arg until out of chunk while len(tmp_cur_prepose_nodes) > 0: if prepose_flag == False: break tmp_next_prepose_nodes = [] tmp_cur_related_prepose_nodes.extend(tmp_cur_prepose_nodes) for cur_prepose_node in tmp_cur_prepose_nodes: if prepose_flag == False: break for cur_prepose_node_arg in cur_prepose_node.all_input_nodes: if type(cur_prepose_node_arg) != type(cur_prepose_node): continue # out of loop if not (start_idx <= self.node_mgr.find_node_idx(cur_prepose_node_arg) < end_idx): continue # compute op in loop elif cur_prepose_node_arg in all_node_info: if all_node_info[cur_prepose_node_arg]["chunk_dim"] is None: tmp_next_prepose_nodes.append(cur_prepose_node_arg) else: prepose_flag = False break # non compute op else: tmp_next_prepose_nodes.append(cur_prepose_node_arg) tmp_cur_prepose_nodes = tmp_next_prepose_nodes if prepose_flag == False: maybe_prepose_nodes.remove(maybe_prepose_nodes[0]) continue else: for n in tmp_cur_related_prepose_nodes: if n not in prepose_nodes: prepose_nodes.append(n) if n in maybe_prepose_nodes: maybe_prepose_nodes.remove(n) # sort by index prepose_nodes.sort(key=lambda x: self.node_mgr.find_node_idx(x)) chunk_info["args"]["prepose_nodes"] = prepose_nodes def _get_non_chunk_inputs(self, chunk_info, start_idx, end_idx): # we need to log input nodes to avoid deleteing them in the loop chunk_node_list = self.node_mgr.get_node_slice_by_idx(start_idx, end_idx + 1) # also need to get some prepose node's arg out of non_chunk_inputs for n in chunk_info["args"]["prepose_nodes"]: chunk_node_list.remove(n) non_chunk_inputs = find_chunk_all_input_nodes(chunk_node_list) for i in non_chunk_inputs: if i not in chunk_info["inputs"]: chunk_info["inputs_non_chunk"].append(i) return chunk_info def flow_search(self, start_idx, start_dim, end_idx, end_dim): inputs, outputs = find_chunk_compute_input_and_output_nodes( self.node_mgr.get_node_slice_by_idx(start_idx, end_idx + 1)) # get every node's chunk dim and fix dim all_node_info = self._get_all_node_info(end_dim, start_idx, end_idx) if all_node_info is None: return None chunk_info = { "region": (start_idx, end_idx), "inputs": [], "inputs_non_chunk": [], "inputs_dim": [], "outputs": [self.node_mgr.get_node_by_idx(end_idx)], "outputs_non_tensor": {}, "outputs_dim": [end_dim], "node_chunk_dim": all_node_info, "args": {}, } # find chunk info for other outputs if len(find_tensor_shape_node(outputs)) > 1: chunk_info = self._get_other_output_info(outputs, start_idx, start_dim, end_idx, end_dim, chunk_info) if chunk_info is None: return None # get input nodes' chunk dim inputs, inputs_dim = self._get_input_nodes_dim(inputs, start_idx, end_idx, all_node_info) if inputs is None: return None chunk_info["inputs"] = inputs chunk_info["inputs_dim"] = inputs_dim # move useless nodes ahead of loop self._get_prepose_nodes(all_node_info, start_idx, end_idx, chunk_info) # find non chunk inputs chunk_info = self._get_non_chunk_inputs(chunk_info, start_idx, end_idx) # reassgin reshape size, some size may have changed due to chunk chunk_info = self._reassgin_reshape_size(chunk_info) return chunk_info def _get_other_output_info(self, outputs: List[Node], start_idx: int, start_dim: int, end_idx: int, end_dim: int, chunk_info: Dict): start_node = self.node_mgr.get_node_by_idx(start_idx) # loop all outputs for output in outputs: output_legal = False output_idx = self.node_mgr.find_node_idx(output) # skip the origin output if output_idx == end_idx: continue # skip non tensor if get_node_shape(output) is None: # log shape tensor if len(output.meta['fwd_out']) > 0 and isinstance(output.meta['fwd_out'][0], int): chunk_info["outputs_non_tensor"][output] = str(output.meta['fwd_out']) continue # loop every dim of outputs, try to find a legal one for output_dim in range(len(get_node_shape(output))): if not self.check_region_start_end(start_node, start_dim, start_idx, output, output_dim, output_idx): continue new_all_node_info = self._get_all_node_info(output_dim, start_idx, output_idx) if new_all_node_info is None: continue # check node info legal if self._update_chunk_info(chunk_info, new_all_node_info, output, output_dim) == True: output_legal = True break # not legal if output_legal == False: return None return chunk_info def _update_chunk_info(self, chunk_info: Dict, new_all_node_info: Dict, output: Node, output_dim: int) -> bool: """ check if there is conflict between new node info and old chunk info. If not, update old chunk info """ # check if conflict overlap_flag = False for k, v in new_all_node_info.items(): if k in chunk_info["node_chunk_dim"]: overlap_flag = True if chunk_info["node_chunk_dim"][k]["chunk_dim"] != v["chunk_dim"]: return False # if no overlap, we just consider them as prepose nodes, instead of new output if overlap_flag == False: return True # update chunk info for k, v in new_all_node_info.items(): if k in chunk_info["node_chunk_dim"]: chunk_info["node_chunk_dim"][k]["fix_dim"] = list( set(chunk_info["node_chunk_dim"][k]["fix_dim"] + v["fix_dim"])) else: chunk_info["node_chunk_dim"][k] = v chunk_info["outputs"].append(output) chunk_info["outputs_dim"].append(output_dim) return True def _reassgin_reshape_size(self, chunk_info): """ Some shape args in reshape may have changed due to chunk reassgin those changed shape """ chunk_region = chunk_info["region"] reshape_size = {} chunk_shape = get_node_shape(chunk_info["outputs"][0])[chunk_info["outputs_dim"][0]] for node in self.node_mgr.get_node_slice_by_idx(chunk_region[0], chunk_region[1] + 1): if any(i == get_node_name(node) for i in ["reshape", "view"]): if node in chunk_info["args"]["prepose_nodes"]: continue if node.args[0] in chunk_info["inputs_non_chunk"]: continue reshape_args = flat_list(node.args[1:]) if len(reshape_args) == 1 and get_node_shape(reshape_args[0]) is None and len( reshape_args[0].meta['fwd_out']) > 1: continue chunk_dim = chunk_info["node_chunk_dim"][node]["chunk_dim"] new_shape = "" for reshape_arg_dim, reshape_arg in enumerate(reshape_args): if reshape_arg_dim == chunk_dim: new_shape += "min(chunk_size, %d - chunk_idx), " % chunk_shape else: if isinstance(reshape_arg, int): new_shape += "%s, " % str(reshape_arg) else: new_shape += "%s, " % reshape_arg.name new_shape = new_shape[:-2] origin_shape = str(reshape_args)[1:-1] reshape_size[node.name] = [origin_shape, new_shape] chunk_info["reshape_size"] = reshape_size return chunk_info def check_region_start_end(self, start_node: Node, start_dim: int, start_idx: int, end_node: Node, end_dim: int, end_idx: int) -> bool: """ check if region start and end is legal """ # dim cannot be None if (get_node_shape(end_node) is None or get_node_shape(start_node) is None): return False # dim size cannot be 1 if (get_node_shape(end_node)[end_dim] == 1 or get_node_shape(start_node)[start_dim] == 1): return False # must have users if len(end_node.users) == 0: return False # check index source align if not self.check_index_source(start_dim, start_node, start_idx, end_dim, end_node): return False # check index copmute if not self.check_index_compute(start_idx, end_dim, end_node, end_idx): return False return True
from .estimate_memory import EstimateMemory from .reorder_graph import ReorderGraph from .trace_indice import TraceIndice from .utils import NodeMgr, is_non_compute_node class SelectChunk(object): def __init__( self, trace_indice: TraceIndice, estimate_memory: EstimateMemory, reorder_graph: ReorderGraph, node_mgr: NodeMgr, max_memory=None, ): self.trace_indice = trace_indice self.estimate_memory = estimate_memory self.reorder_graph = reorder_graph self.node_mgr = node_mgr if max_memory is not None: self.stratge = "fit_memory" self.max_memory = max_memory # MB else: self.stratge = "min_memory" def _select_best_chunk_region(self, possible_chunk_regions, chunk_infos, peak_node, max_chunk_region, mem_peak): if self.stratge == "min_memory": best_region = self._select_min_memory_chunk_region( possible_chunk_regions, chunk_infos, peak_node, max_chunk_region, mem_peak, ) elif self.stratge == "fit_memory": best_region = self._select_fit_memory_chunk_region( possible_chunk_regions, chunk_infos, peak_node, max_chunk_region, mem_peak, ) else: raise RuntimeError() return best_region def _select_fit_memory_chunk_region(self, possible_chunk_regions, chunk_infos, peak_node, max_chunk_region, mem_peak): # stop chunk if max memory satisfy memory limit if max(mem_peak) < self.max_memory: return None # remove illegal regions illegal_regions = [] for i in possible_chunk_regions: if not self._is_legal_region(i, chunk_infos): illegal_regions.append(i) for i in illegal_regions: if i in possible_chunk_regions: possible_chunk_regions.remove(i) if len(possible_chunk_regions) == 0: return None max_possible_chunk_region = (min([i["region"][0] for i in possible_chunk_regions]), max([i["region"][1] for i in possible_chunk_regions])) # get mem for chunk region regions_dict = [] for region in possible_chunk_regions: cur_region = region.copy() cur_node_list, cur_region = self.reorder_graph.tmp_reorder(self.node_mgr.get_node_list(), cur_region) cur_chunk_infos = chunk_infos + [cur_region] cur_mem_peak = self.estimate_memory.estimate_chunk_inference_mem(cur_node_list, cur_chunk_infos)[0] cur_chunk_region_peak = cur_mem_peak[max_possible_chunk_region[0]:max_possible_chunk_region[1] + 1] cur_chunk_region_max_peak = max(cur_chunk_region_peak) if cur_chunk_region_max_peak < self.max_memory: regions_dict.append({ "chunk_info": region, "chunk_max_mem": cur_chunk_region_max_peak, "chunk_len": self._get_compute_node_num(region["region"][0], region["region"][1]), "reorder_chunk_info": cur_region, "reorder_node_list": cur_node_list, }) # no region found if len(regions_dict) == 0: raise RuntimeError("Search failed. Try a larger memory threshold.") # select the min chunk len chunk_len = [i["chunk_len"] for i in regions_dict] best_region_idx = chunk_len.index(min(chunk_len)) best_region = regions_dict[best_region_idx] # get max chunk size best_region = self._get_fit_chunk_size(best_region, chunk_infos) return best_region def _get_fit_chunk_size(self, chunk_region_dict, chunk_infos): chunk_size = 1 reorder_chunk_info = chunk_region_dict["reorder_chunk_info"] reorder_chunk_info["chunk_size"] = chunk_size cur_chunk_max_mem = 0 # search a region while cur_chunk_max_mem < self.max_memory: chunk_size *= 2 reorder_chunk_info["chunk_size"] = chunk_size cur_chunk_infos = chunk_infos + [reorder_chunk_info] cur_mem_peak = self.estimate_memory.estimate_chunk_inference_mem(chunk_region_dict["reorder_node_list"], cur_chunk_infos)[0] cur_chunk_max_mem = max(cur_mem_peak[reorder_chunk_info["region"][0]:reorder_chunk_info["region"][1] + 1]) # search exact size chunk_info = chunk_region_dict["chunk_info"] chunk_info["chunk_size"] = self._chunk_size_binary_search(chunk_size // 2, chunk_size, chunk_region_dict, chunk_infos) return chunk_info def _chunk_size_binary_search(self, left, right, chunk_region_dict, chunk_infos): if left >= 16: gap = 4 else: gap = 1 chunk_info = chunk_region_dict["reorder_chunk_info"] while right >= left + gap: mid = int((left + right) / 2 + 0.5) chunk_info["chunk_size"] = mid cur_chunk_infos = chunk_infos + [chunk_info] cur_mem_peak = self.estimate_memory.estimate_chunk_inference_mem(chunk_region_dict["reorder_node_list"], cur_chunk_infos)[0] cur_chunk_max_mem = max(cur_mem_peak[chunk_info["region"][0]:chunk_info["region"][1] + 1]) if cur_chunk_max_mem >= self.max_memory: right = mid - gap else: left = mid + gap return left def _get_compute_node_num(self, start, end): count = 0 for i in self.node_mgr.get_node_slice_by_idx(start, end + 1): if not is_non_compute_node(i): count += 1 return count def _select_min_memory_chunk_region(self, possible_chunk_regions, chunk_infos, peak_node, max_chunk_region, mem_peak): # remove illegal regions illegal_regions = [] for i in possible_chunk_regions: if not self._is_legal_region(i, chunk_infos): illegal_regions.append(i) for i in illegal_regions: if i in possible_chunk_regions: possible_chunk_regions.remove(i) if len(possible_chunk_regions) == 0: return None # get max possible chunk region max_possible_chunk_region = (min([i["region"][0] for i in possible_chunk_regions]), max([i["region"][1] for i in possible_chunk_regions])) # get mem for chunk region regions_dict_list = [] for region in possible_chunk_regions: cur_region = region.copy() cur_node_list, cur_region = self.reorder_graph.tmp_reorder(self.node_mgr.get_node_list(), cur_region) cur_chunk_infos = chunk_infos + [cur_region] cur_mem_peak = self.estimate_memory.estimate_chunk_inference_mem(cur_node_list, cur_chunk_infos)[0] cur_chunk_region_peak = cur_mem_peak[max_possible_chunk_region[0]:max_possible_chunk_region[1] + 1] cur_chunk_region_max_peak = max(cur_chunk_region_peak) regions_dict_list.append({ "chunk_info": region, "chunk_max_mem": cur_chunk_region_max_peak, "chunk_len": self._get_compute_node_num(region["region"][0], region["region"][1]), "reorder_chunk_info": cur_region, "reorder_node_list": cur_node_list, }) # select the min mem chunk_max_mem = [i["chunk_max_mem"] for i in regions_dict_list] best_region_idx = chunk_max_mem.index(min(chunk_max_mem)) best_region = regions_dict_list[best_region_idx]["chunk_info"] if best_region is not None: best_region["chunk_size"] = 1 return best_region def _is_legal_region(self, cur_chunk_info, chunk_infos): (chunk_region_start, chunk_region_end) = cur_chunk_info["region"] if cur_chunk_info in chunk_infos: return False if chunk_region_end < chunk_region_start: return False for i in chunk_infos: region = i["region"] if not ((chunk_region_start > region[1] and chunk_region_end > region[1]) or (chunk_region_start < region[0] and chunk_region_end < region[0])): return False return True
from typing import Any, Callable, Dict, Iterable, List, Tuple, Union from torch.fx.node import Node from colossalai.logging import get_dist_logger NON_COMPUTE_OP = ["placeholder", "get_attr", "output"] NON_COMPUTE_NAME = ["getattr", "eq", "_assert_is_none", "_assert", "finfo", "size"] logger = get_dist_logger() class NodeMgr(object): def __init__(self, gm) -> None: self._node_list = list(gm.graph.nodes) self._node_dict = {} self._set_node_dict() def _set_node_dict(self) -> None: """ create a dict {node_name: node_idx} """ self._node_dict.clear() for idx, node in enumerate(self._node_list): self._node_dict[node.name] = idx def find_node_idx(self, node: Node) -> int: """ find node's index """ return self._node_dict[node.name] def find_node_idx_by_name(self, node_name: str) -> int: """ find node's index """ return self._node_dict[node_name] def get_node_by_idx(self, idx: int) -> Node: """ get a node by index """ return self._node_list[idx] def get_node_slice_by_idx(self, start: int, end: int) -> List[Node]: """ get a slice of node by index """ return self._node_list[start:end] def get_node_list(self) -> List: """ get full node list """ return self._node_list def update_node_list(self, node_list: List) -> None: """ update node list, reset node dict """ self._node_list = node_list self._set_node_dict() def get_logger() -> Any: return logger def flat_list(inputs: Any) -> List: """ flat a list by recursion """ if not (isinstance(inputs, list) or isinstance(inputs, set) or isinstance(inputs, tuple)): return [inputs] res = [] for i in inputs: if isinstance(i, list) or isinstance(i, set) or isinstance(i, tuple): res.extend(flat_list(i)) else: res.append(i) return res def find_first_tensor_arg(node: Node) -> Node: """ Find the first input tensor arg for a node """ for arg in node.args: if type(arg) == type(node): return arg raise RuntimeError() def is_non_compute_node(node: Node) -> bool: if any(i == node.op for i in NON_COMPUTE_OP) or any(i == get_node_name(node) for i in NON_COMPUTE_NAME): return True if "getitem" in node.name: if get_node_shape(node) is not None: return False node_args = flat_list(node.args[1:]) for node_arg in node_args: if any(i == str(node_arg) for i in ["None", "Ellipsis"]): return False if "slice" in str(node_arg): return False return True return False def get_node_shape(node: Node) -> List: if get_node_name(node) == "split": return node.meta["tensor_meta"][0].shape if hasattr(node.meta["tensor_meta"], "shape"): return node.meta["tensor_meta"].shape return None def is_non_memory_node(node: Node) -> bool: if "getitem" in node.name: return True if "output" in node.op: return True return is_non_compute_node(node) def is_non_compute_node_except_placeholder(node: Node) -> bool: if "placeholder" in node.op: return False return is_non_compute_node(node) def is_non_compute_node_except_placeholder_output(node: Node) -> bool: if "output" in node.op: return False return is_non_compute_node_except_placeholder(node) def find_node_idx(name: str, nodes_list: List) -> int: for idx, node in enumerate(nodes_list): if node.name == name: return idx raise RuntimeError("name %s not found in node list" % name) def delete_free_var_from_last_use(user_to_last_uses: Dict) -> None: for key, value in user_to_last_uses.items(): for n in value: if n.op == "placeholder": user_to_last_uses[key].remove(n) def find_chunk_all_input_nodes(nodes: List[Node]) -> List: """ Find non-compute input and output node names. input nodes are nodes used in the list output nodes are nodes will use nodes in the list """ input_nodes = [] for node in nodes: for input_node in node._input_nodes.keys(): if input_node not in nodes and input_node not in input_nodes: input_nodes.append(input_node) return input_nodes def find_chunk_compute_input_and_output_nodes(nodes: List[Node]) -> Union[List, List]: """ Find non-compute input and output node names. input nodes are nodes used in the list output nodes are nodes will use nodes in the list """ input_nodes = [] output_nodes = [] # if a node has an input node which is not in the node list # we treat that input node as the input of the checkpoint function for node in nodes: for input_node in node._input_nodes.keys(): if (input_node not in nodes and input_node not in input_nodes and not is_non_compute_node_except_placeholder(input_node)): input_nodes.append(input_node) # if a node has a user node which is not in the node list # we treat that user node as the node receiving the current node output for node in nodes: for output_node in node.users.keys(): if (output_node not in nodes and node not in output_nodes and not is_non_compute_node_except_placeholder_output(output_node)): output_nodes.append(node) return input_nodes, output_nodes def get_module_node_name(node: Node) -> str: """ get module class name """ node_targets = node.target.split(".") module = node.graph.owning_module for i in node_targets: module = getattr(module, i) module_name = str(module.__class__).split(".")[-1][:-2] module_name = module_name.lower() return module_name def get_node_name(node: Node) -> str: """ get node name """ node_name = node.name if "_" in node_name: for i in range(len(node_name) - 1, -1, -1): if node_name[i] == "_": node_name = node_name[:i] break elif node_name[i] in ["1", "2", "3", "4", "5", "6", "7", "8", "9", "0"]: continue else: break return node_name def find_tensor_node(node_list: List[Node]) -> List[Node]: """ find tensor nodes from a node list """ out = [] for node in node_list: if get_node_shape(node) is not None: out.append(node) return out def find_tensor_shape_node(node_list: List[Node]) -> List[Node]: """ find tensor and shape nodes from a node list """ out = [] for node in node_list: if get_node_shape(node) is not None: out.append(node) elif len(node.meta['fwd_out']) > 0 and isinstance(node.meta['fwd_out'], list) and isinstance( node.meta['fwd_out'][0], int): out.append(node) return out
import copy from typing import Dict, List, Tuple from torch.fx.node import Node from .utils import NodeMgr, find_first_tensor_arg, flat_list, get_module_node_name, get_node_name, get_node_shape class TraceIndice(object): """ Trace all indice infomation for every node. Indice is a logical concept. Equal dims can been treated as one indice. eg. dim(x1) = [a, b, c] dim(x2) = [d, e, f] and we have x3 = x1 * x2. then a=d, b=e, c=f, due to the broadcast property, dim(x1)=dim(x2)=dim(x3)=[a, b, c] This class will record every node's dims' indice, compute and source. Attibutes: node_list (List) indice_trace_list (List): [{"indice": [...], "compute": [...], "source": [...]}, {...}] indice_view_list (Dict): not used for now indice_count (int): record indice number Args: node_list (List) """ def __init__(self, node_mgr: NodeMgr) -> None: self.node_mgr = node_mgr self.indice_trace_list = self._init_indice_trace_list() self.indice_view_list = {} self.indice_count = -1 self.trace_range = [] self.active_node_list = [] def _init_indice_trace_list(self) -> List: indice_trace_list = [] for n in self.node_mgr.get_node_list(): if get_node_shape(n) != None: cur_trace = { "indice": [None for _ in range(len(get_node_shape(n)))], "compute": [[] for _ in range(len(get_node_shape(n)))], "source": [{} for _ in range(len(get_node_shape(n)))], } else: cur_trace = {"indice": [], "compute": [], "source": []} indice_trace_list.append(cur_trace) return indice_trace_list def set_trace_range(self, trace_range: List, active_node_list: List) -> None: self.trace_range = trace_range self.active_node_list = active_node_list def _add_indice(self) -> int: """ Update the count and return it. To record the idx number. Returns: indice_count: int """ self.indice_count += 1 return self.indice_count def _del_dim(self, idx: int, dim_idx: int) -> None: """ delete a dim for indice, compute and source """ self.indice_trace_list[idx]["indice"].pop(dim_idx) self.indice_trace_list[idx]["compute"].pop(dim_idx) self.indice_trace_list[idx]["source"].pop(dim_idx) def _add_dim(self, node_idx: int, dim_idx: int) -> None: """ add a dim for indice, compute and source """ self.indice_trace_list[node_idx]["indice"].insert(dim_idx, self._add_indice()) self.indice_trace_list[node_idx]["compute"].insert(dim_idx, []) self.indice_trace_list[node_idx]["source"].insert(dim_idx, {}) def _add_source( self, node_from: Node, node_from_dim: int, node_to: Node, node_to_dim: int, init=False, ) -> None: node_from_dim = self._transform_indice(node_from, node_from_dim) node_from_trace_source = self._find_source_trace_from_node(node_from) node_to_dim = self._transform_indice(node_to, node_to_dim) node_to_trace_source = self._find_source_trace_from_node(node_to) node_from_idx = self.node_mgr.find_node_idx(node_from) if init: node_to_trace_source[node_to_dim] = {} # add dim to cur new source if node_from_idx not in node_to_trace_source[node_to_dim]: node_to_trace_source[node_to_dim][node_from_idx] = [node_from_dim] else: if node_from_dim not in node_to_trace_source[node_to_dim][node_from_idx]: node_to_trace_source[node_to_dim][node_from_idx].append(node_from_dim) # update inputs source for node_idx, node_dim in node_from_trace_source[node_from_dim].items(): if node_idx not in node_to_trace_source[node_to_dim]: node_to_trace_source[node_to_dim][node_idx] = copy.deepcopy(node_dim) else: for d in node_dim: if d not in node_to_trace_source[node_to_dim][node_idx]: node_to_trace_source[node_to_dim][node_idx].append(d) def _transform_indice(self, node: Node, node_dim: int) -> int: node_idx = self._find_indice_trace_from_node(node) dims = list(range(len(node_idx))) return dims[node_dim] def _inherit_indice( self, node_from: Node, node_from_dim: int, node_to: Node, node_to_dim: int, init: bool = True, ) -> None: """ node_to's node_to_dim inherit node_from's node_from_dim by indice, compute and source """ node_from_dim = self._transform_indice(node_from, node_from_dim) node_to_dim = self._transform_indice(node_to, node_to_dim) node_from_trace = self._find_trace_from_node(node_from) node_to_trace = self._find_trace_from_node(node_to) if init: node_to_trace["indice"][node_to_dim] = node_from_trace["indice"][node_from_dim] node_to_trace["compute"][node_to_dim] = copy.deepcopy(node_from_trace["compute"][node_from_dim]) else: for j in node_from_trace["compute"][node_from_dim]: if j not in node_to_trace["compute"][node_to_dim]: node_to_trace["compute"][node_to_dim].append(j) self._add_source(node_from, node_from_dim, node_to, node_to_dim, init) def _inherit_all_indice(self, node_from: Node, node_to: Node) -> None: """ inherit all dims with init """ # find indice just for assert length node_from_indice = self._find_indice_trace_from_node(node_from) node_to_indice = self._find_indice_trace_from_node(node_to) assert len(node_from_indice) == len(node_to_indice) for i in range(len(node_from_indice)): self._inherit_indice(node_from, i, node_to, i, init=True) def _inherit_more_indice_from_node_with_exclude(self, node_from: Node, node_to: Node, exclude: List = None) -> None: """ inheirt indice from node without init """ if exclude == None: exclude = [] else: exclude = [self._transform_indice(node_to, i) for i in exclude] node_from_compute = self._find_compute_trace_from_node(node_from) node_to_compute = self._find_compute_trace_from_node(node_to) # assert len(node_from_compute) == len(node_to_compute) for i in range(-1, -min(len(node_from_compute), len(node_to_compute)) - 1, -1): if self._transform_indice(node_to, i) in exclude: continue self._inherit_indice(node_from, i, node_to, i, init=False) def _mark_computation(self, node: Node, idx: int, dim: int) -> None: """ Mark some dims of node as computed. Args: node (node) idx (int): node index dim (list or int): dims to be marked as computed """ if isinstance(dim, int): dim = [dim] dims = list(range(len(get_node_shape(node)))) for d in dim: cur_dim = dims[d] if idx not in self.indice_trace_list[idx]["compute"][cur_dim]: self.indice_trace_list[idx]["compute"][cur_dim].append(idx) def _find_trace_from_node(self, node: Node) -> Dict: """ Find node idx and compute trace by the node. Args: node (node) Returns: idx (list): idx of the node compute (list): computed idx of the node. """ node_idx = self.node_mgr.find_node_idx(node) node_dict = self.indice_trace_list[node_idx] return node_dict def _find_source_trace_from_node(self, node: Node) -> List: """ Find node source trace by the node. Args: node (node) Returns: idx (list): idx of the node compute (list): computed idx of the node. """ node_idx = self.node_mgr.find_node_idx(node) node_dict = self.indice_trace_list[node_idx] return node_dict["source"] def _find_indice_trace_from_node(self, node) -> List: """ Find node idx trace by the node. Args: node (node) Returns: idx (list): idx of the node """ node_idx = self.node_mgr.find_node_idx(node) return self.indice_trace_list[node_idx]["indice"] def _find_compute_trace_from_node(self, node: Node) -> List: """ Find node compute trace by the node. Args: node (node) Returns: compute (list): computed idx of the node. """ node_idx = self.node_mgr.find_node_idx(node) return self.indice_trace_list[node_idx]["compute"] def _assign_indice_as_input(self, node: Node, node_idx: int, input_node=None) -> None: """ Assign node's trace as its input node. Args: node (node) node_idx (int) """ if input_node == None: input_node = find_first_tensor_arg(node) self._inherit_all_indice(input_node, node) def _assign_all_indice(self, node: Node, node_idx: int) -> None: """ Add new indice for all node's dims. Args: node (node) node_idx (int) """ shape = node.meta["tensor_meta"].shape if shape is None: return new_trace = [] for _ in shape: new_trace.append(self._add_indice()) self.indice_trace_list[node_idx]["indice"] = new_trace def _assign_transpose_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for transpose op. 1. swap input's dim according to transpose args 2. inherit input's computation Args: node (node) node_idx (int) """ input_node = node.args[0] tranpose_dim = node.args[1:] self._assign_indice_as_input(node, node_idx, input_node) self._inherit_indice(input_node, tranpose_dim[1], node, tranpose_dim[0]) self._inherit_indice(input_node, tranpose_dim[0], node, tranpose_dim[1]) def _assign_permute_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for permute op. 1. swap input's dim according to permute args 2. inherit input's computation Args: node (node) node_idx (int) """ permute_dim = flat_list(node.args[1:]) input_node = node.args[0] self._assign_indice_as_input(node, node_idx, input_node) for idx, d in enumerate(permute_dim): self._inherit_indice(input_node, d, node, idx) def _assign_linear_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for linear op. 1. copy trace from input node and change last indice accroding to weight 2. mark equal for input node last indice, weight first dim and bias dim. 3. inherit input's computation, mark computation for last dim. Args: node (node) node_idx (int) """ self._assign_indice_as_input(node, node_idx) if len(node.args) >= 2: weight = node.args[1] self._inherit_indice(weight, 1, node, -1) else: self._del_dim(node_idx, -1) self._add_dim(node_idx, -1) self._mark_computation(node, node_idx, [-1]) def _assign_addmm_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for addmm op. Args: node (node) node_idx (int) """ bias, input_node, weight = node.args assert len(get_node_shape(bias)) == 1 and len(get_node_shape(weight)) == 2 self._assign_indice_as_input(node, node_idx, input_node) self._inherit_indice(weight, 1, node, -1) self._inherit_more_indice_from_node_with_exclude(bias, node) self._mark_computation(node, node_idx, [-1]) def _assign_baddbmm_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for baddbmm(batch add and batch matmul) op. add, matmul_left, matmul_right = args out = add + (matmul_left x matmul_right) Args: node (node) node_idx (int) """ add, matmul_left, matmul_right = node.args assert get_node_shape(add) == get_node_shape(node) assert len(get_node_shape(matmul_left)) == len(get_node_shape(matmul_right)) self._assign_indice_as_input(node, node_idx, matmul_left) # matmul self._inherit_indice(matmul_right, -1, node, -1) self._inherit_more_indice_from_node_with_exclude(matmul_right, node, [-2, -1]) self._mark_computation(node, node_idx, [-1]) # add self._inherit_more_indice_from_node_with_exclude(add, node) def _assign_matmul_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for matmul op. 1. copy trace from matmul_left and change last indice accroding to matmul_right. (assert they have same length) 2. mark equal for input matmul_left -1 indice and matmul_right -2 dim. 3. inherit matmul_left and matmul_right computation, mark computation for last dim. Args: node (node) node_idx (int) """ matmul_left, matmul_right = node.args assert len(get_node_shape(matmul_left)) == len(get_node_shape(matmul_right)) self._assign_indice_as_input(node, node_idx, matmul_left) self._inherit_indice(matmul_right, -1, node, -1) self._inherit_more_indice_from_node_with_exclude(matmul_right, node, [-1, -2]) self._mark_computation(node, node_idx, [-1]) def _assign_conv2d_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for conv2d op. Args: node (node) node_idx (int) """ # get conv module node_targets = node.target.split(".") conv_module = node.graph.owning_module for i in node_targets: conv_module = getattr(conv_module, i) assert conv_module.dilation == (1, 1), "dilation for conv2d not implemented" # get conv input assert len(node.args) == 1 input_node = node.args[0] assert len(get_node_shape(input_node)) == 4 # assgin index self._assign_indice_as_input(node, node_idx, input_node) self._del_dim(node_idx, 1) self._add_dim(node_idx, 1) self._mark_computation(node, node_idx, [1, 2, 3]) def _assign_interpolate_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for interpolate op. Args: node (node) node_idx (int) """ # get conv input assert node.kwargs['size'] is None assert len(get_node_shape(node)) == 4 # assgin index self._assign_indice_as_input(node, node_idx) self._mark_computation(node, node_idx, [-1, -2]) def _assign_layernorm_indice(self, node, idx): """ Assign indice for layernorm op. 1. assign indice as input node 2. inherit computation and mark last 2 dims as computed. Args: node (node) node_idx (int) """ self._assign_indice_as_input(node, idx) self._mark_computation(node, idx, [-1]) def _assign_groupnorm_indice(self, node, idx): """ Assign indice for groupnorm op. Args: node (node) node_idx (int) """ assert len(get_node_shape(node)) == 4 self._assign_indice_as_input(node, idx) self._mark_computation(node, idx, [-1, -2, -3]) def _assign_elementwise_indice(self, node, idx): """ Assign indice for element-wise op (eg. relu sigmoid add mul). 1. assign indice as input node 2. inherit computation from all input nodes. Args: node (node) node_idx (int) """ self._assign_indice_as_input(node, idx) nodes_in = [] for node_in in node.args: if type(node_in) == type(node): nodes_in.append(node_in) self._inherit_more_indice_from_node_with_exclude(node_in, node) def _assgin_no_change_indice(self, node, idx): self._assign_indice_as_input(node, idx) for node_in in node.args: if type(node_in) == type(node): self._inherit_more_indice_from_node_with_exclude(node_in, node) def _assign_einsum_indice(self, node, idx): """ Assign indice for einsum op. Args: node (node) node_idx (int) """ patterns = node.args[0] input_nodes = node.args[1:] patterns = patterns.replace(" ", "") left, right = patterns.split("->") left = left.split(",") if "..." in right: replace_list = "!@#$%^&*" target_len = len(get_node_shape(node)) add_len = target_len - len(right) + 3 replace_str = replace_list[:add_len] right = right.replace("...", replace_str) for ll in range(len(left)): left[ll] = left[ll].replace("...", replace_str) all_index = [] for i in left: for c in i: all_index.append(c) all_index = set(all_index) for right_idx, right_indice in enumerate(right): for left_idx, left_str in enumerate(left): if right_indice in left_str: source_idx = left_str.index(right_indice) self._inherit_indice(input_nodes[left_idx], source_idx, node, right_idx) def _assign_softmax_indice(self, node, idx): """ Assign indice for softmax op. 1. assign indice as input node 2. inherit computation and mark softmax dim as computed. Args: node (node) node_idx (int) """ self._assign_indice_as_input(node, idx) self._mark_computation(node, idx, [node.kwargs["dim"]]) def _assign_split_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for split op. Args: node (node) node_idx (int) """ self._assign_indice_as_input(node, node_idx) dim_idx = node.kwargs["dim"] self._del_dim(node_idx, dim_idx) self._add_dim(node_idx, dim_idx) def _assign_unsqueeze_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for unsqueeze op. 1. assign new indice for unsqueeze dim Args: node (node) node_idx (int) """ self._del_dim(node_idx, -1) self._assign_indice_as_input(node, node_idx) dim_idx = node.args[1] # unsqueeze(-1) = unsqueeze(shape_num + 1) if dim_idx < 0: dim_idx = list(range(len(get_node_shape(node))))[dim_idx] self._add_dim(node_idx, dim_idx) def _assign_cat_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for cat op. Args: node (node) node_idx (int) """ nodes_in = flat_list(node.args[0]) self._assign_indice_as_input(node, node_idx, input_node=nodes_in[0]) for n in nodes_in[1:]: self._inherit_more_indice_from_node_with_exclude(n, node) cat_dim = node.kwargs["dim"] self._del_dim(node_idx, cat_dim) self._add_dim(node_idx, cat_dim) def _assign_sum_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for sum op. Args: node (node) node_idx (int) """ nodes_in = flat_list(node.args[0]) self._add_dim(node_idx, 0) self._assign_indice_as_input(node, node_idx, input_node=nodes_in[0]) for n in nodes_in[1:]: self._inherit_more_indice_from_node_with_exclude(n, node) cat_dim = node.kwargs["dim"] self._del_dim(node_idx, cat_dim) def _assign_embedding_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for embedding op. Args: node (node) node_idx (int) """ self._del_dim(node_idx, -1) self._assign_indice_as_input(node, node_idx) self._add_dim(node_idx, -1) def _assign_getitem_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for getitem. getitem can act like slice sometimes Args: node (node) node_idx (int) """ node_args = flat_list(node.args[1:]) # deal with split if get_node_name(node.args[0]) == "split": self._assign_indice_as_input(node, node_idx) self._del_dim(node_idx, node.args[0].kwargs["dim"]) self._add_dim(node_idx, node.args[0].kwargs["dim"]) return # skip non tensor if get_node_shape(node) is None: return # find if slice flag = False for node_arg in node_args: node_arg_str = str(node_arg) if any(i == node_arg_str for i in ["None", "Ellipsis"]): flag = True break if "slice" in node_arg_str: flag = True break if flag == False: return # node args should be like [Ellipsis, slice(start, step, end), None] node_shape = get_node_shape(node) origin_idx_count = 0 new_idx_count = 0 new_dim_num = sum([1 if str(i) == "None" else 0 for i in node_args]) for _ in range(new_dim_num): self._del_dim(node_idx, 0) delete_dim_num = sum([1 if str(i) == "0" else 0 for i in node_args]) for _ in range(delete_dim_num): self._add_dim(node_idx, 0) self._assign_indice_as_input(node, node_idx) for _, node_arg in enumerate(node_args): node_arg_str = str(node_arg) # Ellipsis means [..., ] if "Ellipsis" == node_arg_str: shape_gap = len(node_shape) - len(node_args) + 1 origin_idx_count += shape_gap new_idx_count += shape_gap # slice(None, None, None) means all indexes elif "slice" in node_arg_str: if "slice(None, None, None)" != node_arg_str: self._del_dim(node_idx, new_idx_count) self._add_dim(node_idx, new_idx_count) origin_idx_count += 1 new_idx_count += 1 # None means a new dim elif "None" == node_arg_str: self._add_dim(node_idx, new_idx_count) new_idx_count += 1 elif "0" == node_arg_str: self._del_dim(node_idx, new_idx_count) origin_idx_count += 1 else: raise NotImplementedError() def _assign_view_reshape_indice(self, node: Node, node_idx: int) -> None: """ Assign indice for view and reshape op. 1. get origin shape and target shape by meta info. 2. compute the real value of -1 in target shape. 3. determine changed dim, and assgin indice for generated dim. 4. log changed dim and generated dim for restore 5. inherit computation. 6. look into view list to see whether the view is associated with other, if so assgin equal dim according to previous view. Args: node (node) node_idx (int) """ # get data, turn into number origin_node = node.args[0] origin_shape = origin_node.meta["tensor_meta"].shape target_shape = [] unflated_args = flat_list(node.args) for i in range(1, len(unflated_args)): if isinstance(unflated_args[i], int): target_shape.append(unflated_args[i]) else: target_shape.extend(unflated_args[i].meta["fwd_out"]) # compute the value of -1 if -1 in target_shape: origin_product = 1 for i in origin_shape: origin_product *= i target_product = -1 for i in target_shape: target_product *= i shape_idx = target_shape.index(-1) target_shape[shape_idx] = origin_product // target_product # determine changed dim len_diff = len(origin_shape) - len(target_shape) if len_diff == 1: # dim merge dim_equal = [i == j for i, j in zip(origin_shape[:-1], target_shape)] dim_to = [dim_equal.index(False)] dim_from = [dim_equal.index(False), dim_equal.index(False) + 1] self._add_dim(node_idx, -1) elif len_diff == -1: # dim expand dim_equal = [i == j for i, j in zip(origin_shape, target_shape[:-1])] dim_from = [dim_equal.index(False)] dim_to = [dim_equal.index(False), dim_equal.index(False) + 1] self._del_dim(node_idx, -1) elif len_diff == 0: # dim equal dim_equal = [i == j for i, j in zip(origin_shape, target_shape[:-1])] dim_from = [] dim_to = [] else: raise NotImplementedError("shape" + str(origin_shape) + "and" + str(target_shape) + "view not implemented") # get new indice origin_trace = self._find_indice_trace_from_node(origin_node) self._assign_indice_as_input(node, node_idx, origin_node) idx_from = [origin_trace[i] for i in dim_from] dim_from.reverse() for i in dim_from: self._del_dim(node_idx, i) for i in dim_to: self._add_dim(node_idx, i) dim_from.reverse() # search view list for view_node, view_dict in self.indice_view_list.items(): if (view_dict["idx_to"] == idx_from and view_dict["dim_to"] == dim_from and view_dict["dim_from"] == dim_to): # inheirt indice from current node if len_diff == 1: if origin_shape[dim_from[0]] == 1: self._inherit_indice(origin_node, dim_from[1], node, dim_to[0], init=False) elif origin_shape[dim_from[1]] == 1: self._inherit_indice(origin_node, dim_from[0], node, dim_to[0], init=False) elif len_diff == -1: if target_shape[dim_to[0]] == 1: self._inherit_indice(origin_node, dim_from[0], node, dim_to[1], init=False) elif target_shape[dim_to[1]] == 1: self._inherit_indice(origin_node, dim_from[0], node, dim_to[0], init=False) # inherid indice from input node of last view for dim_to_i in dim_to: self._inherit_indice(view_node.args[0], dim_to_i, node, dim_to_i, init=False) # log view, not used now view_dict = { "idx_from": [origin_trace[i] for i in dim_from], "dim_from": dim_from, "idx_to": [self.indice_trace_list[node_idx]["indice"][i] for i in dim_to], "dim_to": dim_to, } self.indice_view_list[node] = view_dict def _clear_trace(self, node_idx: int) -> None: """ clear too far trace to speed up computation """ trace_range = None for i in range(len(self.trace_range)): if self.trace_range[i][1] == node_idx: trace_range = (self.trace_range[i][0], self.trace_range[i][1]) break if self.trace_range[i][1] > node_idx: break if trace_range is None: return active_nodes = self.active_node_list[trace_range[0]:trace_range[1] + 1] active_nodes = set(flat_list(active_nodes)) active_nodes = [self.node_mgr.find_node_idx_by_name(i) for i in active_nodes] for i in range(trace_range[0], trace_range[1] + 1): trace = self.indice_trace_list[i] # clear compute for dim_compute in trace["compute"]: for i in range(len(dim_compute) - 1, -1, -1): if (dim_compute[i] < trace_range[0] and dim_compute[i] not in active_nodes): dim_compute.pop(i) continue # clear source for dim_source in trace["source"]: for k in list(dim_source.keys()): if k < trace_range[0] and k not in active_nodes: dim_source.pop(k) def trace_indice(self) -> None: for idx, node in enumerate(self.node_mgr.get_node_list()): node_name = get_node_name(node) if node.op == "placeholder": self._assign_all_indice(node, idx) elif node.op == "call_method": if "transpose" == node_name: self._assign_transpose_indice(node, idx) elif "permute" == node_name: self._assign_permute_indice(node, idx) elif "view" == node_name or "reshape" == node_name: self._assign_view_reshape_indice(node, idx) elif "unsqueeze" == node_name: self._assign_unsqueeze_indice(node, idx) elif "split" == node_name: self._assign_split_indice(node, idx) elif any(i == node_name for i in ["to", "contiguous", "clone", "type", "float"]): self._assgin_no_change_indice(node, idx) elif "new_ones" == node_name: self._assign_all_indice(node, idx) elif any(i == node_name for i in ["size"]): continue else: raise NotImplementedError(node_name, "method not implemented yet!") elif node.op == "call_function": if "linear" == node_name: self._assign_linear_indice(node, idx) elif "cat" == node_name: self._assign_cat_indice(node, idx) elif any(n == node_name for n in ["matmul", "bmm"]): self._assign_matmul_indice(node, idx) elif "softmax" == node_name: self._assign_softmax_indice(node, idx) elif any(n == node_name for n in [ "mul", "add", "sigmoid", "relu", "sub", "truediv", "pow", "dropout", "where", "tanh", "exp", "sin", "cos" ]): self._assign_elementwise_indice(node, idx) elif "einsum" == node_name: self._assign_einsum_indice(node, idx) elif "sum" == node_name: self._assign_sum_indice(node, idx) elif "layer_norm" == node_name: self._assign_layernorm_indice(node, idx) elif "getitem" == node_name: self._assign_getitem_indice(node, idx) elif "addmm" == node_name: self._assign_addmm_indice(node, idx) elif "baddbmm" == node_name: self._assign_baddbmm_indice(node, idx) elif "interpolate" == node_name: self._assign_interpolate_indice(node, idx) elif any(i == node_name for i in ["arange", "ones", "ones_like", "tensor", "empty"]): self._assign_all_indice(node, idx) elif any(i == node_name for i in ["getattr", "eq", "_assert_is_none", "_assert", "finfo"]): continue else: raise NotImplementedError(node_name, "function not implemented yet!") elif node.op == "call_module": node_name = get_module_node_name(node) if "layernorm" == node_name: self._assign_layernorm_indice(node, idx) elif "groupnorm" == node_name: self._assign_groupnorm_indice(node, idx) elif "embedding" == node_name: self._assign_embedding_indice(node, idx) elif "linear" == node_name: self._assign_linear_indice(node, idx) elif "conv2d" == node_name: self._assign_conv2d_indice(node, idx) elif any(n == node_name for n in ["sigmoid", "dropout", "relu", "silu"]): self._assign_elementwise_indice(node, idx) else: raise NotImplementedError(node_name, "module not implemented yet!") elif node.op == "get_attr": self._assign_all_indice(node, idx) # get param elif node.op == "output": continue else: raise NotImplementedError(node.op, "op not implemented yet!") # limit trace range self._clear_trace(idx)
from typing import Any, Dict, Iterable, List, Tuple import torch import colossalai from colossalai.fx._compatibility import is_compatible_with_meta from colossalai.fx.codegen.activation_checkpoint_codegen import CODEGEN_AVAILABLE AUTOCHUNK_AVAILABLE = CODEGEN_AVAILABLE and is_compatible_with_meta() if AUTOCHUNK_AVAILABLE: from torch.fx.graph import CodeGen, PythonCode, _custom_builtins, _CustomBuiltin, _format_target, _is_from_torch, _Namespace, _origin_type_map, inplace_methods, magic_methods from torch.fx.node import Argument, Node, _get_qualified_name, _type_repr, map_arg from .search_chunk import SearchChunk from .utils import delete_free_var_from_last_use, get_logger, get_node_name, get_node_shape def _gen_chunk_slice_dim(chunk_dim: int, chunk_indice_name: str, shape: List) -> str: """ Generate chunk slice string, eg. [:, :, chunk_idx_name:chunk_idx_name + chunk_size, :] Args: chunk_dim (int) chunk_indice_name (str): chunk indice name shape (List): node shape Returns: new_shape (str): return slice """ new_shape = "[" for idx, _ in enumerate(shape): if idx == chunk_dim: new_shape += "%s:%s + chunk_size" % (chunk_indice_name, chunk_indice_name) else: new_shape += ":" new_shape += ", " new_shape = new_shape[:-2] + "]" return new_shape def _gen_loop_start(chunk_input: List[Node], chunk_output: List[Node], chunk_ouput_dim: int, chunk_size=2) -> str: """ Generate chunk loop start eg. chunk_result = torch.empty([100, 100], dtype=input_node.dtype, device=input_node.device) chunk_size = 32 for chunk_idx in range(0, 100, 32): ...... Args: chunk_input (List[Node]): chunk input node chunk_output (Node): chunk output node chunk_ouput_dim (int): chunk output node chunk dim chunk_size (int): chunk size. Defaults to 2. Returns: context (str): generated str """ input_node = chunk_input[0] context = "" for i in range(len(chunk_output)): shape_str = str(list(get_node_shape(chunk_output[i]))) if get_node_name(chunk_output[i]) == "split": tensor_str = "torch.empty(%s, dtype=%s.dtype, device=%s.device), " % (shape_str, input_node.name, input_node.name) tensor_str = tensor_str * len(chunk_output[i].meta['tensor_meta']) tensor_str = "[" + tensor_str[:-2] + "]" context += "%s = %s; " % (chunk_output[i].name, tensor_str) else: context += "%s = torch.empty(%s, dtype=%s.dtype, device=%s.device); " % (chunk_output[i].name, shape_str, input_node.name, input_node.name) out_shape = get_node_shape(chunk_output[0]) chunk_shape = out_shape[chunk_ouput_dim[0]] context += "chunk_size = %d\nfor chunk_idx in range(0, %d, chunk_size):\n" % (chunk_size, chunk_shape) return context def _gen_loop_end(chunk_inputs: List[Node], chunk_non_compute_inputs: List[Node], node_list: List[Node], chunk_outputs_idx: int, chunk_outputs_non_tensor: List[Node], search_chunk: SearchChunk) -> str: """ Generate chunk loop end eg. chunk_result[chunk_idx:chunk_idx + chunk_size] = output_node output_node = chunk_result; xx = None; xx = None Args: chunk_inputs (List[Node]): chunk input node chunk_non_compute_inputs (List[Node]): input node without chunk chunk_outputs (Node): chunk output node chunk_outputs_dim (int): chunk output node chunk dim node_list (List) Returns: context (str): generated str """ context = "chunk_size = None" # determine if its the last use for chunk input for chunk_input in chunk_inputs + chunk_non_compute_inputs: if all([search_chunk.node_mgr.find_node_idx(user) <= chunk_outputs_idx for user in chunk_input.users.keys()]): context += "; %s = None" % chunk_input.name for chunk_output_non_tensor, chunk_output_non_tensor_val in chunk_outputs_non_tensor.items(): context += "; %s = %s" % (chunk_output_non_tensor.name, chunk_output_non_tensor_val) context += "\n" return context def _replace_name(context: str, name_from: str, name_to: str) -> str: """ replace node name """ patterns = [(" ", " "), (" ", "."), (" ", ","), ("(", ")"), ("(", ","), (" ", ")"), (" ", ""), ("", " ")] for p in patterns: source = p[0] + name_from + p[1] target = p[0] + name_to + p[1] if source in context: context = context.replace(source, target) break return context def _replace_reshape_size(context: str, node_name: str, reshape_size_dict: Dict) -> str: """ replace reshape size, some may have changed due to chunk """ if node_name not in reshape_size_dict: return context context = context.replace(reshape_size_dict[node_name][0], reshape_size_dict[node_name][1]) return context def _replace_new_tensor_like_shape( search_chunk: SearchChunk, chunk_infos: List[Dict], region_idx: int, node_idx: int, node: Node, body: List[str], ) -> List[str]: """ add chunk slice for new tensor op such as ones like """ if get_node_name(node) in ["ones_like", "zeros_like", "empty_like"]: meta_node = search_chunk.node_mgr.get_node_by_idx(node_idx) chunk_dim = chunk_infos[region_idx]["node_chunk_dim"][meta_node]["chunk_dim"] if get_node_shape(meta_node)[chunk_dim] != 1: source_node = meta_node.args[0].args[0] if (source_node not in chunk_infos[region_idx]["node_chunk_dim"] or chunk_infos[region_idx]["node_chunk_dim"][source_node]["chunk_dim"] is None): chunk_slice = _gen_chunk_slice_dim(chunk_dim, "chunk_idx", get_node_shape(node)) body[-1] = _replace_name(body[-1], node.args[0].name, node.args[0].name + chunk_slice) return body def _replace_new_tensor_shape( search_chunk: SearchChunk, chunk_infos: List[Dict], region_idx: int, node_idx: int, node: Node, body: List[str], ) -> List[str]: """ add chunk slice for new tensor op such as ones """ if get_node_name(node) in ["ones", "zeros", "empty"]: meta_node = search_chunk.node_mgr.get_node_by_idx(node_idx) chunk_dim = chunk_infos[region_idx]["node_chunk_dim"][meta_node]["chunk_dim"] if chunk_dim is None: return if get_node_shape(meta_node)[chunk_dim] == 1: return origin_shape = str(node.args) new_shape = list(node.args) new_shape[chunk_dim] = "min(chunk_size, %d - chunk_idx)" % get_node_shape(meta_node)[chunk_dim] new_shape = str(new_shape) new_shape = new_shape.replace("'", "") body[-1] = _replace_name(body[-1], origin_shape[1:-1], new_shape[1:-1]) return body def _add_node_slice( chunk_nodes: List[Node], region_idx: int, chunk_nodes_dim: Dict, node_idx: int, body: List[str], node: Node, ) -> List[str]: """ add chunk slice for input nodes """ for chunk_node_idx, chunk_node in enumerate(chunk_nodes[region_idx]): # inputs node if isinstance(chunk_nodes_dim[region_idx][chunk_node_idx], dict): for idx, dim in chunk_nodes_dim[region_idx][chunk_node_idx].items(): if idx == node_idx: chunk_slice = _gen_chunk_slice_dim(dim[0], "chunk_idx", get_node_shape(chunk_node)) body[-1] = _replace_name(body[-1], chunk_node.name, chunk_node.name + chunk_slice) # outputs node else: if chunk_node.name == node.name or (chunk_node.name in [i.name for i in node.all_input_nodes]): chunk_slice = _gen_chunk_slice_dim(chunk_nodes_dim[region_idx][chunk_node_idx], "chunk_idx", get_node_shape(chunk_node)) if get_node_name(chunk_node) == "split": split_chunk_slice = "" for i in range(len(chunk_node.meta['tensor_meta'])): split_chunk_slice += "%s[%d]%s, " % (chunk_node.name, i, chunk_slice) split_chunk_slice = split_chunk_slice[:-2] body[-1] = _replace_name(body[-1], chunk_node.name, split_chunk_slice) else: body[-1] = _replace_name(body[-1], chunk_node.name, chunk_node.name + chunk_slice) return body def emit_code_with_chunk( body: List[str], nodes: Iterable[Node], emit_node_func, delete_unused_value_func, search_chunk: SearchChunk, chunk_infos: List, ): """ Emit code with chunk according to chunk_infos. It will generate a for loop in chunk regions, and replace inputs and outputs of regions with chunked variables. Args: body: forward code nodes: graph.nodes emit_node_func: function to emit node delete_unused_value_func: function to remove the unused value search_chunk: the class to search all chunks chunk_infos: store all information about all chunks. """ node_list = list(nodes) # chunk region chunk_starts = [i["region"][0] for i in chunk_infos] chunk_ends = [i["region"][1] for i in chunk_infos] # chunk inputs chunk_inputs = [i["inputs"] for i in chunk_infos] # input with chunk chunk_inputs_non_chunk = [i["inputs_non_chunk"] for i in chunk_infos] # input without chunk chunk_inputs_dim = [i["inputs_dim"] for i in chunk_infos] # input chunk dim chunk_inputs_names = [j.name for i in chunk_inputs for j in i] + [j.name for i in chunk_inputs_non_chunk for j in i] # chunk outputs chunk_outputs = [i["outputs"] for i in chunk_infos] chunk_outputs_non_tensor = [i["outputs_non_tensor"] for i in chunk_infos] chunk_outputs_dim = [i["outputs_dim"] for i in chunk_infos] node_list = search_chunk.reorder_graph.reorder_node_list(node_list) node_idx = 0 region_idx = 0 within_chunk_region = False while node_idx < len(node_list): node = node_list[node_idx] # if is chunk start, generate for loop start if node_idx in chunk_starts: within_chunk_region = True region_idx = chunk_starts.index(node_idx) body.append( _gen_loop_start( chunk_inputs[region_idx], chunk_outputs[region_idx], chunk_outputs_dim[region_idx], chunk_infos[region_idx]["chunk_size"], )) if within_chunk_region: emit_node_func(node, body) # replace input var with chunk var body = _add_node_slice(chunk_inputs, region_idx, chunk_inputs_dim, node_idx, body, node) # replace output var with chunk var body = _add_node_slice(chunk_outputs, region_idx, chunk_outputs_dim, node_idx, body, node) # new tensor like body = _replace_new_tensor_like_shape(search_chunk, chunk_infos, region_idx, node_idx, node, body) # new tensor body = _replace_new_tensor_shape(search_chunk, chunk_infos, region_idx, node_idx, node, body) # reassgin reshape size body[-1] = _replace_reshape_size(body[-1], node.name, chunk_infos[region_idx]["reshape_size"]) body[-1] = " " + body[-1] delete_unused_value_func(node, body, chunk_inputs_names) else: emit_node_func(node, body) if node_idx not in chunk_inputs: delete_unused_value_func(node, body, chunk_inputs_names) # generate chunk region end if node_idx in chunk_ends: body.append( _gen_loop_end(chunk_inputs[region_idx], chunk_inputs_non_chunk[region_idx], node_list, chunk_ends[region_idx], chunk_outputs_non_tensor[region_idx], search_chunk)) within_chunk_region = False node_idx += 1 if AUTOCHUNK_AVAILABLE: class AutoChunkCodeGen(CodeGen): def __init__(self, meta_graph, max_memory: int = None, print_mem: bool = False, print_progress: bool = False) -> None: super().__init__() # find the chunk regions self.search_chunk = SearchChunk(meta_graph, max_memory, print_mem, print_progress) self.chunk_infos = self.search_chunk.search_region() if print_progress: get_logger().info("AutoChunk start codegen") def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace) -> PythonCode: free_vars: List[str] = [] body: List[str] = [] globals_: Dict[str, Any] = {} wrapped_fns: Dict[str, None] = {} # Wrap string in list to pass by reference maybe_return_annotation: List[str] = [""] def add_global(name_hint: str, obj: Any): """Add an obj to be tracked as a global. We call this for names that reference objects external to the Graph, like functions or types. Returns: the global name that should be used to reference 'obj' in generated source. """ if (_is_from_torch(obj) and obj != torch.device): # to support registering torch.device # HACK: workaround for how torch custom ops are registered. We # can't import them like normal modules so they must retain their # fully qualified name. return _get_qualified_name(obj) # normalize the name hint to get a proper identifier global_name = namespace.create_name(name_hint, obj) if global_name in globals_: assert globals_[global_name] is obj return global_name globals_[global_name] = obj return global_name # set _custom_builtins here so that we needn't import colossalai in forward _custom_builtins["colossalai"] = _CustomBuiltin("import colossalai", colossalai) # Pre-fill the globals table with registered builtins. for name, (_, obj) in _custom_builtins.items(): add_global(name, obj) def type_repr(o: Any): if o == (): # Empty tuple is used for empty tuple type annotation Tuple[()] return "()" typename = _type_repr(o) if hasattr(o, "__origin__"): # This is a generic type, e.g. typing.List[torch.Tensor] origin_type = _origin_type_map.get(o.__origin__, o.__origin__) origin_typename = add_global(_type_repr(origin_type), origin_type) if hasattr(o, "__args__"): # Assign global names for each of the inner type variables. args = [type_repr(arg) for arg in o.__args__] if len(args) == 0: # Bare type, such as `typing.Tuple` with no subscript # This code-path used in Python < 3.9 return origin_typename return f'{origin_typename}[{",".join(args)}]' else: # Bare type, such as `typing.Tuple` with no subscript # This code-path used in Python 3.9+ return origin_typename # Common case: this is a regular module name like 'foo.bar.baz' return add_global(typename, o) def _format_args(args: Tuple[Argument, ...], kwargs: Dict[str, Argument]) -> str: def _get_repr(arg): # Handle NamedTuples (if it has `_fields`) via add_global. if isinstance(arg, tuple) and hasattr(arg, "_fields"): qualified_name = _get_qualified_name(type(arg)) global_name = add_global(qualified_name, type(arg)) return f"{global_name}{repr(tuple(arg))}" return repr(arg) args_s = ", ".join(_get_repr(a) for a in args) kwargs_s = ", ".join(f"{k} = {_get_repr(v)}" for k, v in kwargs.items()) if args_s and kwargs_s: return f"{args_s}, {kwargs_s}" return args_s or kwargs_s # Run through reverse nodes and record the first instance of a use # of a given node. This represents the *last* use of the node in the # execution order of the program, which we will use to free unused # values node_to_last_use: Dict[Node, Node] = {} user_to_last_uses: Dict[Node, List[Node]] = {} def register_last_uses(n: Node, user: Node): if n not in node_to_last_use: node_to_last_use[n] = user user_to_last_uses.setdefault(user, []).append(n) for node in reversed(nodes): map_arg(node.args, lambda n: register_last_uses(n, node)) map_arg(node.kwargs, lambda n: register_last_uses(n, node)) delete_free_var_from_last_use(user_to_last_uses) # NOTE: we add a variable to distinguish body and ckpt_func def delete_unused_values(user: Node, body, to_keep=[]): """ Delete values after their last use. This ensures that values that are not used in the remainder of the code are freed and the memory usage of the code is optimal. """ if user.op == "placeholder": return if user.op == "output": body.append("\n") return nodes_to_delete = user_to_last_uses.get(user, []) nodes_to_delete = [i for i in nodes_to_delete if i.name not in to_keep] if len(nodes_to_delete): to_delete_str = " = ".join([repr(n) for n in nodes_to_delete] + ["None"]) body.append(f"; {to_delete_str}\n") else: body.append("\n") # NOTE: we add a variable to distinguish body and ckpt_func def emit_node(node: Node, body): maybe_type_annotation = ("" if node.type is None else f" : {type_repr(node.type)}") if node.op == "placeholder": assert isinstance(node.target, str) maybe_default_arg = ("" if not node.args else f" = {repr(node.args[0])}") free_vars.append(f"{node.target}{maybe_type_annotation}{maybe_default_arg}") raw_name = node.target.replace("*", "") if raw_name != repr(node): body.append(f"{repr(node)} = {raw_name}\n") return elif node.op == "call_method": assert isinstance(node.target, str) body.append( f"{repr(node)}{maybe_type_annotation} = {_format_target(repr(node.args[0]), node.target)}" f"({_format_args(node.args[1:], node.kwargs)})") return elif node.op == "call_function": assert callable(node.target) # pretty print operators if (node.target.__module__ == "_operator" and node.target.__name__ in magic_methods): assert isinstance(node.args, tuple) body.append(f"{repr(node)}{maybe_type_annotation} = " f"{magic_methods[node.target.__name__].format(*(repr(a) for a in node.args))}") return # pretty print inplace operators; required for jit.script to work properly # not currently supported in normal FX graphs, but generated by torchdynamo if (node.target.__module__ == "_operator" and node.target.__name__ in inplace_methods): body.append(f"{inplace_methods[node.target.__name__].format(*(repr(a) for a in node.args))}; " f"{repr(node)}{maybe_type_annotation} = {repr(node.args[0])}") return qualified_name = _get_qualified_name(node.target) global_name = add_global(qualified_name, node.target) # special case for getattr: node.args could be 2-argument or 3-argument # 2-argument: attribute access; 3-argument: fall through to attrib function call with default value if (global_name == "getattr" and isinstance(node.args, tuple) and isinstance(node.args[1], str) and node.args[1].isidentifier() and len(node.args) == 2): body.append( f"{repr(node)}{maybe_type_annotation} = {_format_target(repr(node.args[0]), node.args[1])}") return body.append( f"{repr(node)}{maybe_type_annotation} = {global_name}({_format_args(node.args, node.kwargs)})") if node.meta.get("is_wrapped", False): wrapped_fns.setdefault(global_name) return elif node.op == "call_module": assert isinstance(node.target, str) body.append(f"{repr(node)}{maybe_type_annotation} = " f"{_format_target(root_module, node.target)}({_format_args(node.args, node.kwargs)})") return elif node.op == "get_attr": assert isinstance(node.target, str) body.append(f"{repr(node)}{maybe_type_annotation} = {_format_target(root_module, node.target)}") return elif node.op == "output": if node.type is not None: maybe_return_annotation[0] = f" -> {type_repr(node.type)}" body.append(self.generate_output(node.args[0])) return raise NotImplementedError(f"node: {node.op} {node.target}") # Modified for activation checkpointing ckpt_func = [] # if any node has a list of labels for activation_checkpoint, we # will use nested type of activation checkpoint codegen emit_code_with_chunk( body, nodes, emit_node, delete_unused_values, self.search_chunk, self.chunk_infos, ) if len(body) == 0: # If the Graph has no non-placeholder nodes, no lines for the body # have been emitted. To continue to have valid Python code, emit a # single pass statement body.append("pass\n") if len(wrapped_fns) > 0: wrap_name = add_global("wrap", torch.fx.wrap) wrap_stmts = "\n".join([f'{wrap_name}("{name}")' for name in wrapped_fns]) else: wrap_stmts = "" if self._body_transformer: body = self._body_transformer(body) for name, value in self.additional_globals(): add_global(name, value) # as we need colossalai.utils.checkpoint, we need to import colossalai # in forward function prologue = self.gen_fn_def(free_vars, maybe_return_annotation[0]) prologue = "".join(ckpt_func) + prologue prologue = prologue code = "".join(body) code = "\n".join(" " + line for line in code.split("\n")) fn_code = f""" {wrap_stmts} {prologue} {code}""" # print(fn_code) return PythonCode(fn_code, globals_)
from .cli import cli __all__ = ['cli']
import click from .benchmark import benchmark from .check import check from .launcher import run class Arguments(): def __init__(self, arg_dict): for k, v in arg_dict.items(): self.__dict__[k] = v @click.group() def cli(): pass cli.add_command(run) cli.add_command(check) cli.add_command(benchmark) if __name__ == '__main__': cli()
from functools import partial from typing import Dict, List import click import torch.multiprocessing as mp import colossalai from colossalai.cli.benchmark.utils import find_all_configs, get_batch_data, profile_model from colossalai.context import Config from colossalai.context.random import reset_seeds from colossalai.core import global_context as gpc from colossalai.logging import disable_existing_loggers, get_dist_logger from colossalai.utils import MultiTimer, free_port from .models import MLP def run_benchmark(args: Config) -> None: """ Run benchmarking with torch.multiprocessing. """ # sanity checks if args.gpus is None: click.echo("Error: --num_gpus is not given") exit() if args.gpus <= 1: click.echo("Warning: tensor parallel will be activated with at least 2 devices.") click.echo("=== Benchmarking Parameters ===") for k, v in args.items(): click.echo(f'{k}: {v}') click.echo('') config_list = find_all_configs(args.gpus) avail_ports = [free_port() for _ in range(len(config_list))] run_func = partial(run_dist_profiling, world_size=args.gpus, port_list=avail_ports, config_list=config_list, hyperparams=args) mp.spawn(run_func, nprocs=args.gpus) def run_dist_profiling(rank: int, world_size: int, port_list: List[int], config_list: List[Dict], hyperparams: Config) -> None: """ A function executed for profiling, this function should be spawn by torch.multiprocessing. Args: rank (int): rank of the process world_size (int): the number of processes port_list (List[int]): a list of free ports for initializing distributed networks config_list (List[Dict]): a list of configuration hyperparams (Config): the hyperparameters given by the user """ # disable logging for clean output disable_existing_loggers() logger = get_dist_logger() logger.set_level('WARNING') for config, port in zip(config_list, port_list): colossalai.launch(config=config, rank=rank, world_size=world_size, host='localhost', port=port, backend='nccl') timer = MultiTimer() # 1D parallel should be skipped if in_features or out_features is not able to be divided exactly by 1D parallel size. if config.parallel.tensor.mode == '1d' and hyperparams.dimension % config.parallel.tensor.size != 0: click.echo( "1D parallel will be skipped because in_features or out_features is not able to be divided exactly by 1D parallel size." ) continue if hyperparams.model == 'mlp': model = MLP(dim=hyperparams.dimension, layers=hyperparams.layers) else: if gpc.get_global_rank() == 0: click.echo("Error: Invalid argument for --model") exit() data_func = partial(get_batch_data, dim=hyperparams.dimension, batch_size=hyperparams.batch_size, seq_length=hyperparams.seq_len, mode=config.parallel.tensor.mode) fwd_time, bwd_time, max_allocated, max_cached = profile_model(model=model, warmup_steps=hyperparams.warmup_steps, profile_steps=hyperparams.profile_steps, data_func=data_func, timer=timer) gpc.destroy() reset_seeds() if gpc.get_global_rank() == 0: config_str = ', '.join([f'{k}: {v}' for k, v in config.parallel.tensor.items()]) click.echo(f"=== {config_str} ===") click.echo(f"Average forward time: {fwd_time}") click.echo(f"Average backward time: {bwd_time}") click.echo(f"Max allocated GPU memory: {max_allocated}") click.echo(f"Max cached GPU memory: {max_cached}\n")
import torch import colossalai.nn as col_nn class MLP(torch.nn.Module): def __init__(self, dim: int, layers: int): super().__init__() self.layers = torch.nn.ModuleList() for _ in range(layers): self.layers.append(col_nn.Linear(dim, dim)) def forward(self, x): for layer in self.layers: x = layer(x) return x
import click from colossalai.context import Config from .benchmark import run_benchmark from .utils import * __all__ = ['benchmark'] @click.command() @click.option("-g", "--gpus", type=int, default=None, help="Total number of devices to use.") @click.option("-b", "--batch_size", type=int, default=8, help="Batch size of the input tensor.") @click.option("-s", "--seq_len", type=int, default=512, help="Sequence length of the input tensor.") @click.option("-d", "--dimension", type=int, default=1024, help="Hidden dimension of the input tensor.") @click.option("-w", "--warmup_steps", type=int, default=10, help="The number of warmup steps.") @click.option("-p", "--profile_steps", type=int, default=50, help="The number of profiling steps.") @click.option("-l", "--layers", type=int, default=2) @click.option("-m", "--model", type=click.Choice(['mlp'], case_sensitive=False), default='mlp', help="Select the model to benchmark, currently only supports MLP") def benchmark(gpus: int, batch_size: int, seq_len: int, dimension: int, warmup_steps: int, profile_steps: int, layers: int, model: str): args_dict = locals() args = Config(args_dict) run_benchmark(args)
import math import time import torch from colossalai.utils import MultiTimer from colossalai.context import ParallelMode, Config from typing import List, Dict, Tuple, Callable def get_time_stamp() -> int: """ Return the time stamp for profiling. Returns: time_stamp (int): the time given by time.time() """ torch.cuda.synchronize() time_stamp = time.time() return time_stamp def get_memory_states() -> Tuple[float]: """ Return the memory statistics. Returns: max_allocated (float): the allocated CUDA memory max_cached (float): the cached CUDA memory """ max_allocated = torch.cuda.max_memory_allocated() / (1024**3) max_cached = torch.cuda.max_memory_reserved() / (1024**3) torch.cuda.reset_peak_memory_stats() torch.cuda.empty_cache() return max_allocated, max_cached def find_all_configs(device_cnt: int) -> List[Dict]: """ Find all possible configurations for tensor parallelism Args: device_cnt (int): the number of devices Returns: config_list (List[Dict]): a list of configurations """ def _is_square(num): # 2D parallel should be implemented with at least 2 devices. if num <= 1: return False return math.floor(math.sqrt(num))**2 == num def _is_cube(num): # 3D parallel should be implemented with at least 2 devices. if num <= 1: return False return math.floor(num**(1. / 3.))**3 == num config_list = [] # add non-parallel config config = dict(parallel=dict(tensor=dict(size=device_cnt, mode=None))) config_list.append(config) # add 1D config config = dict(parallel=dict(tensor=dict(size=device_cnt, mode='1d'))) config_list.append(config) # add 2D config only if device_cnt is a square if _is_square(device_cnt): config = dict(parallel=dict(tensor=dict(size=device_cnt, mode='2d'))) config_list.append(config) # check for 2.5D # iterate over depth for depth in range(1, device_cnt): if device_cnt % depth == 0 and _is_square(device_cnt // depth): config = dict(parallel=dict(tensor=dict(size=device_cnt, mode='2.5d', depth=depth))) config_list.append(config) # check for 3D if device_cnt is a cube if _is_cube(device_cnt): config = dict(parallel=dict(tensor=dict(size=device_cnt, mode='3d'))) config_list.append(config) config_list = [Config(cfg) for cfg in config_list] return config_list def profile_model(model: torch.nn.Module, warmup_steps: int, profile_steps: int, data_func: Callable, timer: MultiTimer) -> Tuple[float]: """ Profile the forward and backward of a model Args: model (torch.nn.Module): a PyTorch model warmup_steps (int): the number of steps for warmup profile_steps (int): the number of steps for profiling data_func (Callable): a function to generate random data timer (colossalai.utils.Multitimer): a timer instance for time recording Returns: fwd_time (float): the average forward time taken by forward pass in second bwd_time (float): the average backward time taken by forward pass in second max_allocated (float): the maximum GPU memory allocated in GB max_cached (float): the maximum GPU memory cached in GB """ def _run_step(data): timer.start('forward') out = model(data) timer.stop('forward', keep_in_history=True) timer.start('backward') out.mean().backward() timer.stop('backward', keep_in_history=True) data_list = [data_func() for _ in range(warmup_steps)] for data in data_list: _run_step(data) timer.reset('forward') timer.reset('backward') for _ in range(profile_steps): data = data_func() _run_step(data) max_allocated, max_cached = get_memory_states() fwd_time = timer.get_timer('forward').get_history_mean() bwd_time = timer.get_timer('backward').get_history_mean() return fwd_time, bwd_time, max_allocated, max_cached def get_batch_data(dim: int, batch_size: int, seq_length: int, mode: ParallelMode) -> torch.Tensor: """ Return a random data of shape (batch_size, seq_length, dim) for profiling. Args: dim (int): hidden size batch_size (int): the number of data samples seq_length (int): the number of tokens mode (ParallelMode): Colossal-AI ParallelMode enum Returns: data (torch.Tensor): random data """ if mode in ['2d', '2.5d']: batch_size = batch_size // 2 dim = dim // 2 elif mode == '3d': batch_size = batch_size // 4 dim = dim // 2 data = torch.rand(batch_size, seq_length, dim).cuda() return data
import os import sys from typing import List import click import torch from packaging import version from colossalai.context import Config from .hostinfo import HostInfo, HostInfoList from .multinode_runner import MultiNodeRunner # Constants that define our syntax NODE_SEP = ',' def fetch_hostfile(hostfile_path: str, ssh_port: int) -> HostInfoList: """ Parse the hostfile to obtain a list of hosts. A hostfile should look like: worker-0 worker-1 worker-2 ... Args: hostfile_path (str): the path to the hostfile ssh_port (int): the port to connect to the host """ if not os.path.isfile(hostfile_path): click.echo(f"Error: Unable to find the hostfile, no such file: {hostfile_path}") exit() with open(hostfile_path, 'r') as fd: device_pool = HostInfoList() for line in fd.readlines(): line = line.strip() if line == '': # skip empty lines continue # build the HostInfo object hostname = line.strip() hostinfo = HostInfo(hostname=hostname, port=ssh_port) if device_pool.has(hostname): click.echo(f"Error: found duplicate host {hostname} in the hostfile") exit() device_pool.append(hostinfo) return device_pool def parse_device_filter(device_pool: HostInfoList, include_str=None, exclude_str=None) -> HostInfoList: '''Parse an inclusion or exclusion string and filter a hostfile dictionary. Examples: include_str="worker-0,worker-1" will execute jobs only on worker-0 and worker-1. exclude_str="worker-1" will use all available devices except worker-1. Args: device_pool (HostInfoList): a list of HostInfo objects include_str (str): --include option passed by user, default None exclude_str (str): --exclude option passed by user, default None Returns: filtered_hosts (HostInfoList): filtered hosts after inclusion/exclusion ''' # Ensure include/exclude are mutually exclusive if include_str and exclude_str: click.echo("--include and --exclude are mutually exclusive, only one can be used") exit() # no-op if include_str is None and exclude_str is None: return device_pool # Either build from scratch or remove items if include_str: parse_str = include_str filtered_hosts = HostInfoList() elif exclude_str: parse_str = exclude_str filtered_hosts = device_pool # foreach node in the list for node_config in parse_str.split(NODE_SEP): hostname = node_config hostinfo = device_pool.get_hostinfo(hostname) # sanity check hostname if not device_pool.has(hostname): click.echo(f"Error: Hostname '{hostname}' not found in hostfile") exit() if include_str: filtered_hosts.append(hostinfo) elif exclude_str: filtered_hosts.remove(hostname) return filtered_hosts def get_launch_command( master_addr: str, master_port: int, nproc_per_node: int, user_script: str, user_args: List[str], node_rank: int, num_nodes: int, extra_launch_args: str = None, ) -> str: """ Generate a command for distributed training. Args: master_addr (str): the host of the master node master_port (str): the port of the master node nproc_per_node (str): the number of processes to launch on each node user_script (str): the user Python file user_args (str): the arguments for the user script node_rank (int): the unique ID for the node num_nodes (int): the number of nodes to execute jobs Returns: cmd (str): the command the start distributed training """ def _arg_dict_to_list(arg_dict): ret = [] for k, v in arg_dict.items(): if v: ret.append(f'--{k}={v}') else: ret.append(f'--{k}') return ret if extra_launch_args: extra_launch_args_dict = dict() for arg in extra_launch_args.split(','): if '=' in arg: k, v = arg.split('=') extra_launch_args_dict[k] = v else: extra_launch_args_dict[arg] = None extra_launch_args = extra_launch_args_dict else: extra_launch_args = dict() torch_version = version.parse(torch.__version__) assert torch_version.major == 1 if torch_version.minor < 9: cmd = [ sys.executable, "-m", "torch.distributed.launch", f"--nproc_per_node={nproc_per_node}", f"--master_addr={master_addr}", f"--master_port={master_port}", f"--nnodes={num_nodes}", f"--node_rank={node_rank}" ] else: # extra launch args for torch distributed launcher with torch >= 1.9 default_torchrun_rdzv_args = dict(rdzv_backend="c10d", rdzv_endpoint=f"{master_addr}:{master_port}", rdzv_id="colossalai-default-job") # update rdzv arguments for key in default_torchrun_rdzv_args.keys(): if key in extra_launch_args: value = extra_launch_args.pop(key) default_torchrun_rdzv_args[key] = value if torch_version.minor < 10: cmd = [ sys.executable, "-m", "torch.distributed.run", f"--nproc_per_node={nproc_per_node}", f"--nnodes={num_nodes}", f"--node_rank={node_rank}" ] else: cmd = [ "torchrun", f"--nproc_per_node={nproc_per_node}", f"--nnodes={num_nodes}", f"--node_rank={node_rank}" ] cmd += _arg_dict_to_list(default_torchrun_rdzv_args) cmd += _arg_dict_to_list(extra_launch_args) + [user_script] + user_args cmd = ' '.join(cmd) return cmd def launch_multi_processes(args: Config) -> None: """ Launch multiple processes on a single node or multiple nodes. The overall logic can be summarized as the pseudo code below: if hostfile given: hostinfo = parse_hostfile(hostfile) hostinfo = include_or_exclude_hosts(hostinfo) launch_on_multi_nodes(hostinfo) elif hosts given: hostinfo = parse_hosts(hosts) launch_on_multi_nodes(hostinfo) else: launch_on_current_node() Args: args (Config): the arguments taken from command line """ assert isinstance(args, Config) if args.nproc_per_node is None: click.echo("--nproc_per_node did not receive any value") exit() # cannot accept hosts and hostfile at the same time if args.host and args.hostfile: click.echo("Error: hostfile and hosts are mutually exclusive, only one is required") # check if hostfile is given if args.hostfile: device_pool = fetch_hostfile(args.hostfile, ssh_port=args.ssh_port) active_device_pool = parse_device_filter(device_pool, args.include, args.exclude) if args.num_nodes > 0: # only keep the first num_nodes to execute jobs updated_active_device_pool = HostInfoList() for count, hostinfo in enumerate(active_device_pool): if args.num_nodes == count: break updated_active_device_pool.append(hostinfo) active_device_pool = updated_active_device_pool else: active_device_pool = None env = os.environ.copy() # use hosts if hostfile is not given if args.host and active_device_pool is None: active_device_pool = HostInfoList() host_list = args.host.strip().split(NODE_SEP) for hostname in host_list: hostinfo = HostInfo(hostname=hostname, port=args.ssh_port) active_device_pool.append(hostinfo) if not active_device_pool: # run on local node if not hosts or hostfile is given # add local node to host info list active_device_pool = HostInfoList() localhost_info = HostInfo(hostname='127.0.0.1', port=args.ssh_port) active_device_pool.append(localhost_info) # launch distributed processes runner = MultiNodeRunner() curr_path = os.path.abspath('.') # collect current path env env = dict() for k, v in os.environ.items(): # do not support multi-line env var if v and '\n' not in v: env[k] = v # establish remote connection runner.connect(host_info_list=active_device_pool, workdir=curr_path, env=env) # execute distributed launching command for node_id, hostinfo in enumerate(active_device_pool): cmd = get_launch_command(master_addr=args.master_addr, master_port=args.master_port, nproc_per_node=args.nproc_per_node, user_script=args.user_script, user_args=args.user_args, node_rank=node_id, num_nodes=len(active_device_pool), extra_launch_args=args.extra_launch_args) runner.send(hostinfo=hostinfo, cmd=cmd) # start training msg_from_node = runner.recv_from_all() has_error = False # print node status click.echo("\n====== Training on All Nodes =====") for hostname, msg in msg_from_node.items(): click.echo(f"{hostname}: {msg}") # check if a process failed if msg == "failure": has_error = True # stop all nodes runner.stop_all() # receive the stop status msg_from_node = runner.recv_from_all() # printe node status click.echo("\n====== Stopping All Nodes =====") for hostname, msg in msg_from_node.items(): click.echo(f"{hostname}: {msg}") # give the process an exit code # so that it behaves like a normal process if has_error: sys.exit(1) else: sys.exit(0)
from multiprocessing import Pipe, Process from multiprocessing import connection as mp_connection import click import fabric from .hostinfo import HostInfo, HostInfoList def run_on_host(hostinfo: HostInfo, workdir: str, recv_conn: mp_connection.Connection, send_conn: mp_connection.Connection, env: dict) -> None: """ Use fabric connection to execute command on local or remote hosts. Args: hostinfo (HostInfo): host information workdir (str): the directory to execute the command recv_conn (multiprocessing.connection.Connection): receive messages from the master sender send_conn (multiprocessing.connection.Connection): send messages to the master receiver env (dict): a dictionary for environment variables """ fab_conn = fabric.Connection(hostinfo.hostname, port=hostinfo.port) finish = False env_msg = ' '.join([f'{k}=\"{v}\"' for k, v in env.items()]) # keep listening until exit while not finish: # receive cmd cmds = recv_conn.recv() if cmds == 'exit': # exit from the loop finish = True break else: # execute the commands try: # cd to execute directory with fab_conn.cd(workdir): # propagate the runtime environment with fab_conn.prefix(f"export {env_msg}"): if hostinfo.is_local_host: # execute on the local machine fab_conn.local(cmds, hide=False) else: # execute on the remote machine fab_conn.run(cmds, hide=False) send_conn.send('success') except Exception as e: click.echo( f"Error: failed to run {cmds} on {hostinfo.hostname}, is localhost: {hostinfo.is_local_host}, exception: {e}" ) send_conn.send('failure') # shutdown send_conn.send("finish") fab_conn.close() class MultiNodeRunner: """ A runner to execute commands on an array of machines. This runner is inspired by Nezha (https://github.com/zhuzilin/NeZha). """ def __init__(self): self.processes = {} self.master_send_conns = {} self.master_recv_conns = {} def connect(self, host_info_list: HostInfoList, workdir: str, env: dict) -> None: """ Establish connections to a list of hosts Args: host_info_list (HostInfoList): a list of HostInfo objects workdir (str): the directory where command is executed env (dict): environment variables to propagate to hosts """ for hostinfo in host_info_list: master_send_conn, worker_recv_conn = Pipe() master_recv_conn, worker_send_conn = Pipe() p = Process(target=run_on_host, args=(hostinfo, workdir, worker_recv_conn, worker_send_conn, env)) p.start() self.processes[hostinfo.hostname] = p self.master_recv_conns[hostinfo.hostname] = master_recv_conn self.master_send_conns[hostinfo.hostname] = master_send_conn def send(self, hostinfo: HostInfo, cmd: str) -> None: """ Send a command to a local/remote host. Args: hostinfo (HostInfo): host information cmd (str): the command to execute """ assert hostinfo.hostname in self.master_send_conns, \ f'{hostinfo} is not found in the current connections' conn = self.master_send_conns[hostinfo.hostname] conn.send(cmd) def stop_all(self) -> None: """ Stop connections to all hosts. """ for hostname, conn in self.master_send_conns.items(): conn.send('exit') def recv_from_all(self) -> dict: """ Receive messages from all hosts Returns: msg_from_node (dict): a dictionry which contains messages from each node """ msg_from_node = dict() for hostname, conn in self.master_recv_conns.items(): msg_from_node[hostname] = conn.recv() return msg_from_node
import click from colossalai.context import Config from .run import launch_multi_processes @click.command(help="Launch distributed training on a single node or multiple nodes", context_settings=dict(ignore_unknown_options=True)) @click.option("-H", "-host", "--host", type=str, default=None, help="the list of hostnames to launch in the format <host1>,<host2>") @click.option( "--hostfile", type=str, default=None, help="Hostfile path that defines the device pool available to the job, each line in the file is a hostname") @click.option("--include", type=str, default=None, help="Specify computing devices to use during execution. String format is <host1>,<host2>," " only effective when used with --hostfile.") @click.option( "--exclude", type=str, default=None, help= "Specify computing devices to NOT use during execution. Mutually exclusive with --include. Formatting is the same as --includ," " only effective when used with --hostfile.") @click.option("--num_nodes", type=int, default=-1, help="Total number of worker nodes to use, only effective when used with --hostfile.") @click.option("--nproc_per_node", type=int, default=None, help="Number of GPUs to use on each node.") @click.option("--master_port", type=int, default=29500, help="(optional) Port used by PyTorch distributed for communication during distributed training.") @click.option("--master_addr", type=str, default="127.0.0.1", help="(optional) IP address of node 0, will be inferred via 'hostname -I' if not specified.") @click.option( "--extra_launch_args", type=str, default=None, help= "Set additional torch distributed launcher arguments such as --standalone. The format is --extra_launch_args arg1=1,arg2=2. " "This will be converted to --arg1=1 --arg2=2 during execution") @click.option("--ssh-port", type=int, default=None, help="(optional) the port used for ssh connection") @click.argument("user_script", type=str) @click.argument('user_args', nargs=-1) def run(host: str, hostfile: str, num_nodes: int, nproc_per_node: int, include: str, exclude: str, master_addr: str, master_port: int, extra_launch_args: str, ssh_port: int, user_script: str, user_args: str) -> None: """ To launch multiple processes on a single node or multiple nodes via command line. Usage:: # run with 4 GPUs on the current node use default port 29500 colossalai run --nprocs_per_node 4 train.py # run with 2 GPUs on the current node at port 29550 colossalai run --nprocs_per_node 4 --master_port 29550 train.py # run on two nodes colossalai run --host <host1>,<host2> --master_addr host1 --nprocs_per_node 4 train.py # run with hostfile colossalai run --hostfile <file_path> --master_addr <host> --nprocs_per_node 4 train.py # run with hostfile with only included hosts colossalai run --hostfile <file_path> --master_addr host1 --include host1,host2 --nprocs_per_node 4 train.py # run with hostfile excluding the hosts selected colossalai run --hostfile <file_path> --master_addr host1 --exclude host2 --nprocs_per_node 4 train.py """ if not user_script.endswith('.py'): click.echo(f'Error: invalid Python file {user_script}. Did you use a wrong option? Try colossalai run --help') exit() args_dict = locals() args = Config(args_dict) args.user_args = list(args.user_args) launch_multi_processes(args)
import socket from typing import List class HostInfo: """ A data class to store host connection-related data. Args: hostname (str): name or IP address of the host port (str): the port for ssh connection """ def __init__( self, hostname: str, port: str = None, ): self.hostname = hostname self.port = port self.is_local_host = HostInfo.is_host_localhost(hostname, port) @staticmethod def is_host_localhost(hostname: str, port: str = None) -> None: """ Check if the host refers to the local machine. Args: hostname (str): name or IP address of the host port (str): the port for ssh connection Returns: bool: True if it is local, False otherwise """ if port is None: port = 22 # no port specified, lets just use the ssh port # socket.getfqdn("127.0.0.1") does not return localhost # on some users' machines # thus, we directly return True if hostname is locahost, 127.0.0.1 or 0.0.0.0 if hostname in ("localhost", "127.0.0.1", "0.0.0.0"): return True hostname = socket.getfqdn(hostname) localhost = socket.gethostname() localaddrs = socket.getaddrinfo(localhost, port) targetaddrs = socket.getaddrinfo(hostname, port) for (family, socktype, proto, canonname, sockaddr) in localaddrs: for (rfamily, rsocktype, rproto, rcanonname, rsockaddr) in targetaddrs: if rsockaddr[0] == sockaddr[0]: return True return False def __str__(self): return f'hostname: {self.hostname}, port: {self.port}' def __repr__(self): return self.__str__() class HostInfoList: """ A data class to store a list of HostInfo objects. """ def __init__(self): self.hostinfo_list = [] def append(self, hostinfo: HostInfo) -> None: """ Add an HostInfo object to the list. Args: hostinfo (HostInfo): host information """ self.hostinfo_list.append(hostinfo) def remove(self, hostname: str) -> None: """ Add an HostInfo object to the list. Args: hostname (str): the name of the host """ hostinfo = self.get_hostinfo(hostname) self.hostinfo_list.remove(hostinfo) def get_hostinfo(self, hostname: str) -> HostInfo: """ Return the HostInfo object which matches with the hostname. Args: hostname (str): the name of the host Returns: hostinfo (HostInfo): the HostInfo object which matches with the hostname """ for hostinfo in self.hostinfo_list: if hostinfo.hostname == hostname: return hostinfo raise Exception(f"Hostname {hostname} is not found") def has(self, hostname: str) -> bool: """ Check if the hostname has been added. Args: hostname (str): the name of the host Returns: bool: True if added, False otherwise """ for hostinfo in self.hostinfo_list: if hostinfo.hostname == hostname: return True return False def __iter__(self): return iter(self.hostinfo_list) def __len__(self): return len(self.hostinfo_list)
import click from .check_installation import check_installation __all__ = ['check'] @click.command(help="Check if Colossal-AI is correct based on the given option") @click.option('-i', '--installation', is_flag=True, help="Check if Colossal-AI is built correctly") def check(installation): if installation: check_installation() return click.echo("No option is given")
import subprocess import click import torch from torch.utils.cpp_extension import CUDA_HOME import colossalai def to_click_output(val): # installation check output to understandable symbols for readability VAL_TO_SYMBOL = {True: u'\u2713', False: 'x', None: 'N/A'} if val in VAL_TO_SYMBOL: return VAL_TO_SYMBOL[val] else: return val def check_installation(): """ This function will check the installation of colossalai, specifically, the version compatibility of colossalai, pytorch and cuda. Example: ```text ``` Returns: A table of installation information. """ found_aot_cuda_ext = _check_aot_built_cuda_extension_installed() cuda_version = _check_cuda_version() torch_version, torch_cuda_version = _check_torch_version() colossalai_verison, torch_version_required, cuda_version_required = _parse_colossalai_version() # if cuda_version is None, that means either # CUDA_HOME is not found, thus cannot compare the version compatibility if not cuda_version: sys_torch_cuda_compatibility = None else: sys_torch_cuda_compatibility = _is_compatible([cuda_version, torch_cuda_version]) # if cuda_version or cuda_version_required is None, that means either # CUDA_HOME is not found or AOT compilation is not enabled # thus, there is no need to compare the version compatibility at all if not cuda_version or not cuda_version_required: sys_colossalai_cuda_compatibility = None else: sys_colossalai_cuda_compatibility = _is_compatible([cuda_version, cuda_version_required]) # if torch_version_required is None, that means AOT compilation is not enabled # thus there is no need to compare the versions if torch_version_required is None: torch_compatibility = None else: torch_compatibility = _is_compatible([torch_version, torch_version_required]) click.echo(f'#### Installation Report ####') click.echo(f'\n------------ Environment ------------') click.echo(f"Colossal-AI version: {to_click_output(colossalai_verison)}") click.echo(f"PyTorch version: {to_click_output(torch_version)}") click.echo(f"CUDA version: {to_click_output(cuda_version)}") click.echo(f"CUDA version required by PyTorch: {to_click_output(torch_cuda_version)}") click.echo("") click.echo(f"Note:") click.echo(f"1. The table above checks the versions of the libraries/tools in the current environment") click.echo(f"2. If the CUDA version is N/A, you can set the CUDA_HOME environment variable to locate it") click.echo(f'\n------------ CUDA Extensions AOT Compilation ------------') click.echo(f"Found AOT CUDA Extension: {to_click_output(found_aot_cuda_ext)}") click.echo(f"PyTorch version used for AOT compilation: {to_click_output(torch_version_required)}") click.echo(f"CUDA version used for AOT compilation: {to_click_output(cuda_version_required)}") click.echo("") click.echo(f"Note:") click.echo( f"1. AOT (ahead-of-time) compilation of the CUDA kernels occurs during installation when the environment varialbe CUDA_EXT=1 is set" ) click.echo(f"2. If AOT compilation is not enabled, stay calm as the CUDA kernels can still be built during runtime") click.echo(f"\n------------ Compatibility ------------") click.echo(f'PyTorch version match: {to_click_output(torch_compatibility)}') click.echo(f"System and PyTorch CUDA version match: {to_click_output(sys_torch_cuda_compatibility)}") click.echo(f"System and Colossal-AI CUDA version match: {to_click_output(sys_colossalai_cuda_compatibility)}") click.echo(f"") click.echo(f"Note:") click.echo(f"1. The table above checks the version compatibility of the libraries/tools in the current environment") click.echo( f" - PyTorch version mistach: whether the PyTorch version in the current environment is compatible with the PyTorch version used for AOT compilation" ) click.echo( f" - System and PyTorch CUDA version match: whether the CUDA version in the current environment is compatible with the CUDA version required by PyTorch" ) click.echo( f" - System and Colossal-AI CUDA version match: whether the CUDA version in the current environment is compatible with the CUDA version used for AOT compilation" ) def _is_compatible(versions): """ Compare the list of versions and return whether they are compatible. """ if None in versions: return False # split version into [major, minor, patch] versions = [version.split('.') for version in versions] for version in versions: if len(version) == 2: # x means unknown version.append('x') for idx, version_values in enumerate(zip(*versions)): equal = len(set(version_values)) == 1 if idx in [0, 1] and not equal: return False elif idx == 1: return True else: continue def _parse_colossalai_version(): """ Get the Colossal-AI version information. Returns: colossalai_version: Colossal-AI version. torch_version_for_aot_build: PyTorch version used for AOT compilation of CUDA kernels. cuda_version_for_aot_build: CUDA version used for AOT compilation of CUDA kernels. """ # colossalai version can be in two formats # 1. X.X.X+torchX.XXcuXX.X (when colossalai is installed with CUDA extensions) # 2. X.X.X (when colossalai is not installed with CUDA extensions) # where X represents an integer. colossalai_verison = colossalai.__version__.split('+')[0] try: torch_version_for_aot_build = colossalai.__version__.split('torch')[1].split('cu')[0] cuda_version_for_aot_build = colossalai.__version__.split('cu')[1] except: torch_version_for_aot_build = None cuda_version_for_aot_build = None return colossalai_verison, torch_version_for_aot_build, cuda_version_for_aot_build def _check_aot_built_cuda_extension_installed(): """ According to `op_builder/README.md`, the CUDA extension can be built with either AOT (ahead-of-time) or JIT (just-in-time) compilation. AOT compilation will build CUDA extensions to `colossalai._C` during installation. JIT (just-in-time) compilation will build CUDA extensions to `~/.cache/colossalai/torch_extensions` during runtime. """ try: import colossalai._C.fused_optim found_aot_cuda_ext = True except ImportError: found_aot_cuda_ext = False return found_aot_cuda_ext def _check_torch_version(): """ Get the PyTorch version information. Returns: torch_version: PyTorch version. torch_cuda_version: CUDA version required by PyTorch. """ # get torch version torch_version = torch.__version__.split('+')[0] # get cuda version in pytorch build torch_cuda_major = torch.version.cuda.split(".")[0] torch_cuda_minor = torch.version.cuda.split(".")[1] torch_cuda_version = f'{torch_cuda_major}.{torch_cuda_minor}' return torch_version, torch_cuda_version def _check_cuda_version(): """ Get the CUDA version information. Returns: cuda_version: CUDA version found on the system. """ # get cuda version if CUDA_HOME is None: cuda_version = CUDA_HOME else: raw_output = subprocess.check_output([CUDA_HOME + "/bin/nvcc", "-V"], universal_newlines=True) output = raw_output.split() release_idx = output.index("release") + 1 release = output[release_idx].split(".") bare_metal_major = release[0] bare_metal_minor = release[1][0] cuda_version = f'{bare_metal_major}.{bare_metal_minor}' return cuda_version
from abc import ABC, abstractmethod from copy import deepcopy from typing import Any, List import torch from torch.fx import Graph, Node from colossalai.auto_parallel.passes.runtime_apply_pass import ( runtime_apply, runtime_apply_for_iterable_object, runtime_comm_spec_apply, ) from colossalai.fx.codegen.activation_checkpoint_codegen import ActivationCheckpointCodeGen __all___ = ['CheckpointSolverBase'] def _copy_output(src: Graph, dst: Graph): """Copy the output node from src to dst""" for n_src, n_dst in zip(src.nodes, dst.nodes): if n_src.op == 'output': n_dst.meta = n_src.meta def _get_param_size(module: torch.nn.Module): """Get the size of the parameters in the module""" return sum([p.numel() * torch.tensor([], dtype=p.dtype).element_size() for p in module.parameters()]) class CheckpointSolverBase(ABC): def __init__( self, graph: Graph, free_memory: float = -1.0, requires_linearize: bool = False, cnode: List[str] = None, optim_multiplier: float = 1.0, ): """``CheckpointSolverBase`` class will integrate information provided by the components and use an existing solver to find a possible optimal strategies combination for target computing graph. Existing Solvers: Chen's Greedy solver: https://arxiv.org/abs/1604.06174 (CheckpointSolverChen) Rotor solver: https://hal.inria.fr/hal-02352969 (CheckpointSolverRotor) Args: graph (Graph): The computing graph to be optimized. free_memory (float): Memory constraint for the solution. requires_linearize (bool): Whether the graph needs to be linearized. cnode (List[str], optional): Common node List, should be the subset of input. Default to None. optim_multiplier (float, optional): The multiplier of extra weight storage for the ``torch.optim.Optimizer``. Default to 1.0. Warnings: Meta information of the graph is required for any ``CheckpointSolver``. """ # super-dainiu: this graph is a temporary graph which can refer to # the owning module, but we will return another deepcopy of it after # the solver is executed. self.graph = deepcopy(graph) self.graph.owning_module = graph.owning_module _copy_output(graph, self.graph) self.graph.set_codegen(ActivationCheckpointCodeGen()) # check if has meta information if any(len(node.meta) == 0 for node in self.graph.nodes): raise RuntimeError( "Nodes meta information hasn't been prepared! Please extract from graph before constructing the solver!" ) # parameter memory = parameter size + optimizer extra weight storage self.free_memory = free_memory - _get_param_size(self.graph.owning_module) * (optim_multiplier + 1) self.cnode = cnode self.requires_linearize = requires_linearize if self.requires_linearize: self.node_list = self._linearize_graph() else: self.node_list = self.get_node_list() @abstractmethod def solve(self): """Solve the checkpointing problem and return the solution. """ pass def get_node_list(self): """Get the node list. """ return [[node] for node in self.graph.nodes] def _linearize_graph(self) -> List[List[Node]]: """Linearizing the graph Args: graph (Graph): The computing graph to be optimized. Returns: List[List[Node]]: List of list, each inside list of Node presents the actual 'node' in linearized manner. Remarks: Do merge the inplace ops and shape-consistency ops into the previous node. """ # Common nodes are type of nodes that could be seen as attributes and remain # unchanged throughout the whole model, it will be used several times by # different blocks of model, so that it is hard for us to linearize the graph # when we encounter those kinds of nodes. We let users to annotate some of the # input as common node, such as attention mask, and the followings are some of # the ops that could actually be seen as common nodes. With our common node prop, # we could find some of the "real" common nodes (e.g. the real attention mask # used in BERT and GPT), the rule is simple, for node who's parents are all common # nodes or it's op belongs to the following operations, we view this node as a # newly born common node. # List of target name that could be seen as common node common_ops = ["getattr", "getitem", "size"] def _is_cop(target: Any) -> bool: """Check if an op could be seen as common node Args: target (Any): node target Returns: bool """ if isinstance(target, str): return target in common_ops else: return target.__name__ in common_ops def _is_sink() -> bool: """Check if we can free all dependencies Returns: bool """ def _is_inplace(n: Node): """Get the inplace argument from ``torch.fx.Node`` """ inplace = False if n.op == "call_function": inplace = n.kwargs.get("inplace", False) elif n.op == "call_module": inplace = getattr(n.graph.owning_module.get_submodule(n.target), "inplace", False) return inplace def _is_shape_consistency(n: Node): """Check if this node is shape-consistency node (i.e. ``runtime_apply`` or ``runtime_apply_for_iterable_object``) """ return n.target in [runtime_apply, runtime_apply_for_iterable_object, runtime_comm_spec_apply] return not sum([v for _, v in deps.items()]) and not any(map(_is_inplace, n.users)) and not any( map(_is_shape_consistency, n.users)) # make sure that item in cnode is valid if self.cnode: for name in self.cnode: try: assert next(node for node in self.graph.nodes if node.name == name).op == "placeholder", \ f"Common node {name} is not an input of the model." except StopIteration: raise ValueError(f"Common node name {name} not in graph.") else: self.cnode = [] deps = {} node_list = [] region = [] for n in self.graph.nodes: if n.op != "placeholder" and n.op != "output": for n_par in n.all_input_nodes: if n_par.op != "placeholder" and n_par.name not in self.cnode: deps[n_par] -= 1 region.append(n) # if the node could free all dependencies in graph # we could begin a new node if _is_sink(): node_list.append(region) region = [] # propagate common node attr if possible if len(n.all_input_nodes) == len([node for node in n.all_input_nodes if node.name in self.cnode ]) or _is_cop(n.target): self.cnode.append(n.name) else: deps[n] = len([user for user in n.users if user.op != "output"]) return node_list
from .ckpt_solver_base import CheckpointSolverBase from .ckpt_solver_chen import CheckpointSolverChen from .ckpt_solver_rotor import CheckpointSolverRotor
import math from abc import ABC from typing import Any, Iterable, List from torch.utils._pytree import tree_map class Chain: def __init__(self, ftime: List[float], btime: List[float], x: List[int], xbar: List[int], ftmp: List[int], btmp: List[int], check_consistency: bool = True): """The chain is a basic linearized structure for solving the dynamic programming problem for activation checkpoint. See paper https://hal.inria.fr/hal-02352969 for details. Args: ftime (List[float]): The forward time of each node. btime (List[float]): The backward time of each node. x (List[int]): The forward memory of each node (if save_output). Same as `a` in the paper. xbar (List[int]): The forward memory of each node (if save_all). Same as `a_bar` in the paper. ftmp (List[int]): The temporary forward memory of each node. btmp (List[int]): The temporary backward memory of each node, can be used to control memory budget. check_consistency (bool, optional): Check the lengths consistency for the `Chain`. Defaults to True. """ self.ftime = ftime self.btime = btime self.x = x self.xbar = xbar self.ftmp = ftmp self.btmp = btmp if check_consistency and not self.check_lengths(): raise AttributeError("In Chain, input lists do not have consistent lengths") def check_lengths(self): return ((len(self.ftime) == len(self)) and (len(self.btime) == len(self) + 1) and (len(self.x) == len(self) + 1) and (len(self.ftmp) == len(self)) and (len(self.btmp) == len(self) + 1) and (len(self.xbar) == len(self) + 1)) def __repr__(self): chain_list = [] for i in range(len(self)): chain_list.append((self.ftime[i], self.btime[i], self.x[i], self.xbar[i], self.ftmp[i], self.btmp[i])) i = len(self) chain_list.append((None, self.btime[i], self.x[i], self.xbar[i], None, self.btmp[i])) return chain_list.__repr__() def __len__(self): return len(self.ftime) def discretize_all(self, unit: int): """Discretize the chain into a list of chains according to unit size.""" discretizer = lambda val: math.ceil(val / unit) self.x = tree_map(discretizer, self.x) self.xbar = tree_map(discretizer, self.xbar) self.ftmp = tree_map(discretizer, self.ftmp) self.btmp = tree_map(discretizer, self.btmp) class Operation(ABC): name = "Op" def __repr__(self) -> str: return f"{self.name}_{self.index}" def shift(self, value): if type(self.index) is tuple: self.index = tuple(x + value for x in self.index) else: self.index += value class Forward(Operation): name = "F" def __init__(self, index): self.index = index def cost(self, chain: Chain): if chain is not None: return chain.ftime[self.index] else: return 1 class ForwardEnable(Forward): name = "Fe" class ForwardNograd(Forward): name = "Fn" class ForwardCheck(Forward): name = "CF" class Forwards(Operation): def __init__(self, start, end): self.index = (start, end) def __repr__(self): return "F_{i}->{j}".format(i=self.index[0], j=self.index[1]) def cost(self, chain: Chain): if chain is not None: return sum(chain.ftime[self.index[0]:self.index[1] + 1]) else: return (self.index[1] - self.index[0] + 1) def isForward(op): return type(op) is Forward or type(op) is Forwards class Backward(Operation): name = "B" def __init__(self, index): self.index = index def cost(self, chain: Chain): if chain is not None: return chain.btime[self.index] else: return 1 class Loss(Operation): def __init__(self): pass def __repr__(self): return "L" def cost(self, chain): return 0 class MemoryAccess(Operation): name = "MA" def __init__(self, index): self.index = index def cost(self, chain: Chain): return 0 class WriteMemory(MemoryAccess): name = "WM" class ReadMemory(MemoryAccess): name = "RM" class DiscardMemory(MemoryAccess): name = "DM" class Sequence(list): def __init__(self): super().__init__() def __repr__(self): return repr(self.list_operations()) def list_operations(self): op_list = [] for x in self: if isinstance(x, Operation): op_list.append(x) else: assert isinstance(x, Sequence) op_list += x.list_operations() return op_list
from copy import deepcopy from typing import Any, Dict, List, Tuple from torch import Tensor from torch.fx import Graph, Node from colossalai.auto_parallel.passes.runtime_apply_pass import runtime_apply, runtime_comm_spec_apply from colossalai.fx.codegen.activation_checkpoint_codegen import _find_nested_ckpt_regions from colossalai.fx.profiler import ( activation_size, calculate_bwd_time, calculate_fwd_out, calculate_fwd_time, calculate_fwd_tmp, ) from colossalai.logging import get_dist_logger from .ckpt_solver_base import CheckpointSolverBase from .operation import Backward, Chain, ForwardCheck, ForwardEnable, ForwardNograd, Loss, Sequence __all__ = ['CheckpointSolverRotor'] class CheckpointSolverRotor(CheckpointSolverBase): def __init__(self, graph: Graph, free_memory: float = -1, cnode: List[str] = None, memory_slots: int = 500, optim_multiplier: float = 1.0): """This is the simple implementation of dynamic programming algorithm rotor in https://hal.inria.fr/hal-02352969. Some code are adapted from https://gitlab.inria.fr/hiepacs/rotor. Usage: Assume that we have a ``GraphModule``, and we have already done the extractions to the graph to retrieve all information needed, then we could use the following code to find a solution using ``CheckpointSolverRotor``: >>> solver = CheckpointSolverRotor(gm.graph, free_memory=torch.cuda.mem_get_info(device=0)[0]) >>> rotor_graph = solver.solve(force_python=True) # otherwise use C solver >>> gm.graph = rotor_graph # set the graph to a new graph Args: graph (Graph): The computing graph to be optimized. free_memory (float, optional): Memory constraint for the solution, unit is byte. Use ``torch.cuda.mem_get_info(device=0)[0]`` to estimate the free_memory. Defaults to -1. cnode (List[str], optional): Common node List, should be the subset of input. Defaults to None. memory_slots (int, optional): Number of slots for discretizing memory budget. Defaults to 500. optim_multiplier (float, optional): The multiplier of extra weight storage for the ``torch.optim.Optimizer``. Default to 1.0. """ super().__init__(graph, free_memory, True, cnode, optim_multiplier) self.memory_slots = memory_slots # construct chain unit = self.free_memory // self.memory_slots self.chain = self._construct_chain(self.graph, self.node_list) self.chain.discretize_all(unit) self.cost_table = None self.back_ptr = None self.sequence = None def solve(self, force_python: bool = False, verbose: bool = False) -> Graph: """Solve the checkpointing problem using rotor algorithm. Args: force_python (bool, optional): Use Python version of solver, else use C version. Defaults to False. verbose (bool, optional): Print verbose information. Defaults to False. Returns: graph (Graph): The optimized graph, should be a copy of the original graph. """ chain = self.chain # compute cost table if force_python: self.cost_table, self.back_ptr = self._compute_table(chain, self.memory_slots) else: self.cost_table, self.back_ptr = self._compute_table_c(chain, self.memory_slots) if verbose: self.print_chain() # backtrack try: self.sequence = self._backtrack(chain, 0, len(chain), self.memory_slots - chain.x[0], self.cost_table, self.back_ptr) self._annotate_from_sequence(self.sequence, self.node_list) except ValueError as e: # using logger to annonce that the solver is failed logger = get_dist_logger() logger.warning(f'Checkpoint solver failed: {e}') raise ValueError if verbose: self.print_sequence() return deepcopy(self.graph) def print_chain(self): print('[input]', self.chain.x[0], self.chain.xbar[0], self.chain.ftmp[0], self.chain.btmp[0]) for idx in range(len(self.node_list) - 1): print(self.node_list[idx], self.chain.x[idx + 1], self.chain.xbar[idx + 1], self.chain.ftmp[idx], self.chain.btmp[idx]) print(f'Chain = {self.chain}') def print_sequence(self): print(f'Sequence = {self.sequence}') @classmethod def _construct_chain(cls, graph: Graph, node_list: List[List[Node]]) -> Chain: input_tensors = cls._extract_input(graph) ftime, btime, ftmp, btmp = list(), list(), list(), list() xbar, x = [activation_size(input_tensors)], [activation_size(input_tensors)] for node in node_list: node_info = cls._extract_node_info(node) ftime.append(node_info[0]) btime.append(node_info[1]) x.append(node_info[2]) xbar.append(node_info[3]) ftmp.append(node_info[4]) btmp.append(node_info[5]) # currently we view loss backward temp as zero btime.append(0) btmp.append(0) return Chain(ftime, btime, x, xbar, ftmp, btmp) @classmethod def _extract_node_info(cls, node: List[Node]) -> Tuple[int, ...]: """Extract node info from a list of nodes""" xbar = 0 ftime = 0 btime = 0 fwd_mem_peak = 0 for n in node: assert isinstance(n, Node), f'{n} is not a Node' if n.target == runtime_apply or n.target == runtime_comm_spec_apply: # in this case we need to calculate memory usage directly based on the statics that hooked in node.meta xbar += n.meta['fwd_mem_out'] fwd_mem_peak = max(fwd_mem_peak, xbar + n.meta['fwd_mem_tmp']) else: xbar += calculate_fwd_tmp(n) + calculate_fwd_out(n) fwd_mem_peak = max(fwd_mem_peak, xbar + n.meta['fwd_mem_tmp'] + cls._extract_unused_output(n)) # minimum flop count is required ftime += max(calculate_fwd_time(n), 1.0) btime += max(calculate_bwd_time(n), 1.0) x = calculate_fwd_out(node[-1]) xbar = max(x, xbar) ftmp = fwd_mem_peak - xbar btmp = cls._extract_btmp(node) return ftime, btime, x, xbar, ftmp, btmp @staticmethod def _extract_input(graph: Graph) -> Tuple[Tensor, ...]: """Extract input tensors from a Graph""" input_tensors = [] for node in graph.nodes: if node.op == 'placeholder': input_tensors.append(node.meta['fwd_out']) return input_tensors @staticmethod def _extract_unused_output(node: Node) -> int: """Extract unused output from `torch.fx.Node`""" return activation_size(node.meta['fwd_out']) - calculate_fwd_out(node) @staticmethod def _extract_btmp(node: List[Node]) -> int: """Extract btmp from a list of nodes""" def _extract_deps_size(): deps_size = 0 for k, v in deps.items(): k: Node if v > 0: deps_size += k.meta['bwd_mem_out'] if v == float('-inf'): deps_size -= calculate_fwd_tmp(k) + calculate_fwd_out(k) return deps_size btmp = 0 deps = {} for n in reversed(node): deps[n] = len(n.all_input_nodes) btmp = max(btmp, _extract_deps_size() + n.meta['bwd_mem_tmp']) for child in n.users: if child in deps: deps[child] -= 1 if deps[child] <= 0: deps[child] = float('-inf') # free return btmp @staticmethod def _compute_table(chain: Chain, mmax: int) -> Tuple: """Compute the table using dynamic programming. Returns the cost table and the backtracking pointer. Args: chain (Chain): A basic linearized structure for solving the dynamic programming problem. mmax (int): Maximum number of memory slots. Returns: cost_table (List): cost_table[m][lhs][rhs] with lhs = 0...chain.length and rhs = lhs...chain.length (lhs is not included) and m = 0...mmax back_ptr (List): back_ptr[m][lhs][rhs] is (True,) if the optimal choice is a chain checkpoint (False, j) if the optimal choice is a leaf checkpoint of length j """ ftime = chain.ftime + [0.0] btime = chain.btime x = chain.x + [0] xbar = chain.xbar + [0] ftmp = chain.ftmp + [0] btmp = chain.btmp + [0] # Build table cost_table = [[{} for _ in range(len(chain) + 1)] for _ in range(mmax + 1)] back_ptr = [[{} for _ in range(len(chain) + 1)] for _ in range(mmax + 1)] # Last one is a dict because its indices go from i to l. Renumbering will wait for C implementation # Initialize borders of the tables for lmax-lmin = 0 for m in range(mmax + 1): for i in range(len(chain) + 1): limit = max(x[i + 1] + xbar[i + 1] + ftmp[i], x[i + 1] + xbar[i + 1] + btmp[i]) if m >= limit: # Equation (1) cost_table[m][i][i] = ftime[i] + btime[i] else: cost_table[m][i][i] = float("inf") # Compute everything for m in range(mmax + 1): for d in range(1, len(chain) + 1): for i in range(len(chain) + 1 - d): idx = i + d mmin = x[idx + 1] + x[i + 1] + ftmp[i] if idx > i + 1: mmin = max(mmin, x[idx + 1] + max(x[j] + x[j + 1] + ftmp[j] for j in range(i + 1, idx))) if m < mmin: cost_table[m][i][idx] = float("inf") else: leaf_checkpoints = [(j, sum(ftime[i:j]) + cost_table[m - x[j]][j][idx] + cost_table[m][i][j - 1]) for j in range(i + 1, idx + 1) if m >= x[j]] if leaf_checkpoints: best_leaf = min(leaf_checkpoints, key=lambda t: t[1]) else: best_leaf = None if m >= xbar[i + 1]: chain_checkpoint = cost_table[m][i][i] + cost_table[m - xbar[i + 1]][i + 1][idx] else: chain_checkpoint = float("inf") if best_leaf and best_leaf[1] <= chain_checkpoint: cost_table[m][i][idx] = best_leaf[1] back_ptr[m][i][idx] = (False, best_leaf[0]) else: cost_table[m][i][idx] = chain_checkpoint back_ptr[m][i][idx] = (True,) return cost_table, back_ptr @staticmethod def _compute_table_c(chain: Chain, mmax: int) -> Tuple: try: from .rotorc import compute_table # build module if module not found except ModuleNotFoundError: import os import subprocess import sys logger = get_dist_logger() logger.info("rotorc hasn't been built! Building library...", ranks=[0]) this_dir = os.path.dirname(os.path.abspath(__file__)) result = subprocess.Popen( [ f"{sys.executable}", f"{os.path.join(this_dir, 'build_c_ext.py')}", "build_ext", f"--build-lib={this_dir}" ], stdout=subprocess.PIPE, stderr=subprocess.PIPE, ) if result.wait() == 0: logger.info("rotorc has been built!", ranks=[0]) from .rotorc import compute_table else: logger.warning("rotorc built failed! Using python version!", ranks=[0]) return CheckpointSolverRotor._compute_table(chain, mmax) return compute_table(chain, mmax) @staticmethod def _backtrack(chain: Chain, lhs: int, rhs: int, budget: int, cost_table: List[Any], back_ptr: List[Any]) -> "Sequence": """Backtrack the cost table and retrieve the optimal checkpointing strategy. Args: chain (Chain): A basic linearized structure for solving the dynamic programming problem. lhs (int): The left index of the interval to backtrack. rhs (int): The right index of the interval to backtrack. budget (int): The memory budget for processing this interval. cost_table (List[Any]): See ``._compute_table()`` for definitions back_ptr (List[Any]): See ``._compute_table()`` for definitions Raises: ValueError: Can not process the chain. Returns: sequence (Sequence): The sequence of executing nodes with checkpoints. """ if budget <= 0: raise ValueError(f"Can not process a chain with negative memory {budget}") elif cost_table[budget][lhs][rhs] == float("inf"): raise ValueError(f"Can not process this chain from index {lhs} to {rhs} with memory {budget}") sequence = Sequence() if rhs == lhs: if lhs == len(chain): sequence += [Loss()] else: sequence += [ForwardEnable(lhs), Backward(lhs)] return sequence if back_ptr[budget][lhs][rhs][0]: sequence += [ ForwardEnable(lhs), CheckpointSolverRotor._backtrack(chain, lhs + 1, rhs, budget - chain.xbar[lhs + 1], cost_table, back_ptr), Backward(lhs), ] else: best_leaf = back_ptr[budget][lhs][rhs][1] sequence += [ForwardCheck(lhs)] sequence += [ForwardNograd(k) for k in range(lhs + 1, best_leaf)] sequence += [ CheckpointSolverRotor._backtrack(chain, best_leaf, rhs, budget - chain.x[best_leaf], cost_table, back_ptr), CheckpointSolverRotor._backtrack(chain, lhs, best_leaf - 1, budget, cost_table, back_ptr), ] return sequence @staticmethod def _annotate_from_sequence(sequence: Sequence, node_list: List[List[Node]]): """Annotate the nodes in the ``node_list`` with activation checkpoint from the sequence. Args: sequence (Sequence): The sequence of executing nodes with activation checkpoint annotations. node_list (List[List[Node]]): The list of nodes to annotate. """ op_list = sequence.list_operations() loss_op = next(op for op in op_list if isinstance(op, Loss)) fwd_list = op_list[:op_list.index(loss_op)] bwd_list = op_list[op_list.index(loss_op) + 1:] ckpt_idx = 0 in_ckpt = False ckpt_region = [] # forward annotation for idx, op in enumerate(fwd_list, 0): if in_ckpt: if isinstance(op, ForwardNograd): ckpt_region.append(idx) elif isinstance(op, ForwardEnable): in_ckpt = False for node_idx in ckpt_region: for n in node_list[node_idx]: n.meta['activation_checkpoint'] = [ckpt_idx] ckpt_idx += 1 ckpt_region = [] elif isinstance(op, ForwardCheck): for node_idx in ckpt_region: for n in node_list[node_idx]: n.meta['activation_checkpoint'] = [ckpt_idx] ckpt_idx += 1 ckpt_region = [idx] else: if isinstance(op, ForwardCheck): in_ckpt = True ckpt_region.append(idx) # annotate the backward if there is any nested activation checkpoint in_recompute = False for op in bwd_list: if in_recompute: if isinstance(op, ForwardNograd): ckpt_region.append(op.index) elif isinstance(op, ForwardEnable): for node_idx in ckpt_region: for n in node_list[node_idx]: n.meta['activation_checkpoint'].append(ckpt_idx) ckpt_idx += 1 ckpt_region = [] elif isinstance(op, ForwardCheck): for node_idx in ckpt_region: for n in node_list[node_idx]: n.meta['activation_checkpoint'].append(ckpt_idx) ckpt_idx += 1 ckpt_region = [op.index] elif isinstance(op, Backward): for node_idx in ckpt_region: for n in node_list[node_idx]: n.meta['activation_checkpoint'].append(ckpt_idx) in_recompute = False else: if not isinstance(op, Backward): in_recompute = True ckpt_idx = 0 ckpt_region = [] if isinstance(op, ForwardCheck): ckpt_region.append(op.index) # postprocess, make sure every activation checkpoint label in the # same activation checkpoint region (level = 0) has the same length op_list = [] for node in node_list: op_list += node ckpt_regions = _find_nested_ckpt_regions(op_list) for (start_idx, end_idx) in ckpt_regions: nested_length = max( len(op_list[idx].meta['activation_checkpoint']) for idx in range(start_idx, end_idx + 1)) for idx in range(start_idx, end_idx + 1): op_list[idx].meta['activation_checkpoint'] += [None] * (nested_length - len(op_list[idx].meta['activation_checkpoint']))
import math from copy import deepcopy from typing import List, Set, Tuple from torch.fx import Graph, Node from colossalai.fx.profiler import calculate_fwd_in, calculate_fwd_tmp from .ckpt_solver_base import CheckpointSolverBase __all__ = ['CheckpointSolverChen'] class CheckpointSolverChen(CheckpointSolverBase): def __init__(self, graph: Graph, cnode: List[str] = None, num_grids: int = 6): """ This is the simple implementation of Algorithm 3 in https://arxiv.org/abs/1604.06174. Note that this algorithm targets at memory optimization only, using techniques in appendix A. Usage: Assume that we have a ``GraphModule``, and we have already done the extractions to the graph to retrieve all information needed, then we could use the following code to find a solution using ``CheckpointSolverChen``: >>> solver = CheckpointSolverChen(gm.graph) >>> chen_graph = solver.solve() >>> gm.graph = chen_graph # set the graph to a new graph Args: graph (Graph): The computing graph to be optimized. cnode (List[str], optional): Common node List, should be the subset of input. Defaults to None. num_grids (int, optional): Number of grids to search for b. Defaults to 6. """ super().__init__(graph, 0, 0, True, cnode) self.num_grids = num_grids def solve(self) -> Graph: """Solve the checkpointing problem using Algorithm 3. Returns: graph (Graph): The optimized graph, should be a copy of the original graph. """ checkpointable_op = ['call_module', 'call_method', 'call_function', 'get_attr'] ckpt = self.grid_search() for i, seg in enumerate(ckpt): for idx in range(*seg): nodes = self.node_list[idx] for n in nodes: if n.op in checkpointable_op: n.meta['activation_checkpoint'] = i return deepcopy(self.graph) def run_chen_greedy(self, b: int = 0) -> Tuple[Set, int]: """ This is the simple implementation of Algorithm 3 in https://arxiv.org/abs/1604.06174. """ ckpt_intv = [] temp = 0 x = 0 y = 0 prev_idx = 2 for idx, nodes in enumerate(self.node_list): for n in nodes: n: Node temp += calculate_fwd_in(n) + calculate_fwd_tmp(n) y = max(y, temp) if temp > b and idx > prev_idx: x += calculate_fwd_in(nodes[0]) temp = 0 ckpt_intv.append((prev_idx, idx + 1)) prev_idx = idx + 1 return ckpt_intv, math.floor(math.sqrt(x * y)) def grid_search(self) -> Set: """ Search ckpt strategy with b = 0, then run the allocation algorithm again with b = √xy. Grid search over [√2/2 b, √2 b] for ``ckpt_opt`` over ``num_grids`` as in appendix A. """ _, b_approx = self.run_chen_greedy(0) b_min, b_max = math.floor(b_approx / math.sqrt(2)), math.ceil(b_approx * math.sqrt(2)) b_opt = math.inf for b in range(b_min, b_max, (b_max - b_min) // self.num_grids): ckpt_intv, b_approx = self.run_chen_greedy(b) if b_approx < b_opt: b_opt = b_approx ckpt_opt = ckpt_intv return ckpt_opt
import os from setuptools import Extension, setup this_dir = os.path.dirname(os.path.abspath(__file__)) ext_modules = [Extension( 'rotorc', sources=[os.path.join(this_dir, 'ckpt_solver_rotor.c')], )] setup( name='rotor c extension', version='0.1', description='rotor c extension for faster dp computing', ext_modules=ext_modules, )
from typing import Callable, List import torch from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( MemoryCost, OperationData, OperationDataType, ShardingStrategy, StrategiesVector, TrainCycleItem, ) from colossalai.tensor.sharding_spec import ShardingSpec from .constants import INPLACE_MODULE, INPLACE_OPS, NO_SAVE_ACTIVATION from .registry import meta_register __all__ = ['MetaInfo'] class MetaInfo: """MetaInfo class This class is used to store meta info based on sharding strategy and the given target function. """ def __init__(self, strategy: ShardingStrategy = None, target: Callable = None) -> None: # compute cost of forward and backward computation self.compute_cost: TrainCycleItem # compute memory cost of forward and backward phase self.memory_cost: TrainCycleItem # list of input tensors self.fwd_in: List[torch.Tensor] # list of buffer tensors self.fwd_buffer: List[torch.Tensor] # list of output tensors self.fwd_out: List[torch.Tensor] # sharding strategy self._strategy = strategy # target function self._target = target # compute metainfo if possible if self._strategy is not None and self._target is not None: self.compute_metainfo() @property def strategy(self) -> ShardingStrategy: return self._strategy @property def target(self) -> Callable: return self._target @strategy.setter def strategy(self, strategy: ShardingStrategy) -> None: self._strategy = strategy if self._strategy is not None and self._target is not None: self.compute_metainfo() @target.setter def target(self, target: Callable) -> None: self._target = target if self._strategy is not None and self._target is not None: self.compute_metainfo() def compute_sharded_opdata(self, operation_data: OperationData, sharding_spec: ShardingSpec) -> torch.Tensor: """ Compute sharded opdata based on the given data and sharding spec. """ return OperationData(name=operation_data.name, data=torch.zeros(sharding_spec.get_sharded_shape_per_device(), device="meta"), type=operation_data.type, logical_shape=operation_data.logical_shape) def compute_metainfo(self): """ Compute meta info based on sharding strategy and the given target function. """ assert meta_register.has(self._target.__class__) or meta_register.has(self._target), \ f"Meta info for {self._target} is not registered." if meta_register.has(self._target.__class__): # module meta_func = meta_register.get(self._target.__class__) # check whether the target in the list that we don't need to save activation save_fwd_in = self._target.__class__ not in NO_SAVE_ACTIVATION else: # function meta_func = meta_register.get(self._target) # check whether the target in the list that we don't need to save activation save_fwd_in = self._target.__class__ not in NO_SAVE_ACTIVATION # construct args for meta_func args = [self.compute_sharded_opdata(k, v) for k, v in self._strategy.sharding_specs.items()] # construct kwargs if self.target in INPLACE_MODULE: kwargs = {'inplace': self.target.inplace} elif self.target in INPLACE_OPS: kwargs = {'inplace': True} else: kwargs = {'inplace': False} # compute metainfo with meta_func self.compute_cost, self.memory_cost, self.fwd_in, self.fwd_buffer, self.fwd_out = meta_func(*args, **kwargs) # process corner case for NO_SAVE_ACTIVATION if not save_fwd_in: self.fwd_in = []
__all__ = ['Registry'] class Registry: def __init__(self, name): self.name = name self.store = {} def register(self, source): def wrapper(func): if isinstance(source, (list, tuple)): # support register a list of items for this func for element in source: self.store[element] = func else: self.store[source] = func return func return wrapper def get(self, source): assert source in self.store, f'{source} not found in the {self.name} registry' target = self.store[source] return target def has(self, source): return source in self.store meta_register = Registry('meta')
import operator import torch import torch.nn as nn from ..tensor_shard.constants import * # list of inplace module INPLACE_MODULE = [nn.ReLU] # list of inplace operations INPLACE_OPS = [torch.flatten] # list of operations that do not save forward activations NO_SAVE_ACTIVATION = [torch.add, torch.sub, operator.add, operator.sub]
from .meta_registry import * from .metainfo import * from .registry import meta_register
from functools import reduce from typing import Callable, Dict, List, Tuple, Union import torch from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( MemoryCost, OperationData, OperationDataType, ShardingStrategy, StrategiesVector, TrainCycleItem, ) from colossalai.fx.profiler.memory_utils import activation_size from colossalai.fx.profiler.opcount import flop_mapping from colossalai.tensor.sharding_spec import ShardingSpec from ..registry import meta_register __all__ = ['linear_meta_info', 'matmul_meta_info'] @meta_register.register(torch.nn.functional.linear) @meta_register.register(torch.nn.Linear) def linear_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """torch.nn.Linear & torch.nn.functional.linear meta info generator NOTE: currently we separate the bias part from the biased linear ops, we will consider the memory consumption in add metainfo generator, but we will hold the bias mechanism in the linear metainfo generator for future use. graph(): %input_2 : [#users=2] = placeholder[target=placeholder](default=) %addmm_default : [#users=1] = call_function[target=torch.ops.aten.addmm.default](args = (None, %input_2, None), kwargs = {}) %zeros_like_default : [#users=3] = call_function[target=torch.ops.aten.zeros_like.default](args = (%addmm_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%input_2,), kwargs = {}) %mm_default : [#users=1] = call_function[target=torch.ops.aten.mm.default](args = (%zeros_like_default, None), kwargs = {}) %t_default : [#users=1] = call_function[target=torch.ops.aten.t.default](args = (%zeros_like_default,), kwargs = {}) %mm_default_1 : [#users=1] = call_function[target=torch.ops.aten.mm.default](args = (%t_default, %detach_default), kwargs = {}) %t_default_1 : [#users=1] = call_function[target=torch.ops.aten.t.default](args = (%mm_default_1,), kwargs = {}) %sum_dim_int_list : [#users=1] = call_function[target=torch.ops.aten.sum.dim_IntList](args = (%zeros_like_default, [None], None), kwargs = {}) %view_default : [#users=1] = call_function[target=torch.ops.aten.view.default](args = (%sum_dim_int_list, [None]), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%view_default,), kwargs = {}) %detach_default_2 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_1,), kwargs = {}) %detach_default_3 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%mm_default,), kwargs = {}) %detach_default_4 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_3,), kwargs = {}) %t_default_2 : [#users=1] = call_function[target=torch.ops.aten.t.default](args = (%t_default_1,), kwargs = {}) %detach_default_5 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%t_default_2,), kwargs = {}) %detach_default_6 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_5,), kwargs = {}) The one without bias is graph(): %input_2 : [#users=2] = placeholder[target=placeholder](default=) %mm_default : [#users=1] = call_function[target=torch.ops.aten.mm.default](args = (%input_2, None), kwargs = {}) %zeros_like_default : [#users=2] = call_function[target=torch.ops.aten.zeros_like.default](args = (%mm_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%input_2,), kwargs = {}) %t_default : [#users=1] = call_function[target=torch.ops.aten.t.default](args = (%zeros_like_default,), kwargs = {}) %mm_default_1 : [#users=1] = call_function[target=torch.ops.aten.mm.default](args = (%t_default, %detach_default), kwargs = {}) %t_default_1 : [#users=1] = call_function[target=torch.ops.aten.t.default](args = (%mm_default_1,), kwargs = {}) %mm_default_2 : [#users=1] = call_function[target=torch.ops.aten.mm.default](args = (%zeros_like_default, None), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%mm_default_2,), kwargs = {}) %detach_default_2 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_1,), kwargs = {}) %t_default_2 : [#users=1] = call_function[target=torch.ops.aten.t.default](args = (%t_default_1,), kwargs = {}) %detach_default_3 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%t_default_2,), kwargs = {}) %detach_default_4 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_3,), kwargs = {}) Returns: Tuple[TrainCycleItem, TrainCycleItem, bool]: compute cost, memory cost and forward inputs """ has_bias: bool = False input_tensor = args[0].data output_tensor = args[2].data if len(args) == 4: weight_tensors = [args[1].data, args[3].data] else: weight_tensors = [args[1].data] # process the dimension of input and output if len(input_tensor.shape) > 2: input_tensor: torch.Tensor input_tensor = input_tensor.view(-1, input_tensor.shape[-1]) if len(output_tensor.shape) > 2: output_tensor: torch.Tensor output_tensor = output_tensor.view(-1, output_tensor.shape[-1]) if len(weight_tensors) > 1: has_bias = True if len(weight_tensors[0].shape) == 2: weight_tensor, bias_tensor = weight_tensors else: bias_tensor, weight_tensor = weight_tensors else: weight_tensor = weight_tensors[0] if has_bias: # calculate cost with bias # the fwd op with compute cost is addmm # the bwd op with compute cost is mm * 2 and sum.dim_IntList # calculate compute cost fwd_compute_cost = flop_mapping[torch.ops.aten.addmm.default]( [bias_tensor, input_tensor, torch.transpose(weight_tensor, 0, 1)], (output_tensor,)) bwd_compute_cost = flop_mapping[torch.ops.aten.mm.default]([output_tensor, weight_tensor], (input_tensor,)) + \ flop_mapping[torch.ops.aten.mm.default]([torch.transpose(output_tensor, 0, 1), input_tensor], (weight_tensor,)) + \ flop_mapping[torch.ops.aten.sum.dim_IntList]([output_tensor], (bias_tensor,)) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost # NOTE: Linear don't have buffer and temp in forward and backward phase # the forward activation cost is the size of output_tensor, parameter cost is the size of weight_tensor and bias_tensor # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, output_tensor]), parameter=activation_size([weight_tensor, bias_tensor]), temp=0, buffer=0) # the backward activation cost is the size of input_tensor, weight_tensor and bias_tensor, parameter cost is 0 bwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, weight_tensor, bias_tensor]), parameter=activation_size([weight_tensor, bias_tensor]), temp=0, buffer=0) # total cost is to sum the forward and backward cost total_cost = MemoryCost(activation=fwd_memory_cost.activation + bwd_memory_cost.activation, parameter=fwd_memory_cost.parameter + bwd_memory_cost.parameter) memory_cost = TrainCycleItem(fwd=fwd_memory_cost, bwd=bwd_memory_cost, total=total_cost) else: # calculate cost without bias # the fwd op with compute cost is mm # the bwd op with compute cost is mm * 2 # calculate compute cost fwd_compute_cost = flop_mapping[torch.ops.aten.mm.default]( [input_tensor, torch.transpose(weight_tensor, 0, 1)], (output_tensor,)) bwd_compute_cost = flop_mapping[torch.ops.aten.mm.default]([output_tensor, weight_tensor], (input_tensor,)) + \ flop_mapping[torch.ops.aten.mm.default]([torch.transpose(output_tensor, 0, 1), input_tensor], (weight_tensor,)) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost # NOTE: Linear don't have buffer and temp in forward and backward phase # the forward activation cost is the size of output_tensor, parameter cost is the size of weight_tensor # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, output_tensor]), parameter=activation_size(weight_tensor), temp=0, buffer=0) # the backward activation cost is the size of input_tensor and weight_tensor, parameter cost is 0 bwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, weight_tensor]), parameter=activation_size(weight_tensor), temp=0, buffer=0) # total cost is to sum the forward and backward cost total_cost = MemoryCost(activation=fwd_memory_cost.activation + bwd_memory_cost.activation, parameter=fwd_memory_cost.parameter + bwd_memory_cost.parameter) memory_cost = TrainCycleItem(fwd=fwd_memory_cost, bwd=bwd_memory_cost, total=total_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [torch.zeros_like(input_tensor, device='meta')] fwd_buffer = [] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out @meta_register.register(torch.matmul) def matmul_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """torch.matmul meta info generator There are several cases for torch.matmul: 1. Vector-vector multiplication => no temp memory, forward memory cost is 1 element (could be neglected), backward memory cost is the same as two input vectors. 2. Matrix-vector multiplication => if the first input is matrix, no temp memory is needed, otherwise, there is a temp memory in the backward phase for the transpose of the matrix. The forward memory cost is the size of output tensor, backward memory cost is the size of the two inputs; if the first input is vector, the forward memory cost is the size of the output tensor, and during the backward phase, it will allocate a temp memory the same size as the input matrix, and allocate memory for the gradient of two inputs. 3. Batched Matrix-vector multiplication => if the first input is the batched matrix, no temp memory, the forward memory cost is the size of output tensor, backward memory cost is the size of the two inputs; if the second input is the batched matrix, the matmul will allocate memory for the gradient of the batched matrix in the forward phase (as they create a new tensor without the former batches), so the forward memory cost is the output tensor and the newly created matrix (take the same amount of memory of the input batched matrix). During the backward phase, it will allocate a temp memory the same size as input batched matrix, and allocate a tensor for the gradient of the input vector. The gradient of the batched matrix will be stored in the memory allocated during the forward phase. 3. Matrix-matrix multiplication => no temp memory, forward memory is the size of output tensor, backward memory is the size of the two inputs 4. Batched matrix-matrix multiplication => if the first input is the batched matrix, no temp memory, the forward memory cost is the size of two inputs and backward memory cost is the size of the output tensor; if the second input is the batched matrix, during the forward phase it will allocate memory for the output and gradient of the second input, and has a temp memory the same size as the output, during the backward phase, it will allocate memory for the gradient of the first input and has a temp memory which is as big as output and the second input. 5. Batched matrix-batched matrix multiplication => if the two inputs have the same batch dimensions, no temp memory, the forward memory cost is the size of output, backward memory cost is the size of the two inputs; it the two inputs have different batch dimensions, during the forward phase it will allocate memory of the expanded inputs (so that the batch dimensions could match) and the output, and during the backward phase, it has a temp memory of the size of two expanded inputs, and it will allocate memory for the gradient of the two inputs and discard the expanded inputs allocated during the forward phase. Returns: Tuple[TrainCycleItem, TrainCycleItem, bool]: compute cost, memory cost and forward inputs """ # Get input and output tensors input_tensors = [args[0].data, args[1].data] output_tensors = [args[-1].data] # Check dimension if all(len(tensor.shape) == 1 for tensor in input_tensors): # Dot fwd_compute_cost = flop_mapping[torch.ops.aten.dot.default](input_tensors, output_tensors) bwd_compute_cost = flop_mapping[torch.ops.aten.mul.Tensor](input_tensors[0], output_tensors) * 2 fwd_mem_cost = MemoryCost(activation=activation_size(output_tensors), parameter=0, temp=0, buffer=0) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors), parameter=0, temp=0, buffer=0) elif len(input_tensors[0].shape) >= 2 and len(input_tensors[1].shape) == 1: # gemv case 1: matrix-vector multiplication # & # batched gemv case 1: batched matrix-vector multiplication fwd_compute_cost = flop_mapping[torch.ops.aten.mv.default]( [input_tensors[0].reshape(-1, input_tensors[0].shape[-1]), input_tensors[1]], output_tensors) # combine the dimensions of output bwd_compute_cost = flop_mapping[torch.ops.aten.mul.Tensor]( [output_tensors[0].reshape(-1), input_tensors[1]], output_tensors) + \ flop_mapping[torch.ops.aten.mv.default]( [input_tensors[0].reshape(-1, input_tensors[0].shape[-1]).transpose(0, 1), output_tensors[0].reshape(-1)], output_tensors) fwd_mem_cost = MemoryCost(activation=activation_size(output_tensors), parameter=0, temp=0, buffer=0) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors), parameter=0, temp=0, buffer=0) elif len(input_tensors[0].shape) == 1 and len(input_tensors[1].shape) == 2: # gemv case 2: vector-matrix multiplication fwd_compute_cost = flop_mapping[torch.ops.aten.mv.default](input_tensors, output_tensors) bwd_compute_cost = flop_mapping[torch.ops.aten.mul.Tensor]([output_tensors[0], input_tensors[0]], output_tensors) + \ flop_mapping[torch.ops.aten.mv.default]([input_tensors[1], output_tensors[0]], output_tensors) fwd_mem_cost = MemoryCost(activation=activation_size(output_tensors), parameter=0, temp=0, buffer=0) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors), parameter=0, temp=activation_size(input_tensors[1]), buffer=0) elif len(input_tensors[0].shape) == 1 and len(input_tensors[1].shape) >= 3: # batched gemv case 2: vector-batched matrix multiplication fwd_compute_cost = flop_mapping[torch.ops.aten.mv.default]( [input_tensors[1].transpose(-2, -1).reshape(-1, input_tensors[1].shape[-2]), input_tensors[0]], [output_tensors[0].reshape(-1)]) # combine the dimensions of output bwd_compute_cost = flop_mapping[torch.ops.aten.mul.Tensor]( [output_tensors[0].reshape(-1), input_tensors[0]], output_tensors ) + \ flop_mapping[torch.ops.aten.mv.default]( [input_tensors[1].transpose(-2, -1).reshape(-1, input_tensors[1].shape[-2]).transpose(0, 1), output_tensors[0].reshape(-1)], output_tensors ) fwd_mem_cost = MemoryCost(activation=activation_size(output_tensors + [input_tensors[1]])) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors[0]), parameter=0, temp=activation_size(input_tensors[1]), buffer=0) elif len(input_tensors[0].shape) >= 2 and len(input_tensors[1].shape) == 2: # gemm & batched gemm case 1: batched matrix-matrix multiplication fwd_compute_cost = flop_mapping[torch.ops.aten.mm.default]( [input_tensors[0].reshape(-1, input_tensors[0].shape[-1]), input_tensors[1]], [output_tensors[0].reshape(-1, output_tensors[0].shape[-1])]) bwd_compute_cost = flop_mapping[torch.ops.aten.mm.default]( [input_tensors[0].reshape(-1, input_tensors[0].shape[-1]).transpose(0, 1), output_tensors[0].reshape(-1, output_tensors[0].shape[-1])], [input_tensors[1]] ) + \ flop_mapping[torch.ops.aten.mm.default]( [output_tensors[0].reshape(-1, output_tensors[0].shape[-1]), input_tensors[1].transpose(0, 1)], [input_tensors[0].reshape(-1, input_tensors[0].shape[-1])] ) fwd_mem_cost = MemoryCost(activation=activation_size(output_tensors), parameter=0, temp=0, buffer=0) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors), parameter=0, temp=0, buffer=0) elif len(input_tensors[0].shape) == 2 and len(input_tensors[1].shape) >= 3: # batched gemm case 2: matrix-batched matrix multiplication fwd_compute_cost = flop_mapping[torch.ops.aten.mm.default]([ input_tensors[1].transpose(-2, -1).reshape(-1, input_tensors[1].shape[-2]), input_tensors[0].transpose( 0, 1) ], [output_tensors[0].transpose(-2, -1)]) bwd_compute_cost = flop_mapping[torch.ops.aten.mm.default]( [output_tensors[0].transpose(-2, -1).reshape(-1, output_tensors[0].shape[-2]).transpose(0, 1), input_tensors[1].transpose(-2, -1).reshape(-1, input_tensors[1].shape[-2])], [input_tensors[0]] ) + \ flop_mapping[torch.ops.aten.mm.default]( [output_tensors[0].transpose(-2, -1).reshape(-1, output_tensors[0].shape[-2]), input_tensors[0]], [input_tensors[1].transpose(-2, -1).reshape(-1, input_tensors[1].shape[-2])] ) fwd_mem_cost = MemoryCost(activation=activation_size(output_tensors) + activation_size(input_tensors[1]), temp=activation_size(output_tensors)) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors[0]), parameter=0, temp=activation_size(input_tensors[1]) + activation_size(output_tensors)) elif all(len(tensor.shape) >= 3 for tensor in input_tensors): # Batched matrix-batched matrix multiplication # Fetch shape of the two inputs and see if the batch dimensions are the same _is_batch_dims_same = True if len(input_tensors[0].shape) == len(input_tensors[1].shape): for (shape_0, shape_1) in zip(input_tensors[0].shape[:-2], input_tensors[1].shape[:-2]): if shape_0 != shape_1: _is_batch_dims_same = False break else: _is_batch_dims_same = False # retireve dimensions input_dim_00 = input_tensors[0].shape[-2] input_dim_01 = input_tensors[0].shape[-1] input_dim_10 = input_tensors[1].shape[-2] input_dim_11 = input_tensors[1].shape[-1] output_dim_0 = output_tensors[0].shape[-2] output_dim_1 = output_tensors[0].shape[-1] if _is_batch_dims_same: # Case 1: batch dimensions are the same # Forward compute cost: C = A * B fwd_compute_cost = flop_mapping[torch.ops.aten.bmm.default]([ input_tensors[0].reshape(-1, input_dim_00, input_dim_01), input_tensors[1].reshape( -1, input_dim_10, input_dim_11) ], [output_tensors[0].reshape(-1, output_dim_0, output_dim_1)]) # Backward compute cost: dB = A^T * dC, dA = dC * B^T bwd_compute_cost = flop_mapping[torch.ops.aten.bmm.default]( [input_tensors[0].transpose(-2, -1).reshape(-1, input_dim_01, input_dim_00), output_tensors[0].reshape(-1, output_dim_0, output_dim_1)], [input_tensors[1].reshape(-1, input_dim_11, input_dim_10)] ) + \ flop_mapping[torch.ops.aten.bmm.default]( [output_tensors[0].reshape(-1, output_dim_0, output_dim_1), input_tensors[1].transpose(-2, -1).reshape(-1, input_dim_11, input_dim_10)], [input_tensors[0].reshape(-1, input_dim_00, input_dim_01)] ) fwd_mem_cost = MemoryCost(activation=activation_size(output_tensors)) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors)) else: # Case 2: batch dimensions are different batch_dims = output_tensors[0].shape[:-2] extended_input_0 = torch.rand(reduce(lambda x, y: x * y, batch_dims), input_dim_00, input_dim_01, device="meta") extended_input_1 = torch.rand(reduce(lambda x, y: x * y, batch_dims), input_dim_10, input_dim_11, device="meta") # Forward compute cost: C = A * B fwd_compute_cost = flop_mapping[torch.ops.aten.bmm.default]( [extended_input_0, extended_input_1], [output_tensors[0].reshape(-1, output_dim_0, output_dim_1)]) # Backward compute cost: dB = A^T * dC, dA = dC * B^T bwd_compute_cost = flop_mapping[torch.ops.aten.bmm.default]( [extended_input_0.transpose(-2, -1), output_tensors[0].reshape(-1, output_dim_0, output_dim_1)], [extended_input_1] ) + \ flop_mapping[torch.ops.aten.bmm.default]( [output_tensors[0].reshape(-1, output_dim_0, output_dim_1), extended_input_1.transpose(-2, -1)], [extended_input_0] ) fwd_mem_cost = MemoryCost( activation=activation_size([output_tensors[0], extended_input_0, extended_input_1])) bwd_mem_cost = MemoryCost(activation=activation_size(input_tensors) - activation_size([extended_input_0, extended_input_1]), temp=activation_size([extended_input_0, extended_input_1])) # compute cost compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # memory cost total_cost = MemoryCost(activation=fwd_mem_cost.activation + bwd_mem_cost.activation, parameter=fwd_mem_cost.parameter + bwd_mem_cost.parameter, temp=fwd_mem_cost.temp + bwd_mem_cost.temp, buffer=fwd_mem_cost.buffer + bwd_mem_cost.buffer) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = input_tensors fwd_buffer = [] fwd_out = output_tensors return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out
from typing import Callable, Dict, List, Tuple, Union import torch from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( MemoryCost, OperationData, OperationDataType, ShardingStrategy, StrategiesVector, TrainCycleItem, ) from colossalai.fx.profiler.memory_utils import activation_size from colossalai.fx.profiler.opcount import flop_mapping from colossalai.tensor.sharding_spec import ShardingSpec from ..registry import meta_register __all__ = ['batchnormnd_meta_info', 'layernorm_meta_info'] @meta_register.register(torch.nn.BatchNorm1d) @meta_register.register(torch.nn.BatchNorm2d) @meta_register.register(torch.nn.BatchNorm3d) def batchnormnd_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """BatchNorm1d, BatchNorm2d, BatchNorm3d, meta info generator The aten graph of BatchNorm2d is like graph(): %input_2 : [#users=2] = placeholder[target=placeholder](default=) %cudnn_batch_norm_default : [#users=4] = call_function[target=torch.ops.aten.cudnn_batch_norm.default](args = (%input_2, None, None, None, None, None, None, None), kwargs = {}) %zeros_like_default : [#users=1] = call_function[target=torch.ops.aten.zeros_like.default](args = (%cudnn_batch_norm_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%input_2,), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%cudnn_batch_norm_default,), kwargs = {}) %detach_default_2 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%cudnn_batch_norm_default,), kwargs = {}) %detach_default_3 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%cudnn_batch_norm_default,), kwargs = {}) %cudnn_batch_norm_backward_default : [#users=3] = call_function[target=torch.ops.aten.cudnn_batch_norm_backward.default](args = (%detach_default, %zeros_like_default, None, None, None, %detach_default_1, %detach_default_2, None, %detach_default_3), kwargs = {}) %detach_default_4 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%cudnn_batch_norm_backward_default,), kwargs = {}) %detach_default_5 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_4,), kwargs = {}) %detach_default_6 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%cudnn_batch_norm_backward_default,), kwargs = {}) %detach_default_7 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_6,), kwargs = {}) %detach_default_8 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%cudnn_batch_norm_backward_default,), kwargs = {}) %detach_default_9 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_8,), kwargs = {}) Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ input_tensor = args[0].data output_tensor = next(filter(lambda x: x.type == OperationDataType.OUTPUT, args)).data weight_tensor = next(filter(lambda x: x.name == "weight", args)).data bias_tensor = next(filter(lambda x: x.name == "bias", args)).data mean_tensor = next(filter(lambda x: x.name == "running_mean", args)).data var_tensor = next(filter(lambda x: x.name == "running_var", args)).data num_batch = next(filter(lambda x: x.name == "num_batches_tracked", args)).data # construct fwd args # the fwd inputs are input, weight, bias, running_mean, running_var and some other args # indicating the status of the module # the fwd outputs are output, saved mean, saved inv std and num batches tracked fwd_in_args = [input_tensor, weight_tensor, bias_tensor, mean_tensor, var_tensor, True, 0.1, 1e-5] fwd_out_args = [output_tensor, mean_tensor, var_tensor, num_batch] # construct bwd args # the bwd inputs are upstream grad, input, weight, running_mean, running_var, saved mean, # saved inv std and some other args indicating the status of the module # the bwd outputs are input grad, weight grad and bias grad bwd_in_args = [ output_tensor, output_tensor, weight_tensor, mean_tensor, var_tensor, mean_tensor, var_tensor, 1e-5, num_batch ] bwd_out_args = [input_tensor, weight_tensor, bias_tensor] # calculate cost fwd_compute_cost = flop_mapping[torch.ops.aten.cudnn_batch_norm.default](fwd_in_args, fwd_out_args) bwd_compute_cost = flop_mapping[torch.ops.aten.cudnn_batch_norm_backward.default](bwd_in_args, bwd_out_args) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost # the fwd activation cost is output plus saved mean and saved inv std # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, output_tensor, mean_tensor, var_tensor]), parameter=activation_size([weight_tensor, bias_tensor]), temp=0, buffer=activation_size([mean_tensor, var_tensor])) # the bwd memory cost is quite tricky here, BatchNorm will remove saved mean # and saved inv std during backward phase bwd_memory_cost = MemoryCost(activation=activation_size([input_tensor]), parameter=activation_size([weight_tensor, bias_tensor]), temp=activation_size([mean_tensor, var_tensor]), buffer=activation_size([mean_tensor, var_tensor])) # total cost is the sum of forward and backward cost total_cost = MemoryCost(activation=fwd_memory_cost.activation + bwd_memory_cost.activation, parameter=fwd_memory_cost.parameter + bwd_memory_cost.parameter) memory_cost = TrainCycleItem(fwd=fwd_memory_cost, bwd=bwd_memory_cost, total=total_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [torch.zeros_like(input_tensor, device='meta')] fwd_buffer = [torch.zeros_like(mean_tensor, device='meta'), torch.zeros_like(var_tensor, device='meta')] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out @meta_register.register(torch.nn.LayerNorm) def layernorm_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """LayerNorm meta information Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ # construct needed tensors input_tensor = next(filter(lambda x: x.type == OperationDataType.ARG, args)).data output_tensor = next(filter(lambda x: x.type == OperationDataType.OUTPUT, args)).data weight_tensor = next(filter(lambda x: x.name == "weight", args)).data bias_tensor = next(filter(lambda x: x.name == "bias", args)).data running_mean = torch.rand(input_tensor.shape[0], 1, device='meta') running_var = torch.rand(input_tensor.shape[0], 1, device='meta') # construct args fwd_in_args = [input_tensor, [input_tensor.shape[0]], weight_tensor] fwd_out_args = [output_tensor] bwd_in_args = [input_tensor, output_tensor, [input_tensor.shape[0]]] bwd_out_args = [weight_tensor, bias_tensor] # compute cost fwd_compute_cost = flop_mapping[torch.ops.aten.native_layer_norm.default](fwd_in_args, fwd_out_args) bwd_compute_cost = flop_mapping[torch.ops.aten.native_layer_norm_backward.default](bwd_in_args, bwd_out_args) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # memory cost # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, output_tensor, weight_tensor, bias_tensor]), parameter=activation_size([weight_tensor, bias_tensor]), temp=0, buffer=activation_size([running_mean, running_var])) bwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, weight_tensor, bias_tensor]), parameter=activation_size([weight_tensor, bias_tensor]), temp=activation_size([running_mean, running_var]), buffer=activation_size([running_mean, running_var])) total_cost = MemoryCost(activation=fwd_memory_cost.activation + bwd_memory_cost.activation, parameter=fwd_memory_cost.parameter + bwd_memory_cost.parameter, temp=fwd_memory_cost.temp + bwd_memory_cost.temp, buffer=fwd_memory_cost.buffer + bwd_memory_cost.buffer) memory_cost = TrainCycleItem(fwd=fwd_memory_cost, bwd=bwd_memory_cost, total=total_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [torch.zeros_like(input_tensor, device='meta')] fwd_buffer = [torch.zeros_like(running_mean, device='meta'), torch.zeros_like(running_var, device='meta')] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out
from typing import List, Tuple import torch from colossalai.auto_parallel.tensor_shard.sharding_strategy import MemoryCost, OperationDataType, TrainCycleItem from colossalai.fx.profiler.memory_utils import activation_size from colossalai.fx.profiler.opcount import flop_mapping from ..constants import BCAST_FUNC_OP, NO_SAVE_ACTIVATION from ..registry import meta_register __all__ = ['binary_elementwise_meta_info'] @meta_register.register(BCAST_FUNC_OP) def binary_elementwise_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """Meta information generator for binary elementwise operations NOTE: Some of the binary elementwise operations will discard the input activation after computation, as they don't need those tensors for back propagation, for example, if there are two tensors being sent for `torch.add`, they will be discarded right after add operation is done. We create a simple API in `MetaInfo` class to identify this behavior, it is critical for better memory estimation. Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ input_op_data = [arg for arg in args if arg.type != OperationDataType.OUTPUT] output_op_data = next(filter(lambda arg: arg.type == OperationDataType.OUTPUT, args)) # construct forward args for flop mapping fwd_in_args = [opdata.data for opdata in input_op_data] fwd_out_args = [output_op_data.data] # calculate cost # calculate compute cost # NOTE: we set bwd_compute_cost two times of fwd_compute_cost in this case fwd_compute_cost = flop_mapping[torch.ops.aten.add.Tensor](fwd_in_args, fwd_out_args) bwd_compute_cost = fwd_compute_cost * 2 compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost param_mem_cost = activation_size([arg.data for arg in input_op_data if arg.type == OperationDataType.PARAM]) fwd_mem_cost = MemoryCost( activation=activation_size(output_op_data.data), parameter=param_mem_cost, ) bwd_mem_cost = MemoryCost( activation=activation_size(fwd_in_args), parameter=param_mem_cost, ) # total cost total_mem_cost = MemoryCost( activation=fwd_mem_cost.activation + bwd_mem_cost.activation, parameter=fwd_mem_cost.parameter + bwd_mem_cost.parameter, ) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [] fwd_buffer = [] fwd_out = [torch.zeros_like(output_op_data.data, device='meta')] return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out
from typing import List, Tuple import torch from colossalai.auto_parallel.tensor_shard.sharding_strategy import MemoryCost, OperationDataType, TrainCycleItem from colossalai.fx.profiler.memory_utils import activation_size from colossalai.fx.profiler.opcount import flop_mapping from ..registry import meta_register __all__ = ["avgpool_meta_info", "maxpool_meta_info"] @meta_register.register(torch.nn.AdaptiveAvgPool1d) @meta_register.register(torch.nn.AdaptiveAvgPool2d) @meta_register.register(torch.nn.AdaptiveAvgPool3d) @meta_register.register(torch.flatten) def avgpool_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """Meta info for AdaptiveAvgPool The aten graph of AdaptiveAvgPool is graph(): %input_2 : [#users=2] = placeholder[target=placeholder](default=) %_adaptive_avg_pool2d_default : [#users=1] = call_function[target=torch.ops.aten._adaptive_avg_pool2d.default](args = (%input_2, [None, None]), kwargs = {}) %zeros_like_default : [#users=1] = call_function[target=torch.ops.aten.zeros_like.default](args = (%_adaptive_avg_pool2d_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%input_2,), kwargs = {}) %_adaptive_avg_pool2d_backward_default : [#users=1] = call_function[target=torch.ops.aten._adaptive_avg_pool2d_backward.default](args = (%zeros_like_default, %detach_default), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%_adaptive_avg_pool2d_backward_default,), kwargs = {}) %detach_default_2 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_1,), kwargs = {}) Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ input_tensor = args[0].data output_tensor = next(filter(lambda x: x.type == OperationDataType.OUTPUT, args)).data is_inplace = kwargs.get("inplace", False) # construct forward args for flop mapping fwd_in_args = [input_tensor] fwd_out_args = [output_tensor] # construct backward args for flop mapping bwd_in_args = [output_tensor] bwd_out_args = [input_tensor] # calculate cost # the fwd op with compute cost is _adaptive_avg_pool2d.default # the bwd op with compute cost is _adaptive_avg_pool2d_backward.default # calculate compute cost fwd_compute_cost = flop_mapping[torch.ops.aten._adaptive_avg_pool2d.default](fwd_in_args, fwd_out_args) bwd_compute_cost = flop_mapping[torch.ops.aten._adaptive_avg_pool2d_backward.default](bwd_in_args, bwd_out_args) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost fwd_mem_cost = MemoryCost() if is_inplace else MemoryCost(activation=activation_size(output_tensor)) bwd_mem_cost = MemoryCost() if is_inplace else MemoryCost(activation=activation_size(input_tensor)) # total cost total_mem_cost = MemoryCost(activation=fwd_mem_cost.activation + bwd_mem_cost.activation) mem_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [] fwd_buffer = [] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, mem_cost, fwd_in, fwd_buffer, fwd_out @meta_register.register(torch.nn.MaxPool1d) @meta_register.register(torch.nn.MaxPool2d) @meta_register.register(torch.nn.MaxPool3d) def maxpool_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """Meta info for MaxPool The aten graph of MaxPool is graph(): %input_2 : [#users=2] = placeholder[target=placeholder](default=) %max_pool2d_with_indices_default : [#users=2] = call_function[target=torch.ops.aten.max_pool2d_with_indices.default](args = (%input_2, [None, None], [None, None]), kwargs = {}) %zeros_like_default : [#users=1] = call_function[target=torch.ops.aten.zeros_like.default](args = (%max_pool2d_with_indices_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%input_2,), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%max_pool2d_with_indices_default,), kwargs = {}) %max_pool2d_with_indices_backward_default : [#users=1] = call_function[target=torch.ops.aten.max_pool2d_with_indices_backward.default](args = (%zeros_like_default, %detach_default, [None, None], [None, None], [None, None], [None, None], None, %detach_default_1), kwargs = {}) %detach_default_2 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%max_pool2d_with_indices_backward_default,), kwargs = {}) %detach_default_3 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_2,), kwargs = {}) Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ input_tensor = next(filter(lambda x: x.type == OperationDataType.ARG, args)).data output_tensor = next(filter(lambda x: x.type == OperationDataType.OUTPUT, args)).data # construct forward args for flop mapping fwd_in_args = [input_tensor] fwd_out_args = [output_tensor] # construct backward args for flop mapping bwd_in_args = [output_tensor] bwd_out_args = [input_tensor] # construct index matrix index_matrix = torch.zeros_like(output_tensor, device="meta", dtype=torch.int64) # calculate cost # the fwd op with compute cost is max_pool2d_with_indices.default # the bwd op with compute cost is max_pool2d_with_indices_backward.default # calculate compute cost fwd_compute_cost = flop_mapping[torch.ops.aten.max_pool2d_with_indices.default](fwd_in_args, fwd_out_args) bwd_compute_cost = flop_mapping[torch.ops.aten.max_pool2d_with_indices_backward.default](bwd_in_args, bwd_out_args) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost # NOTE: the index matrix will be discarded in backward phase # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_mem_cost = MemoryCost(activation=activation_size([input_tensor, output_tensor, index_matrix])) # temp memory for backward is the index matrix to be discarded bwd_mem_cost = MemoryCost(activation=activation_size(input_tensor) - activation_size(index_matrix), temp=activation_size(index_matrix)) # total cost total_mem_cost = MemoryCost(activation=fwd_mem_cost.activation + bwd_mem_cost.activation, temp=bwd_mem_cost.temp) mem_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [torch.zeros_like(input_tensor, device='meta')] fwd_buffer = [torch.zeros_like(index_matrix, device='meta')] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, mem_cost, fwd_in, fwd_buffer, fwd_out
from .activation import * from .binary_elementwise_ops import * from .conv import * from .linear import * from .norm import * from .pooling import *
from typing import List, Tuple import torch from colossalai.auto_parallel.tensor_shard.sharding_strategy import MemoryCost, OperationDataType, TrainCycleItem from colossalai.fx.profiler.memory_utils import activation_size from colossalai.fx.profiler.opcount import flop_mapping from ..registry import meta_register __all__ = ["relu_meta_info"] @meta_register.register(torch.nn.ReLU) def relu_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """torch.nn.ReLU metainfo generator The aten graph of torch.nn.ReLU is graph(): %input_2 : [#users=1] = placeholder[target=placeholder](default=) %relu_default : [#users=2] = call_function[target=torch.ops.aten.relu.default](args = (%input_2,), kwargs = {}) %zeros_like_default : [#users=1] = call_function[target=torch.ops.aten.zeros_like.default](args = (%relu_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%relu_default,), kwargs = {}) %threshold_backward_default : [#users=1] = call_function[target=torch.ops.aten.threshold_backward.default](args = (%zeros_like_default, %detach_default, None), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%threshold_backward_default,), kwargs = {}) %detach_default_2 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_1,), kwargs = {}) Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ input_tensor = args[0].data output_tensor = next(filter(lambda x: x.type == OperationDataType.OUTPUT, args)).data is_inplace = kwargs.get("inplace", False) # construct input args for forward fwd_in_args = [input_tensor] # construct input args for backward bwd_in_args = [output_tensor] # calculate cost # the fwd op with compute cost is relu.default # the bwd op with compute cost is threshold_backward # calculate compute cost fwd_compute_cost = flop_mapping[torch.ops.aten.relu.default](fwd_in_args, (output_tensor,)) bwd_compute_cost = flop_mapping[torch.ops.aten.threshold_backward.default](bwd_in_args, (input_tensor,)) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost # NOTE: the inplace ReLU don't have forward memory cost # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_memory_cost = MemoryCost( activation=activation_size(input_tensor) if is_inplace else activation_size([output_tensor, input_tensor]), parameter=0, temp=0, buffer=0) bwd_memory_cost = MemoryCost(activation=activation_size(input_tensor), parameter=0, temp=0, buffer=0) # total cost is the sum of forward and backward cost total_cost = MemoryCost(activation=fwd_memory_cost.activation + bwd_memory_cost.activation, parameter=fwd_memory_cost.parameter + bwd_memory_cost.parameter) memory_cost = TrainCycleItem(fwd=fwd_memory_cost, bwd=bwd_memory_cost, total=total_cost) # store fwd_in, fwd_buffer, fwd_out # NOTE: It might seems a little bit weird here, we just want to align it with the older version # of MetaInfoProp. In the future we might modify this part to make it clearer. fwd_in = [] fwd_buffer = [torch.zeros_like(output_tensor, device='meta')] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out @meta_register.register(torch.nn.Softmax) @meta_register.register(torch.nn.functional.softmax) def softmax_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """torch.nn.Softmax metainfo generator Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ input_tensor = next( filter( lambda x: (x.type == OperationDataType.ARG or x.type == OperationDataType.PARAM) and x.name != 'softmax_dim', args)).data output_tensor = next(filter(lambda x: x.type == OperationDataType.OUTPUT, args)).data softmax_dim = next(filter(lambda x: x.name == 'softmax_dim', args)).data # calculate cost # calculate compute cost fwd_compute_cost = flop_mapping[torch.ops.aten._softmax.default]([input_tensor], [output_tensor]) bwd_compute_cost = flop_mapping[torch.ops.aten._softmax_backward_data.default]([output_tensor], [input_tensor]) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_memory_cost = MemoryCost(activation=activation_size([input_tensor, output_tensor]), parameter=0, temp=0, buffer=0) bwd_memory_cost = MemoryCost(activation=activation_size(input_tensor), parameter=0, temp=activation_size(input_tensor), buffer=0) # total cost is the sum of forward and backward cost total_cost = MemoryCost(activation=fwd_memory_cost.activation + bwd_memory_cost.activation, parameter=fwd_memory_cost.parameter + bwd_memory_cost.parameter, temp=fwd_memory_cost.temp + bwd_memory_cost.temp, buffer=fwd_memory_cost.buffer + bwd_memory_cost.buffer) memory_cost = TrainCycleItem(fwd=fwd_memory_cost, bwd=bwd_memory_cost, total=total_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [] fwd_buffer = [torch.zeros_like(output_tensor, device='meta')] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out
from typing import Callable, Dict, List, Tuple, Union import torch from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( MemoryCost, OperationData, OperationDataType, ShardingStrategy, StrategiesVector, TrainCycleItem, ) from colossalai.fx.profiler.memory_utils import activation_size from colossalai.fx.profiler.opcount import flop_mapping from colossalai.tensor.sharding_spec import ShardingSpec from ..registry import meta_register __all__ = ['convnd_meta_info'] @meta_register.register(torch.nn.Conv1d) @meta_register.register(torch.nn.Conv2d) @meta_register.register(torch.nn.Conv3d) @meta_register.register(torch.nn.functional.conv1d) @meta_register.register(torch.nn.functional.conv2d) @meta_register.register(torch.nn.functional.conv3d) def convnd_meta_info(*args, **kwargs) -> Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: """torch.nn.Conv1d, torch.nn.Conv2d, torch.nn.Conv3d meta info generator The atens graph of torch.nn.Convnd with bias is graph(): %input_2 : [#users=2] = placeholder[target=placeholder](default=) %convolution_default : [#users=1] = call_function[target=torch.ops.aten.convolution.default](args = (%input_2, None, None, [None, None, None], [None, None, None], [None, None, None], None, [None, None, None], None), kwargs = {}) %zeros_like_default : [#users=1] = call_function[target=torch.ops.aten.zeros_like.default](args = (%convolution_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%input_2,), kwargs = {}) %convolution_backward_default : [#users=3] = call_function[target=torch.ops.aten.convolution_backward.default](args = (%zeros_like_default, %detach_default, None, [None], [None, None, None], [None, None, None], [None, None, None], None, [None, None, None], None, [None, None, None]), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%convolution_backward_default,), kwargs = {}) %detach_default_2 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_1,), kwargs = {}) %detach_default_3 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%convolution_backward_default,), kwargs = {}) %detach_default_4 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_3,), kwargs = {}) %detach_default_5 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%convolution_backward_default,), kwargs = {}) %detach_default_6 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_5,), kwargs = {}) The atens graph of torch.nn.Convnd without bias is graph(): %input_2 : [#users=2] = placeholder[target=placeholder](default=) %convolution_default : [#users=1] = call_function[target=torch.ops.aten.convolution.default](args = (%input_2, None, None, [None, None], [None, None], [None, None], None, [None, None], None), kwargs = {}) %zeros_like_default : [#users=1] = call_function[target=torch.ops.aten.zeros_like.default](args = (%convolution_default,), kwargs = {dtype: None, layout: None, device: None, pin_memory: None}) %detach_default : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%input_2,), kwargs = {}) %convolution_backward_default : [#users=2] = call_function[target=torch.ops.aten.convolution_backward.default](args = (%zeros_like_default, %detach_default, None, [None], [None, None], [None, None], [None, None], None, [None, None], None, [None, None, None]), kwargs = {}) %detach_default_1 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%convolution_backward_default,), kwargs = {}) %detach_default_2 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_1,), kwargs = {}) %detach_default_3 : [#users=1] = call_function[target=torch.ops.aten.detach.default](args = (%convolution_backward_default,), kwargs = {}) %detach_default_4 : [#users=0] = call_function[target=torch.ops.aten.detach.default](args = (%detach_default_3,), kwargs = {}) Returns: Tuple[TrainCycleItem, TrainCycleItem, List[torch.Tensor]]: compute cost, memory cost and forward inputs """ has_bias: bool = False input_tensor = args[0].data output_tensor = next(filter(lambda x: x.type == OperationDataType.OUTPUT, args)).data if len(args) == 4: weight_tensors = [args[1].data, args[3].data] else: weight_tensors = [args[1].data] # check if conv has bias if len(weight_tensors) > 1: has_bias = True # bias tensor's shape only has one dimension if len(weight_tensors[0].shape) == 1: bias_tensor, weight_tensor = weight_tensors else: weight_tensor, bias_tensor = weight_tensors else: weight_tensor = weight_tensors[0] # construct input args for forward fwd_args = [None] * 9 # weight and input fwd_args[0] = input_tensor fwd_args[1] = weight_tensor fwd_args[2] = bias_tensor if has_bias else None # transpose indicator should be set to False fwd_args[6] = False # construct input args for backward bwd_args = [None] * 11 # weight and input bwd_args[0] = output_tensor bwd_args[1] = input_tensor bwd_args[2] = weight_tensor bwd_args[-1] = [True, True, True] if has_bias else [True, True, False] # calculate cost # the fwd op with compute cost is convolution.default # the bwd op with compute cost is convolution_backward.default # calculate compute cost fwd_compute_cost = flop_mapping[torch.ops.aten.convolution.default](fwd_args, (output_tensor,)) bwd_compute_cost = flop_mapping[torch.ops.aten.convolution_backward.default](bwd_args, (input_tensor, weight_tensor, bias_tensor)) if has_bias else \ flop_mapping[torch.ops.aten.convolution_backward.default](bwd_args, (input_tensor, weight_tensor)) compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) # calculate memory cost # TODO: use profiler to check conv temp memory # NOTE: currently in SPMD solver we always believe that there will be a new tensor created in forward fwd_memory_cost = MemoryCost( activation=activation_size([input_tensor, output_tensor]), parameter=activation_size([weight_tensor, bias_tensor]) if has_bias else activation_size(weight_tensor), temp=0, buffer=0) bwd_memory_cost = MemoryCost( activation=activation_size([input_tensor, weight_tensor, bias_tensor]) if has_bias else activation_size([input_tensor, weight_tensor]), parameter=activation_size([weight_tensor, bias_tensor]) if has_bias else activation_size(weight_tensor), temp=0, buffer=0) # total cost is the sum of forward and backward cost total_cost = MemoryCost(activation=fwd_memory_cost.activation + bwd_memory_cost.activation, parameter=fwd_memory_cost.parameter + bwd_memory_cost.parameter) memory_cost = TrainCycleItem(fwd=fwd_memory_cost, bwd=bwd_memory_cost, total=total_cost) # store fwd_in, fwd_buffer, fwd_out fwd_in = [torch.zeros_like(input_tensor, device='meta')] fwd_buffer = [] fwd_out = [torch.zeros_like(output_tensor, device='meta')] return compute_cost, memory_cost, fwd_in, fwd_buffer, fwd_out
from copy import deepcopy from typing import Dict, List import torch from torch.fx.node import Node from colossalai.auto_parallel.meta_profiler import MetaInfo from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, OperationData, OperationDataType, TrainCycleItem, ) from colossalai.device.device_mesh import DeviceMesh from colossalai.tensor.comm_spec import CommSpec from colossalai.tensor.shape_consistency import ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec shape_consistency_manager = ShapeConsistencyManager() def runtime_apply(node: Node, origin_dict: Dict, input_dict: Dict, node_index: int, user_node_index: int): """ This method will be invoked during runtime to do the shape consistency, which make sure the activations is converted into the user node expected form. """ origin_sharding_spec = origin_dict[node_index] target_sharding_spec = input_dict[node_index][user_node_index] return shape_consistency_manager.apply_for_autoparallel_runtime(node, origin_sharding_spec, target_sharding_spec) def runtime_apply_for_iterable_object(node: Node, origin_dict: Dict, input_dict: Dict, node_index: int, user_node_index: int): """ This method will be invoked during runtime to do the shape consistency, which makes sure the activations in type of tuple or list is converted into the user node expected form. """ rst = [] for index, (origin_sharding_spec, target_sharding_spec) in enumerate(zip(origin_dict[node_index], input_dict[node_index][user_node_index])): rst.append( shape_consistency_manager.apply_for_autoparallel_runtime(node[index], origin_sharding_spec, target_sharding_spec)) rst = type(node)(rst) return rst def runtime_comm_spec_apply(tensor: torch.Tensor, comm_actions_dict: Dict, node_index: int, op_data_name: str): """ This method will be invoked during runtime to apply the comm action following the instruction of comm spec. """ comm_action = comm_actions_dict[node_index][op_data_name] if isinstance(comm_action.comm_spec, CommSpec): rst = comm_action.comm_spec.covert_spec_to_action(tensor) else: origin_sharding_spec = comm_action.comm_spec['src_spec'] tgt_sharding_spec = comm_action.comm_spec['tgt_spec'] rst = shape_consistency_manager.apply_for_autoparallel_runtime(tensor, origin_sharding_spec, tgt_sharding_spec) return rst def _preprocess_graph(nodes: List[Node]): """ This method is used to extract all the placeholders with sharding information, and mapping the nodes into the index of the origin graph. """ # mapping the node into the origin graph index node_to_index_dict = {} index = 0 for node in nodes: if node.target == 'sharding_spec_convert_dict': input_dict_node = node continue if node.target == 'origin_node_sharding_spec_dict': origin_dict_node = node continue if node.target == 'comm_actions_dict': comm_actions_dict_node = node continue if not hasattr(node, 'best_strategy'): continue node_to_index_dict[node] = index index += 1 return input_dict_node, origin_dict_node, comm_actions_dict_node, node_to_index_dict def _shape_consistency_apply(gm: torch.fx.GraphModule): """ This pass is used to add the shape consistency node to the origin graph. """ mod_graph = gm.graph nodes = tuple(mod_graph.nodes) input_dict_node, origin_dict_node, _, node_to_index_dict = _preprocess_graph(nodes) for node in nodes: if not hasattr(node, 'best_strategy') or node.op == 'output': continue for user_node_index, user_node in enumerate(node.strategies_vector.successor_nodes): if isinstance(node.sharding_spec, (list, tuple)): assert isinstance( node.target_sharding_specs, (list, tuple)), 'target sharding specs should be tuple or list when node.sharding_spec is tuple or list' total_difference = 0 for sharding_spec, target_sharding_spec in zip(node.sharding_spec, node.target_sharding_specs[user_node_index]): total_difference += sharding_spec.sharding_sequence_difference(target_sharding_spec) if total_difference == 0: continue with mod_graph.inserting_before(user_node): shape_consistency_node = mod_graph.create_node('call_function', runtime_apply_for_iterable_object, args=(node, origin_dict_node, input_dict_node, node_to_index_dict[node], user_node_index)) else: assert isinstance(node.sharding_spec, ShardingSpec), 'node.sharding_spec should be type of ShardingSpec, tuple or list.' if node.sharding_spec.sharding_sequence_difference(node.target_sharding_specs[user_node_index]) == 0: continue with mod_graph.inserting_before(user_node): shape_consistency_node = mod_graph.create_node('call_function', runtime_apply, args=(node, origin_dict_node, input_dict_node, node_to_index_dict[node], user_node_index)) if 'activation_checkpoint' in user_node.meta: shape_consistency_node.meta['activation_checkpoint'] = user_node.meta['activation_checkpoint'] new_args = list(user_node.args) new_kwargs = dict(user_node.kwargs) # the origin node may be a positional argument or key word argument of user node if node in new_args: # substitute the origin node with shape_consistency_node origin_index_args = new_args.index(node) new_args[origin_index_args] = shape_consistency_node user_node.args = tuple(new_args) elif str(node) in new_kwargs: # substitute the origin node with shape_consistency_node new_kwargs[str(node)] = shape_consistency_node user_node.kwargs = new_kwargs return gm def _comm_spec_apply(gm: torch.fx.GraphModule): """ This pass is used to add the comm spec apply node to the origin graph. """ mod_graph = gm.graph nodes = tuple(mod_graph.nodes) _, _, comm_actions_dict_node, node_to_index_dict = _preprocess_graph(nodes) for node in nodes: if not hasattr(node, 'best_strategy') or node.op == 'output': continue comm_actions = node.best_strategy.communication_actions for op_data, comm_action in comm_actions.items(): if comm_action.comm_type == CommType.HOOK: continue if comm_action.comm_type == CommType.BEFORE: if op_data.type == OperationDataType.OUTPUT: comm_object = node elif comm_action.key_for_kwarg is not None: comm_object = node.kwargs[comm_action.key_for_kwarg] else: comm_object = node.args[comm_action.arg_index] with mod_graph.inserting_before(node): comm_spec_apply_node = mod_graph.create_node('call_function', runtime_comm_spec_apply, args=(comm_object, comm_actions_dict_node, node_to_index_dict[node], op_data.name)) # the origin node may be a positional argument or key word argument of user node if comm_action.key_for_kwarg is not None: # substitute the origin node with comm_spec_apply_node new_kwargs = dict(node.kwargs) new_kwargs[comm_action.key_for_kwarg] = comm_spec_apply_node node.kwargs = new_kwargs else: # substitute the origin node with comm_spec_apply_node new_args = list(node.args) new_args[comm_action.arg_index] = comm_spec_apply_node node.args = tuple(new_args) elif comm_action.comm_type == CommType.AFTER: with mod_graph.inserting_after(node): comm_spec_apply_node = mod_graph.create_node('call_function', runtime_comm_spec_apply, args=(node, comm_actions_dict_node, node_to_index_dict[node], op_data.name)) user_list = list(node.users.keys()) for user in user_list: if user == comm_spec_apply_node: continue new_args = list(user.args) new_kwargs = dict(user.kwargs) # the origin node may be a positional argument or key word argument of user node if node in new_args: # substitute the origin node with comm_spec_apply_node new_args[new_args.index(node)] = comm_spec_apply_node user.args = tuple(new_args) elif str(node) in new_kwargs: # substitute the origin node with comm_spec_apply_node new_kwargs[str(node)] = comm_spec_apply_node user.kwargs = new_kwargs if 'activation_checkpoint' in node.meta: comm_spec_apply_node.meta['activation_checkpoint'] = node.meta['activation_checkpoint'] return gm def _act_annotataion_pass(gm: torch.fx.GraphModule): """ This pass is used to add the act annotation to the new inserted nodes. """ mod_graph = gm.graph nodes = tuple(mod_graph.nodes) for node in nodes: if not hasattr(node.meta, 'activation_checkpoint'): from .runtime_preparation_pass import size_processing user_act_annotation = -1 input_act_annotation = -1 for user_node in node.users.keys(): if 'activation_checkpoint' in user_node.meta: user_act_annotation = user_node.meta['activation_checkpoint'] break for input_node in node._input_nodes.keys(): if 'activation_checkpoint' in input_node.meta: input_act_annotation = input_node.meta['activation_checkpoint'] break if user_act_annotation == input_act_annotation and user_act_annotation != -1: node.meta['activation_checkpoint'] = user_act_annotation return gm def runtime_apply_pass(gm: torch.fx.GraphModule): """ The method manages all the passes acting on the distributed training runtime. """ gm = _shape_consistency_apply(gm) gm = _comm_spec_apply(gm) return gm
import uuid from dataclasses import asdict from typing import List import torch import torch.fx from torch.fx import GraphModule from torch.fx.node import Node from colossalai.auto_parallel.meta_profiler import MetaInfo from colossalai.auto_parallel.passes.constants import OUTPUT_SAVED_MOD, OUTPUT_SAVED_OPS from colossalai.fx._compatibility import compatibility from colossalai.fx.profiler import GraphInfo def _normalize_tuple(x): if not isinstance(x, tuple): return (x,) return x @compatibility(is_backward_compatible=False) class MetaInfoProp: def __init__(self, module: GraphModule) -> None: self.module = module self.func_dict = { 'placeholder': self.placeholder_handler, 'get_attr': self.get_attr_handler, 'output': self.output_handler, 'call_function': self.node_handler, 'call_module': self.node_handler, 'call_method': self.node_handler, } def _set_data_ptr(self, x): """ Set uuid to tensor """ if isinstance(x, torch.Tensor): if not x.data_ptr(): data_ptr = uuid.uuid4() x.data_ptr = lambda: data_ptr def _is_inplace(self, node: Node): """ Check if the node is inplace operation. """ if node.op == 'call_module': return node.graph.owning_module.get_submodule(node.target).__class__ in OUTPUT_SAVED_MOD elif node.op == "call_function": return node.target in OUTPUT_SAVED_OPS return False def run(self) -> GraphModule: """ Run the meta information propagation pass on the module. """ for node in self.module.graph.nodes: node: Node self.func_dict[node.op](node) @compatibility(is_backward_compatible=False) def placeholder_handler(self, node: Node) -> None: """ Handle the placeholder node. """ graph_info = GraphInfo() out = _normalize_tuple(getattr(node, '_meta_data', None)) graph_info.fwd_out = list(out) if out[0] is not None else [] node.meta = {**asdict(graph_info)} @compatibility(is_backward_compatible=False) def get_attr_handler(self, node: Node) -> None: """ Handle the get_attr node. """ graph_info = GraphInfo() node.meta = {**asdict(graph_info)} @compatibility(is_backward_compatible=False) def output_handler(self, node: Node) -> None: """ Handle the output node. """ graph_info = GraphInfo() output_tensors = [] for par in node._input_nodes: if par.meta: output_tensors += par.meta["fwd_out"] graph_info.fwd_in = output_tensors node.meta = {**asdict(graph_info)} @compatibility(is_backward_compatible=False) def node_handler(self, node: Node) -> None: """ Handle other kind of nodes """ assert hasattr(node, 'best_metainfo'), f"Cannot find best_metainfo in node {node}, {node.op}" graph_info = GraphInfo() meta_info = node.best_metainfo meta_info: MetaInfo # set data_ptr for input_tensor in MetaInfo class input_tensors: List[torch.Tensor] = meta_info.fwd_in buffer_tensors: List[torch.Tensor] = meta_info.fwd_buffer output_tensors: List[torch.Tensor] = meta_info.fwd_out if self._is_inplace(node): # inplace operation will not create new tensor, and it only has one parent node # TODO: Verify this observation # set data_ptr for input_tensor, buffer_tensor and output_tensor of current node parent_node = list(node._input_nodes.keys())[0] parent_tensor = parent_node.meta.get("fwd_out")[0] parent_tensor: torch.Tensor for tensor in input_tensors: tensor.data_ptr = parent_tensor.data_ptr for tensor in buffer_tensors: tensor.data_ptr = parent_tensor.data_ptr for tensor in output_tensors: tensor.data_ptr = parent_tensor.data_ptr else: for par in node._input_nodes: # set data_ptr for the input_tensor of current node from the output_tensor of its parent node for tensor in par.meta.get("fwd_out", []): tensor: torch.Tensor target_input_tensor = next( (x for x in input_tensors if not x.data_ptr() and x.shape == tensor.shape), None) if target_input_tensor is not None: target_input_tensor.data_ptr = tensor.data_ptr # set data_ptr for tensor in input_tensor that is not set for tensor in input_tensors: if not tensor.data_ptr(): self._set_data_ptr(tensor) # set data_ptr for buffer_tensor for tensor in buffer_tensors: self._set_data_ptr(tensor) # set data_ptr for output_tensor for tensor in output_tensors: self._set_data_ptr(tensor) # attach them to graph_info graph_info.fwd_in = input_tensors graph_info.fwd_tmp = buffer_tensors graph_info.fwd_out = output_tensors # fetch other memory informations memory_cost = meta_info.memory_cost graph_info.fwd_mem_tmp = memory_cost.fwd.temp graph_info.fwd_mem_out = memory_cost.fwd.activation graph_info.bwd_mem_tmp = memory_cost.bwd.temp graph_info.bwd_mem_out = memory_cost.bwd.activation # fetch flop information # here we use fwd_time and bwd_time to deal with the case that # communication cost is a float compute_cost = meta_info.compute_cost graph_info.fwd_time = compute_cost.fwd graph_info.bwd_time = compute_cost.bwd node.meta = {**asdict(graph_info)}
import torch OUTPUT_SAVED_OPS = [torch.nn.functional.relu, torch.nn.functional.softmax, torch.flatten] OUTPUT_SAVED_MOD = [ torch.nn.ReLU, torch.nn.Softmax, ] # SHAPE_ARGUMENT_OPS contains node with (input, *shape) style args. # This list could be extended if any other method has the same # argument style as view and reshape. SHAPE_ARGUMENT_OPS = [torch.Tensor.view, torch.Tensor.reshape, torch.reshape]
from typing import Dict import torch from torch.fx import GraphModule from torch.fx.node import Node from colossalai.auto_parallel.meta_profiler import MetaInfo from colossalai.auto_parallel.passes.runtime_apply_pass import runtime_apply, runtime_comm_spec_apply from colossalai.auto_parallel.tensor_shard.sharding_strategy import MemoryCost, TrainCycleItem from colossalai.tensor.comm_spec import CommSpec from colossalai.tensor.shape_consistency import ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec shape_consistency_manager = ShapeConsistencyManager() def _construct_meta_info(node: Node, origin_sharding_spec: ShardingSpec, target_sharding_spec: ShardingSpec) -> MetaInfo: # get comm_action_sequence and total_cost from shape_consistency_manager _, comm_action_sequence, total_cost = shape_consistency_manager.shape_consistency( origin_sharding_spec, target_sharding_spec) meta_info = MetaInfo() # NOTE: the cost in shape_consistency_manager.mem_cost is the count in number of numel # get mem cost for MetaInfo mem_cost = shape_consistency_manager.mem_cost(comm_action_sequence) # extract user that has _meta_data and extract element length input_node = next(n for n in node._input_nodes if hasattr(n, '_meta_data')) element_length = input_node._meta_data.element_size() mem_cost.fwd.activation *= element_length mem_cost.fwd.temp *= element_length mem_cost.bwd.activation *= element_length mem_cost.bwd.temp *= element_length mem_cost.total.activation *= element_length meta_info.memory_cost = mem_cost # get computation cost for MetaInfo meta_info.compute_cost = TrainCycleItem(total_cost['forward'] * element_length, total_cost['backward'] * element_length, total_cost['total'] * element_length) # get tensor shape for MetaInfo origin_sharding_spec: ShardingSpec target_sharding_spec: ShardingSpec input_shape = origin_sharding_spec.get_sharded_shape_per_device() output_shape = target_sharding_spec.get_sharded_shape_per_device() meta_info.fwd_in = [torch.rand(input_shape, device='meta')] meta_info.fwd_buffer = [] meta_info.fwd_out = [torch.rand(output_shape, device='meta')] return meta_info def _runtime_apply_meta_info(node: Node, origin_spec_dict, sharding_spec_dict) -> MetaInfo: """ This method is used to construct `MetaInto` for shape consistency node """ # extract node index and user node index args = node.args node_index, user_node_index = args[3], args[4] origin_sharding_spec, target_sharding_spec = origin_spec_dict[node_index], sharding_spec_dict[node_index][ user_node_index] return _construct_meta_info(node, origin_sharding_spec, target_sharding_spec) def _runtime_comm_spec_apply_meta_info(node: Node, comm_actions_dict: Dict) -> MetaInfo: # extract node_index and op_data_name node_index, op_data_name = node.args[2], node.args[3] comm_action = comm_actions_dict[node_index][op_data_name] if isinstance(comm_action.comm_spec, CommSpec): # this case is for all_reduce, there will be no memory cost meta_info = MetaInfo() meta_info.memory_cost = TrainCycleItem(MemoryCost(), MemoryCost(), MemoryCost) output_node = next(n for n in node.users if hasattr(n, '_meta_data')) element_length = output_node._meta_data.element_size() total_cost = comm_action.comm_spec.get_comm_cost() meta_info.compute_cost = TrainCycleItem(total_cost['forward'] * element_length, total_cost['backward'] * element_length, total_cost['total'] * element_length) input_shape = output_shape = comm_action.comm_spec.sharding_spec.get_sharded_shape_per_device() meta_info.fwd_in = [torch.rand(input_shape, device='meta')] meta_info.fwd_buffer = [] meta_info.fwd_out = [torch.rand(output_shape, device='meta')] else: # this case will be handled by shape consistency manager origin_sharding_spec, target_sharding_spec = comm_action.comm_spec['src_spec'], comm_action.comm_spec[ 'tgt_spec'] meta_info = _construct_meta_info(node, origin_sharding_spec, target_sharding_spec) return meta_info def comm_metainfo_pass(gm: GraphModule, sharding_spec_dict: Dict, origin_spec_dict: Dict, comm_actions_dict: Dict) -> GraphModule: """ The method manages all the metainfo of the communication node (run_time_apply, runtime_comm_spec_apply) in the graph. """ for node in gm.graph.nodes: if node.target == runtime_apply: setattr(node, 'best_metainfo', _runtime_apply_meta_info(node, origin_spec_dict, sharding_spec_dict)) elif node.target == runtime_comm_spec_apply: setattr(node, 'best_metainfo', _runtime_comm_spec_apply_meta_info(node, comm_actions_dict)) else: pass return gm
import operator from copy import deepcopy from typing import Dict, List, Union import torch from torch.fx import symbolic_trace from torch.fx.node import Node from colossalai.auto_parallel.tensor_shard.constants import RESHAPE_FUNC_OP from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, OperationDataType, ShardingStrategy, ) from colossalai.auto_parallel.tensor_shard.solver.strategies_constructor import StrategiesConstructor from colossalai.device.device_mesh import DeviceMesh from colossalai.tensor.comm_spec import _all_reduce from colossalai.tensor.shape_consistency import ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec from .constants import SHAPE_ARGUMENT_OPS shape_consistency_manager = ShapeConsistencyManager() def size_processing(size: Union[int, torch.Size], dim_partition_dict: Dict[int, List[int]], device_mesh_info: Dict[int, int], target_dim: int = None, node_name: str = None): """ This method will be invoked during runtime to convert size node value depending on distributed information. """ if target_dim is not None: assert isinstance(size, int) if target_dim in dim_partition_dict: total_shard_size = 1 for shard_dim in dim_partition_dict[target_dim]: total_shard_size *= device_mesh_info[shard_dim] size = size * total_shard_size else: size = list(size) for dim, dim_size in enumerate(size): if dim in dim_partition_dict: total_shard_size = 1 for shard_dim in dim_partition_dict[dim]: total_shard_size *= device_mesh_info[shard_dim] size[dim] = dim_size * total_shard_size size = torch.Size(size) return size def solution_annotatation_pass(gm: torch.fx.GraphModule, solution: List[int], strategies_constructor: StrategiesConstructor): """ This method is used to stick the solution strategy to the nodes and add the information required in runtime into graph as placeholder nodes. """ mod_graph = gm.graph nodes = [strategies_vector.node for strategies_vector in strategies_constructor.leaf_strategies] no_strategy_nodes = strategies_constructor.no_strategy_nodes # the dict to get origin sharding spec of node origin_node_sharding_spec_dict = {} for node_index, (node, strategy_index) in enumerate(zip(nodes, solution)): strategies_vector = node.strategies_vector # stick the solution strategy to the corresponding node setattr(node, 'best_strategy', strategies_vector[strategy_index]) setattr(node, 'sharding_spec', strategies_vector[strategy_index].get_sharding_spec_by_name(str(node))) origin_node_sharding_spec_dict[node_index] = strategies_vector[strategy_index].get_sharding_spec_by_name( str(node)) # attach the corresponding metainfo if node has the attribute `metainfo_vector` if hasattr(node, 'metainfo_vector'): setattr(node, 'best_metainfo', node.metainfo_vector[strategy_index]) # the dict to get input sharding specs of user node sharding_spec_convert_dict = {} # the dict to record comm actions of nodes comm_actions_dict = {} for index, node in enumerate(nodes): target_sharding_specs = [] for user_node in node.strategies_vector.successor_nodes: if user_node in no_strategy_nodes: target_sharding_spec = node.best_strategy.get_sharding_spec_by_name(str(node.name)) else: target_sharding_spec = user_node.best_strategy.get_sharding_spec_by_name(str(node.name)) target_sharding_specs.append(target_sharding_spec) sharding_spec_convert_dict[index] = target_sharding_specs setattr(node, 'target_sharding_specs', target_sharding_specs) # the get_attr node strategy is kind of pending strategy, which means we will change it # to the same strategy of the user node. if node.op == 'get_attr': assert len(target_sharding_specs) == 1, f'sharing weight is not supported in current version.' target_node = node.strategies_vector.successor_nodes[0] node_name = str(node) if target_node.op == 'call_function' and target_node.target in RESHAPE_FUNC_OP: node_name = str(target_node) target_node = target_node.strategies_vector.successor_nodes[0] user_strategy = target_node.best_strategy op_data_in_user = user_strategy.get_op_data_by_name(node_name) origin_pending_strategy = node.best_strategy origin_op_data = origin_pending_strategy.get_op_data_by_name(str(node)) new_communication_actions = {} if op_data_in_user in user_strategy.communication_actions: new_communication_action = user_strategy.communication_actions.pop(op_data_in_user) new_communication_action.arg_index = 0 new_communication_actions[origin_op_data] = new_communication_action node.best_strategy.communication_actions = new_communication_actions comm_action_dict = {} for op_data, comm_action in node.best_strategy.communication_actions.items(): comm_action_dict[op_data.name] = comm_action comm_actions_dict[index] = comm_action_dict # add above dicts into graph for node in nodes: if node.op != 'placeholder': with mod_graph.inserting_before(node): input_specs_node = mod_graph.create_node('placeholder', target='sharding_spec_convert_dict') origin_specs_node = mod_graph.create_node('placeholder', target='origin_node_sharding_spec_dict') comm_actions_dict_node = mod_graph.create_node('placeholder', target='comm_actions_dict') break return gm, sharding_spec_convert_dict, origin_node_sharding_spec_dict, comm_actions_dict def size_value_converting_pass(gm: torch.fx.GraphModule, device_mesh: DeviceMesh): """ In the auto parallel system, tensors may get shard on different devices, so the size of tensors need to be converted to the size of original tensor and managed by the users, such as torch.view, torch.reshape, etc. These nodes have enough information like input sharding_spec and output sharding_spec to decide how to convert the size value. """ mod_graph = gm.graph nodes = tuple(mod_graph.nodes) node_pairs = {} # DeviceMesh information instructs the scaling of the size value device_mesh_info = {} for dim, dim_size in enumerate(device_mesh.mesh_shape): device_mesh_info[dim] = dim_size def _extract_target_dim(node): ''' A helper function to etract the target dimension from size node. There are two usages of torch.Tensor.size: 1. tensor.size() 2. tensor.size(dim) If a target_dim is assigned, then the output will be in type of int, instead of torch.Size. Otherwise, the output will be in type of torch.Size and this function will return None. ''' target_dim = None if len(node.args) > 1: target_dim = node.args[1] if target_dim < 0: target_dim += node.args[0]._meta_data.dim() return target_dim def _post_processing(node, size_processing_node): ''' This function is used to process the dependency between the size node and its users after inserting the size_process_node. ''' # store original node and processing node pair in node_pairs dictioanry # It will be used to replace the original node with processing node in slice object node_pairs[node] = size_processing_node size_processing_node._meta_data = node._meta_data if 'activation_checkpoint' in node.meta: size_processing_node.meta['activation_checkpoint'] = node.meta['activation_checkpoint'] user_list = list(node.users.keys()) for user in user_list: if user == size_processing_node: continue new_args = list(user.args) new_kwargs = dict(user.kwargs) # the origin node may be a positional argument or key word argument of user node if node in new_args: # substitute the origin node with size_processing_node new_args[new_args.index(node)] = size_processing_node user.args = tuple(new_args) elif str(node) in new_kwargs: # substitute the origin node with size_processing_node new_kwargs[str(node)] = size_processing_node user.kwargs = new_kwargs def _update_slice_object_args(slice_object): ''' This function is used to update the slice object argument list. If the slice object contains the Node argument, then the size node will be replaced with ''' if isinstance(slice_object, slice): start = slice_object.start stop = slice_object.stop step = slice_object.step if start in node_pairs: start = node_pairs[start] if stop in node_pairs: stop = node_pairs[stop] if step in node_pairs: step = node_pairs[step] return slice(start, stop, step) elif isinstance(slice_object, int): if slice_object in node_pairs: return node_pairs[slice_object] else: return slice_object else: raise RuntimeError(f"Unsupported slice object type: {type(slice_object)}") for node in nodes: if node.op == 'call_method' and node.target == 'size': # extract useful information from size node # dim_partition_dict will instruct the size value on which # dimension should be enlarged. sharding_spec = node.args[0].sharding_spec dim_partition_dict = sharding_spec.dim_partition_dict target_dim = _extract_target_dim(node) # insert size_processing node with mod_graph.inserting_after(node): size_processing_node = mod_graph.create_node('call_function', size_processing, args=(node, dim_partition_dict, device_mesh_info, target_dim, node.name)) _post_processing(node, size_processing_node) if node.op == 'call_function' and node.target == operator.getitem: getitem_index = node.args[1] # slice object is quite special in torch.fx graph, # On one side, we treat slice object same as type of int, # so we do not create a node for slice object. On the other side, # slice object could take fx.Node as its argument. And the user # relationship cannot be tracked in fx graph. # Therefore, I record the node_pairs in this pass, and use the it # to replace the original node argument inside the slice object if # it has been processed in above pass. # There are three main usages of operator.getitem: # getitem(input, int) # getitem(input, slice) # getitem(input, Tuple[slice]) # In this pass, we need process the last two cases because # node arguments may potentially appear in these cases. if isinstance(getitem_index, slice): new_slice_item = _update_slice_object_args(getitem_index) new_args = (node.args[0], new_slice_item) node.args = new_args elif isinstance(getitem_index, (tuple, list)): if not isinstance(getitem_index[0], slice): continue new_slice_items = [] for slice_item in getitem_index: if slice_item is None: new_slice_items.append(None) continue new_slice_item = _update_slice_object_args(slice_item) new_slice_items.append(new_slice_item) new_args = (node.args[0], tuple(new_slice_items)) node.args = new_args return gm def node_args_converting_pass(gm: torch.fx.GraphModule, device_mesh: DeviceMesh): """ This pass will process node args to adapt the distributed tensor layout. """ mod_graph = gm.graph nodes = tuple(mod_graph.nodes) def _extract_info_from_sharding_spec(sharding_spec): ''' This function is used to extract the dim_partition_dict and device_mesh from sharding spec instance or a list of sharding spec. ''' if isinstance(sharding_spec, ShardingSpec): dim_partition_dict = sharding_spec.dim_partition_dict device_mesh = sharding_spec.device_mesh return dim_partition_dict, device_mesh if sharding_spec is None: return None, None assert isinstance(sharding_spec, (tuple, list)), 'sharding_spec should be type of ShardingSpec, tuple, list or None' device_mesh = sharding_spec[0].device_mesh dim_partition_dict = [] for element in sharding_spec: dim_partition_dict.append(_extract_info_from_sharding_spec(element)) return dim_partition_dict, sharding_spec def _process_node_arguments(node): new_args = [] for arg in node.args: # There are two args style: # 1. (input, *shape) # 2. (input, shape) # We will extract the elements from shape and add them into the new_args # Finally, the args style of new_args will be unified to (input, *shape) if isinstance(arg, Node): if isinstance(arg._meta_data, (tuple, list)): new_args.extend(arg._meta_data) elif isinstance(arg._meta_data, int): new_args.append(arg._meta_data) else: new_args.append(arg) else: assert isinstance(arg, (int, tuple, list)), 'The argument in view node should be either type of Node or int.' if isinstance(arg, (tuple, list)): new_args.extend(arg) else: new_args.append(arg) return new_args def _scale_args_adapt_sharding_spec(dim_partition_dict, device_mesh, node): new_args = _process_node_arguments(node) if node.op == 'call_method': args_to_process = list(new_args[1:]) else: args_to_process = list(new_args) for dim, shard_dims in dim_partition_dict.items(): total_shard_size = 1 for shard_dim in shard_dims: total_shard_size *= device_mesh.shape[shard_dim] # we will skip the dim with -1 value if args_to_process[dim] == -1: continue else: # TODO: add assertion here to make sure the dim size is divisible by total_shard_size args_to_process[dim] //= total_shard_size args_to_process = tuple(args_to_process) if node.op == 'call_method': new_args = (new_args[0],) + args_to_process else: new_args = args_to_process node.args = new_args def _filter_node_with_shape_args(node): if node.op == 'call_method': target = getattr(node.args[0]._meta_data.__class__, node.target) elif node.op == 'call_function': target = node.target else: target = None if target in SHAPE_ARGUMENT_OPS: return True return False for node in nodes: # skip the placeholder node added in _solution_annotation pass if not hasattr(node, 'sharding_spec'): continue output_dim_partition_dict, device_mesh = _extract_info_from_sharding_spec(node.sharding_spec) if _filter_node_with_shape_args(node): _scale_args_adapt_sharding_spec(output_dim_partition_dict, device_mesh, node) return gm def module_params_sharding_pass(gm: torch.fx.GraphModule, device_mesh: DeviceMesh, overlap=False): """ Apply the sharding action to the module parameters and buffers following the instructions of solver solution. """ mod_graph = gm.graph nodes = tuple(mod_graph.nodes) # This stream is created for overlaping the communication and computation. reduction_stream = torch.cuda.Stream() def _add_hook_for_grad_communication(node, param): comm_actions = node.best_strategy.communication_actions def _filter_param_to_hook(node, op_data, comm_action): if node.op == 'call_module' and op_data.type == OperationDataType.PARAM and op_data.name == param.name and comm_action.comm_type == CommType.HOOK: return True if node.op == 'get_attr' and isinstance( node._meta_data, torch.nn.parameter.Parameter) and comm_action.comm_type == CommType.HOOK: return True return False for operation_data, comm_action in comm_actions.items(): comm_spec_to_use = comm_action.comm_spec # register hook to the parameters if _filter_param_to_hook(node, operation_data, comm_action): def wrapper(param, comm_spec, stream, overlap): def hook_fn(grad): if overlap: with torch.cuda.stream(stream): _all_reduce(grad, comm_spec, async_op=True) else: _all_reduce(grad, comm_spec, async_op=False) param.register_hook(hook_fn) wrapper(param, comm_spec_to_use, reduction_stream, overlap=overlap) def _shard_param(param, target_sharding_spec): # apply the sharding spec of parameters if target_sharding_spec.dim_partition_dict != {}: origin_sharding_spec = ShardingSpec(device_mesh, param.shape, {}) setattr(param, 'sharding_spec', origin_sharding_spec) # TODO: build a ColoParamter class to manager the distributed parameters # we could use .data here, because all the operations just happen before the real training # loop, so we don't need to track these operations in the autograd graph. param = torch.nn.Parameter( shape_consistency_manager.apply_for_autoparallel_runtime(param.data, param.sharding_spec, target_sharding_spec).detach().clone()) return param for node in nodes: if node.op == 'call_module': target_module = node.graph.owning_module.get_submodule(node.target) # TODO: we need to do more actions to take care of the shared parameters. if hasattr(target_module, 'processed') and target_module.processed: continue setattr(target_module, 'processed', True) for name, param in target_module.named_parameters(): target_sharding_spec = node.best_strategy.get_sharding_spec_by_name(name) param = _shard_param(param, target_sharding_spec) setattr(target_module, name, param) _add_hook_for_grad_communication(node, param) sharded_buffer_dict = {} # apply the sharding spec of buffers for name, buffer in target_module.named_buffers(): origin_sharding_spec = ShardingSpec(device_mesh, buffer.shape, {}) setattr(buffer, 'sharding_spec', origin_sharding_spec) target_sharding_spec = node.best_strategy.get_sharding_spec_by_name(name) buffer_sharded = shape_consistency_manager.apply(buffer, target_sharding_spec) sharded_buffer_dict[name] = buffer_sharded for name, buffer_sharded in sharded_buffer_dict.items(): setattr(target_module, name, buffer_sharded.detach().clone()) if node.op == 'get_attr': root = node.graph.owning_module atoms = node.target.split(".") attr_len = len(atoms) if attr_len == 1: target_module = root target = getattr(root, atoms[0]) else: target_module = root for atom in atoms[:-1]: target_module = getattr(target_module, atom) target = getattr(target_module, atoms[-1]) target_sharding_spec = node.sharding_spec target = _shard_param(target, target_sharding_spec) assert hasattr(target_module, atoms[-1]) setattr(target_module, atoms[-1], target) _add_hook_for_grad_communication(node, target) return gm def implicit_comm_action_apply(gm: torch.fx.GraphModule): """ replace the origin kernel into kernel with implicit communication inside. """ pass def runtime_preparation_pass(gm: torch.fx.GraphModule, solution: List[int], device_mesh: DeviceMesh, strategies_constructor: StrategiesConstructor, overlap=False): gm, sharding_spec_convert_dict, origin_node_sharding_spec_dict, comm_actions_dict = solution_annotatation_pass( gm, solution, strategies_constructor) gm = size_value_converting_pass(gm, device_mesh) gm = node_args_converting_pass(gm, device_mesh) # TODO: the pass below should be uncommented after the implementation of implicit_comm_action_apply_pass completed. # gm = implicit_comm_action_apply(gm) gm = module_params_sharding_pass(gm, device_mesh, overlap=overlap) return gm, sharding_spec_convert_dict, origin_node_sharding_spec_dict, comm_actions_dict
from dataclasses import dataclass from enum import Enum __all__ = ['SolverOptions', 'SolverPerference', 'DataloaderOption', 'ShardOption'] class SolverPerference(Enum): """ This enum class is to define the solver preference. """ STANDARD = 0 DP = 1 TP = 2 class ShardOption(Enum): """ This enum class is to define the shard level required in node strategies. Notes: STANDARD: We do not add any extra shard requirements. SHARD: We require the node to be shard using at least one device mesh axis. SHARD_ONE_AXIS: We require the node to be shard using the last device mesh axis. FULL_SHARD: We require the node to be shard using all device mesh axes. TP_SHARD: We require the node to be shard using tensor parallel strategies on last device mesh axis. TP_FULL_SHARD: We require the node to be shard using tensor parallel strategies on all device mesh axes. """ STANDARD = 0 SHARD = 1 SHARD_LAST_AXIS = 2 FULL_SHARD = 3 class DataloaderOption(Enum): """ This enum class is to define the dataloader option. """ REPLICATED = 0 DISTRIBUTED = 1 @dataclass class SolverOptions: """ SolverOptions is a dataclass used to configure the preferences for the parallel execution plan search. """ solver_perference: SolverPerference = SolverPerference.STANDARD dataloader_option: DataloaderOption = DataloaderOption.REPLICATED shard_option: ShardOption = ShardOption.STANDARD
from copy import deepcopy from dataclasses import dataclass from enum import Enum from typing import Any, Dict, List, Tuple, Union import torch from torch.fx.node import Node from colossalai.tensor.comm_spec import CommSpec from colossalai.tensor.sharding_spec import ShardingSpec from .constants import ( BCAST_FUNC_OP, ELEMENTWISE_FUNC_OP, ELEMENTWISE_METHOD_OP, ELEMENTWISE_MODULE_OP, RESHAPE_FUNC_OP, RESHAPE_METHOD_OP, ) __all__ = ['OperationDataType', 'OperationData', 'TrainCycleItem', 'MemoryCost', 'ShardingStrategy', 'StrategiesVector'] class OperationDataType(Enum): """ An operation can come from the argument list of an operator or the parameter list of a module. """ INPUT = 0 ARG = 1 PARAM = 2 BUFFER = 3 OUTPUT = 4 @dataclass class OperationData: """ OperationData is the data related to an operator, the data can be the operand or the output. Args: name (str): the name of the operation-related data type (OperationDataType): the type of the operation data data (Any): the value for this data, usually it is a meta tensor. logical_shape (Tuple[int]): the logical shape of the data, it can be different from the its actual shape in memory. """ name: str type: OperationDataType data: Any logical_shape: Tuple[int] = None def __post_init__(self): # if no logical shape is specified, use the data shape as the logical shape if self.logical_shape is None: def _infer_logical_shape(data: any): """ This function is used to infer the logical shape of the data. """ if isinstance(data, torch.Tensor): return data.shape elif isinstance(data, torch.Size): return None elif isinstance(data, (tuple, list)): data_type = type(data) return data_type([_infer_logical_shape(d) for d in data]) else: return None self.logical_shape = _infer_logical_shape(self.data) def __repr__(self) -> str: return f'OperationData(name={self.name}, type={self.type})' def __eq__(self, other) -> bool: return other.name == self.name def __hash__(self) -> int: return hash(f'{self.name}') @dataclass class TrainCycleItem: """ TrainCycleItem is a dataclass to store the items which have different values for the forward and backward pass in a training iteration. Args: fwd (float): the item for the forward pass bwd (float): the item for the backward pass """ fwd: Any bwd: Any total: Any @dataclass class MemoryCost: """ MemoryCost is a dataclass which stores the memory usage in the program. Args: activation (int): the memory cost incurred by the activations in bytes. parameter (int): the memory cost incurred by the module parameter in bytes. temp (int): the memory cost incurred by the temporary tensors in bytes. buffer (int): the memory cost incurred by the module buffer in bytes. """ activation: int = 0 parameter: int = 0 temp: int = 0 buffer: int = 0 class CommType(Enum): """ CommType describes the sequential order of a communication action and a computation action. Meaning: BEFORE: the communication action happens just before the computation operation. AFTER: the communication action happens after the computation operation. HOOK: the communication action is used to do the grad all reduce. IMPLICIT: the communication action happens during the kernel execution, such as SyncBatchNorm """ BEFORE = 0 AFTER = 1 HOOK = 2 IMPLICIT = 3 @dataclass class CommAction: """ CommAction is used to record the communication action. Args: comm_spec: express the communication pattern and the process groups to execute the communication action. comm_type: describes the sequential order of a communication action and a computation action. arg_index: record the location of tensor which join the communication, we cannot use name of node or op_data at runtime, because the args of node may be changed by graph transform passes. """ comm_spec: CommSpec = None comm_type: CommType = None arg_index: int = -1 key_for_kwarg: any = None @dataclass class ShardingStrategy: """ ShardingStrategy is a dataclass to store the meta information on tensor sharding for a node. Args: name (str): express the sharding strategies in string, such as 'S0S1 = S0R x RS1'. output_sharding_spec (ShardingSpec): ShardingSpec of the output node. compute_cost (TrainCycleItem): Computation cost to complete this strategy. (default to None) communication_cost (TrainCycleItem): Communication cost to complete this strategy. (default to None) memory_cost (TrainCycleItem): Memory cost of the output node using this strategy. (default to None) input_sharding_specs (List(ShardingSpec)): The ShardingSpecs of the input nodes. """ name: str sharding_specs: Dict[OperationData, Union[ShardingSpec, Tuple[ShardingSpec]]] = None compute_cost: TrainCycleItem = None communication_cost: TrainCycleItem = None memory_cost: TrainCycleItem = None communication_actions: Dict[OperationData, CommAction] = None resharding_costs: Dict[Node, List[TrainCycleItem]] = None @property def input_sharding_specs(self) -> Dict[OperationData, ShardingSpec]: specs = {} specs.update(self._get_sharding_spec(OperationDataType.ARG)) specs.update(self._get_sharding_spec(OperationDataType.PARAM)) return specs @property def argument_sharding_specs(self) -> Dict[OperationData, ShardingSpec]: return self._get_sharding_spec(OperationDataType.ARG) @property def param_sharding_specs(self) -> Dict[OperationData, ShardingSpec]: return self._get_sharding_spec(OperationDataType.PARAM) @property def output_sharding_specs(self) -> Dict[OperationData, ShardingSpec]: return self._get_sharding_spec(OperationDataType.OUTPUT) def _get_sharding_spec(self, operation_data_type: OperationDataType): specs = {k: v for k, v in self.sharding_specs.items() if k.type == operation_data_type} return specs def get_op_data_by_name(self, name: str): for op_data in self.sharding_specs.keys(): if op_data.name == name: return op_data raise KeyError(f"Could not find the OperationData with name {name}") def get_sharding_spec_by_name(self, name: str): for op_data, sharding_spec in self.sharding_specs.items(): if op_data.name == name: return sharding_spec raise KeyError(f"Could not find the ShardingSpec for OperationData with name {name}") def clone(self): def _deepcopy_dict_vals(data: Dict): return {k: deepcopy(v) for k, v in data.items()} sharding_specs = _deepcopy_dict_vals(self.sharding_specs) if self.sharding_specs is not None else None # We need to deepcopy it when self.communication_actions is not None, instead of checking its __bool__ value. # Consider the examples below: # If self.communication_actions is an empty dictionary {}, then self.communication_actions is not None, but its __bool__ value is False. # In this case, if we set None to the new object, program will crash when we try to access the communication_actions.items. communication_actions = _deepcopy_dict_vals( self.communication_actions) if self.communication_actions is not None else None # same reason as communication_actions resharding_costs = _deepcopy_dict_vals(self.resharding_costs) if self.resharding_costs is not None else None compute_cost = deepcopy(self.compute_cost) communication_cost = deepcopy(self.communication_cost) memory_cost = deepcopy(self.memory_cost) return ShardingStrategy(name=self.name, sharding_specs=sharding_specs, compute_cost=compute_cost, communication_cost=communication_cost, memory_cost=memory_cost, communication_actions=communication_actions, resharding_costs=resharding_costs) class StrategiesVector(list): ''' Each node in fx graph will have a corresponding StrategiesVector, to store all the possible strategies of the node. Argument: node (Node): node for which the list of sharding strategies are generated. ''' def __init__(self, node: Node): super().__init__() self.node = node # fetch its input and output nodes # TODO: placeholder input nodes self.predecessor_nodes = list(node._input_nodes.keys()) self.successor_nodes = list(node.users.keys()) def check_merge(self): merge_label = False if self.node.op == 'call_module': target = self.node.target root_module = self.node.graph.owning_module submod = root_module.get_submodule(target) submod_type = type(submod) # merge elementwise module node into source nodes # we could merge element-wise op, because the output sharding spec is always same as the input sharding spec. if submod_type in ELEMENTWISE_MODULE_OP: merge_label = True if self.node.op == 'call_function': # we could merge element-wise op, because the output sharding spec is always same as the input sharding spec. if self.node.target in ELEMENTWISE_FUNC_OP: merge_label = True # we could merge bcast op if the rhs is a scalar, because it will fall back to the element-wise case. # TODO: remove this after we support the fall back logic. # if self.node.target in BCAST_FUNC_OP and len(self.predecessor_nodes) == 1: # merge_label = True # we could merge reshape op, because their computation costs are negligible. if self.node.target in RESHAPE_FUNC_OP: merge_label = True if self.node.op == 'call_method': # we could merge reshape op, because their computation costs are negligible. method = getattr(self.node.args[0]._meta_data.__class__, self.node.target) if method in RESHAPE_METHOD_OP: merge_label = True if method in ELEMENTWISE_METHOD_OP: merge_label = True return merge_label
from typing import Dict, List, Tuple import torch import torch.distributed as dist import torch.nn as nn from torch.fx import GraphModule from torch.fx.graph import Graph from colossalai.auto_parallel.passes.runtime_apply_pass import runtime_apply_pass from colossalai.auto_parallel.passes.runtime_preparation_pass import runtime_preparation_pass from colossalai.auto_parallel.tensor_shard.options import DataloaderOption, ShardOption, SolverOptions, SolverPerference from colossalai.auto_parallel.tensor_shard.sharding_strategy import CommAction from colossalai.auto_parallel.tensor_shard.solver import CostGraph, GraphAnalyser, Solver, StrategiesConstructor from colossalai.device.alpha_beta_profiler import AlphaBetaProfiler from colossalai.device.device_mesh import DeviceMesh from colossalai.fx.graph_module import ColoGraphModule from colossalai.fx.tracer import ColoTracer from colossalai.tensor.sharding_spec import ShardingSpec class ModuleWrapper(nn.Module): ''' This class is used to wrap the original module, and add the sharding_spec_dict, origin_spec_dict, comm_actions_dict into the forward function. ''' def __init__(self, module: ColoGraphModule, sharding_spec_dict: Dict[int, List[ShardingSpec]], origin_spec_dict: Dict[int, ShardingSpec], comm_actions_dict: Dict[int, Dict[str, CommAction]]): ''' Args: module: the original module sharding_spec_dict: The sharding_spec_dict is used to record the target sharding specs of each tensor required in user node. origin_spec_dict: The origin_spec_dict is used to record the original sharding spec of each tensor. comm_actions_dict: The comm_actions_dict is used to record the communication actions of each tensor. ''' super(ModuleWrapper, self).__init__() self.module = module self.sharding_spec_dict = sharding_spec_dict self.origin_spec_dict = origin_spec_dict self.comm_actions_dict = comm_actions_dict def forward(self, *args, **kwargs): return self.module(*args, sharding_spec_convert_dict=self.sharding_spec_dict, origin_node_sharding_spec_dict=self.origin_spec_dict, comm_actions_dict=self.comm_actions_dict, **kwargs) def extract_meta_args_from_dataloader(data_loader: torch.utils.data.DataLoader, data_process_func: callable): ''' This method is used to extract the meta_args from the dataloader under the instruction of the data_process_func. ''' # TODO: implement this function pass def extract_alpha_beta_for_device_mesh(alpha_beta_dict: Dict[Tuple[int], Tuple[float]], logical_mesh_shape: Tuple[int]): ''' This method is used to extract the mesh_alpha and mesh_beta for the given logical_mesh_shape from the alpha_beta_dict. These two values will be used to estimate the communication cost. ''' # TODO: implement this function pass def build_strategy_constructor(graph: Graph, device_mesh: DeviceMesh, solver_preference: str, dataloader_option: str, shard_option: str): ''' This method is used to build the strategy_constructor for the given graph. After this method, each node in the graph will have a strategies_vector which is constructed by the related node handler. ''' if solver_preference == 'standard': solver_preference = SolverPerference.STANDARD elif solver_preference == 'tp': solver_preference = SolverPerference.TP elif solver_preference == 'dp': solver_preference = SolverPerference.DP else: raise ValueError(f'Invalid solver_preference: {solver_preference}') if dataloader_option == 'replicated': dataloader_option = DataloaderOption.REPLICATED elif dataloader_option == 'distributed': dataloader_option = DataloaderOption.DISTRIBUTED else: raise ValueError(f'Invalid dataloader_option: {dataloader_option}') if shard_option == 'standard': shard_option = ShardOption.STANDARD elif shard_option == 'shard': shard_option = ShardOption.SHARD elif shard_option == 'shard_last_axis': shard_option = ShardOption.SHARD_LAST_AXIS elif shard_option == 'full_shard': shard_option = ShardOption.FULL_SHARD else: raise ValueError(f'Invalid shard_option: {shard_option}') solver_options = SolverOptions(solver_perference=solver_preference, dataloader_option=dataloader_option, shard_option=shard_option) strategies_constructor = StrategiesConstructor(graph, device_mesh, solver_options) strategies_constructor.build_strategies_and_cost() return strategies_constructor def solve_solution(gm: ColoGraphModule, strategy_constructor: StrategiesConstructor, memory_budget: float = -1.0): ''' This method is used to solve the best solution for the given graph. The solution is a list of integers, each integer represents the best strategy index of the corresponding node. ''' graph_analyser = GraphAnalyser(gm) liveness_list = graph_analyser.liveness_analysis() cost_graph = CostGraph(strategy_constructor.leaf_strategies) cost_graph.simplify_graph() solver = Solver(gm.graph, strategy_constructor, cost_graph, graph_analyser, memory_budget=memory_budget) ret = solver.call_solver_serialized_args() solution = list(ret[0]) return solution def transform_to_sharded_model(gm: ColoGraphModule, solution: List[int], device_mesh: DeviceMesh, strategies_constructor: StrategiesConstructor, overlap: bool = False): ''' This method is used to transform the original graph to the sharded graph. The model parameters will be sharded according to the solution and the grad hooks will be added to the sharded graph using the runtime_preparation_pass. The communication node will be added into the graph using the runtime_apply_pass. ''' gm, sharding_spec_dict, origin_spec_dict, comm_actions_dict = runtime_preparation_pass(gm, solution, device_mesh, strategies_constructor, overlap=overlap) gm = runtime_apply_pass(gm) gm.recompile() sharding_spec_dicts = (sharding_spec_dict, origin_spec_dict, comm_actions_dict) return gm, sharding_spec_dicts def initialize_device_mesh(world_size: int = -1, physical_devices: List[int] = None, alpha_beta_dict: Dict[Tuple[int], Tuple[float]] = None, logical_mesh_shape: Tuple[int] = None, logical_mesh_id: torch.Tensor = None): ''' This method is used to initialize the device mesh. Args: world_size: the size of device mesh. If the world_size is -1, the world size will be set to the number of GPUs in the current machine. physical_devices: the physical devices used to initialize the device mesh. alpha_beta_dict(optional): the alpha_beta_dict contains the alpha and beta values for each devices. if the alpha_beta_dict is None, the alpha_beta_dict will be generated by profile_alpha_beta function. logical_mesh_shape(optional): the logical_mesh_shape is used to specify the logical mesh shape. logical_mesh_id(optional): the logical_mesh_id is used to specify the logical mesh id. ''' # if world_size is not set, use the world size from torch.distributed if world_size == -1: world_size = dist.get_world_size() if physical_devices is None: physical_devices = [i for i in range(world_size)] physical_mesh = torch.tensor(physical_devices) if alpha_beta_dict is None: # if alpha_beta_dict is not given, use a series of executions to profile alpha and beta values for each device ab_profiler = AlphaBetaProfiler(physical_devices) alpha_beta_dict = ab_profiler.alpha_beta_dict else: ab_profiler = AlphaBetaProfiler(physical_devices, alpha_beta_dict=alpha_beta_dict) if logical_mesh_shape is None and logical_mesh_id is None: # search for the best logical mesh shape logical_mesh_id = ab_profiler.search_best_logical_mesh() logical_mesh_id = torch.Tensor(logical_mesh_id).to(torch.int) logical_mesh_shape = logical_mesh_id.shape # extract alpha and beta values for the chosen logical mesh shape mesh_alpha, mesh_beta = ab_profiler.extract_alpha_beta_for_device_mesh() elif logical_mesh_shape is not None and logical_mesh_id is None: logical_mesh_id = physical_mesh.reshape(logical_mesh_shape) # extract alpha and beta values for the chosen logical mesh shape mesh_alpha, mesh_beta = extract_alpha_beta_for_device_mesh(alpha_beta_dict, logical_mesh_id) device_mesh = DeviceMesh(physical_mesh_id=physical_mesh, logical_mesh_id=logical_mesh_id, mesh_alpha=mesh_alpha, mesh_beta=mesh_beta, init_process_group=True) return device_mesh def initialize_model(model: nn.Module, meta_args: Dict[str, torch.Tensor], device_mesh: DeviceMesh, memory_budget: float = -1.0, overlap: bool = False, solver_preference: str = 'standard', dataloader_option: str = 'replicated', shard_option: str = 'standard', save_solver_solution: bool = False, load_solver_solution: bool = False, solution_path: str = None, return_solution: bool = False): ''' This method is used to initialize the sharded model which could be used as normal pytorch model. Args: model: the model to be sharded. meta_args: the meta_args is used to specify the input shapes of the model. device_mesh: the device mesh to execute the model. memory_budget(optional): the max cuda memory could be used. If the memory budget is -1.0, the memory budget will be infinity. overlap(optional): the overlap is used to specify whether to overlap gradient communication and backward computing. solver_preference(optional): the solver_preference is used to specify which parallelism algorithm has higher priority. The valid solver_preference could be 'standard', 'tp', or 'dp'. dataloader_option(optional): the dataloader_option is used to specify which kind of data_loader will be used. The valid dataloader_option could be 'replicated' or 'distributed'. shard_option(optional): the shard_option is used to specify how many axes will be used to shard the model. The valid shard_option could be 'standard', 'shard', 'shard_last_axis', or 'full_shard'. save_solver_solution(optional): if the save_solver_solution is True, the solution will be saved to the solution_path. load_solver_solution(optional): if the load_solver_solution is True, the solution will be loaded from the solution_path. solution_path(optional): the path to save or load the solution. return_solution(optional): if the return_solution is True, the solution will be returned. The returned solution will be used to debug or help to analyze the sharding result. Therefore, we will not just return a series of integers, but return the best strategies. ''' tracer = ColoTracer(trace_act_ckpt=True) graph = tracer.trace(root=model, meta_args=meta_args) gm = ColoGraphModule(model, graph, model.__class__.__name__) gm.recompile() strategies_constructor = build_strategy_constructor(graph, device_mesh, solver_preference=solver_preference, dataloader_option=dataloader_option, shard_option=shard_option) if load_solver_solution: solution = torch.load(solution_path) else: solution = solve_solution(gm, strategies_constructor, memory_budget) if save_solver_solution: torch.save(solution, solution_path) gm, sharding_spec_dicts = transform_to_sharded_model(gm, solution, device_mesh, strategies_constructor, overlap) model_to_return = ModuleWrapper(gm, *sharding_spec_dicts) if return_solution: solution_to_return = [] nodes = [strategies_vector.node for strategies_vector in strategies_constructor.leaf_strategies] for index, node in enumerate(nodes): solution_to_return.append(f'{node.name} {node.strategies_vector[solution[index]].name}') return model_to_return, solution_to_return else: return model_to_return def autoparallelize(model: nn.Module, meta_args: Dict[str, torch.Tensor] = None, data_loader: torch.utils.data.DataLoader = None, data_process_func: callable = None, alpha_beta_dict: Dict[Tuple[int], Tuple[float]] = None, logical_mesh_shape: Tuple[int] = None, logical_mesh_id: torch.Tensor = None, solver_preference: str = 'standard', dataloader_option: str = 'replicated', shard_option: str = 'standard', save_solver_solution: bool = False, load_solver_solution: bool = False, solver_solution_path: str = None, return_solution: bool = False, memory_budget: float = -1.0): ''' This method is used to initialize the device mesh, extract the meta_args, and use them to create a sharded model. Args: model: the model to be sharded. meta_args(optional): the meta_args is used to specify the input shapes of the model. If the meta_args is None, the meta_args will be extracted from the data_loader. data_loader(optional): the data_loader to be used in normal training loop. data_process_func(optional): the data_process_func is used to process the data from the data_loader. alpha_beta_dict(optional): the alpha_beta_dict contains the alpha and beta values for each devices. if the alpha_beta_dict is None, the alpha_beta_dict will be generated by profile_alpha_beta function. logical_mesh_shape(optional): the logical_mesh_shape is used to specify the logical mesh shape. If the logical_mesh_shape is None, the logical_mesh_shape will be generated by search_best_logical_mesh_shape function. logical_mesh_id(optional): the logical_mesh_id is used to specify the logical mesh id. solver_preference(optional): the solver_preference is used to specify which parallelism algorithm has higher priority. The valid solver_preference could be 'standard', 'tp', or 'dp'. dataloader_option(optional): the dataloader_option is used to specify which kind of data_loader will be used. The valid dataloader_option could be 'replicated' or 'distributed'. shard_option(optional): the shard_option is used to specify how many axes will be used to shard the model. The valid shard_option could be 'standard', 'shard', 'shard_last_axis', or 'full_shard'. save_solver_solution(optional): if the save_solver_solution is True, the solution will be saved to the solution_path. load_solver_solution(optional): if the load_solver_solution is True, the solution will be loaded from the solution_path. solver_solution_path(optional): the path to save or load the solution. return_solution(optional): if the return_solution is True, the solution will be returned. memory_budget(optional): the max cuda memory could be used. If the memory budget is -1.0, the memory budget will be infinity. ''' device_mesh = initialize_device_mesh(alpha_beta_dict=alpha_beta_dict, logical_mesh_shape=logical_mesh_shape, logical_mesh_id=logical_mesh_id) if meta_args is None: meta_args = extract_meta_args_from_dataloader(data_loader, data_process_func) rst_to_unpack = initialize_model(model, meta_args, device_mesh, solver_preference=solver_preference, dataloader_option=dataloader_option, save_solver_solution=save_solver_solution, load_solver_solution=load_solver_solution, solution_path=solver_solution_path, return_solution=return_solution, memory_budget=memory_budget) if return_solution: model, solution = rst_to_unpack return model, solution else: model = rst_to_unpack return model
import operator import torch __all__ = [ 'ELEMENTWISE_MODULE_OP', 'ELEMENTWISE_FUNC_OP', 'RESHAPE_FUNC_OP', 'CONV_MODULE_OP', 'CONV_FUNC_OP', 'LINEAR_MODULE_OP', 'LINEAR_FUNC_OP', 'BATCHNORM_MODULE_OP', 'POOL_MODULE_OP', 'NON_PARAM_FUNC_OP', 'BCAST_FUNC_OP', 'EMBEDDING_MODULE_OP', 'LAYERNORM_MODULE_OP', 'ELEMENTWISE_METHOD_OP', 'RESHAPE_METHOD_OP', 'INFINITY_COST' ] ELEMENTWISE_MODULE_OP = [torch.nn.Dropout, torch.nn.ReLU] ELEMENTWISE_FUNC_OP = [ torch.abs, torch.cos, torch.exp, operator.neg, torch.multiply, torch.nn.functional.relu, torch.nn.functional.dropout, # softmax should not be here torch.nn.functional.softmax ] ELEMENTWISE_METHOD_OP = [ torch.Tensor.to, torch.Tensor.type, # TODO: contiguous maybe need some extra processes. torch.Tensor.contiguous ] RESHAPE_FUNC_OP = [ torch.flatten, torch.reshape, torch.transpose, torch.split, torch.permute, operator.getitem, ] RESHAPE_METHOD_OP = [ torch.Tensor.view, torch.Tensor.unsqueeze, torch.Tensor.split, torch.Tensor.permute, torch.Tensor.transpose, ] BCAST_FUNC_OP = [ torch.add, torch.sub, torch.mul, torch.div, torch.floor_divide, torch.true_divide, operator.add, operator.sub, operator.mul, operator.floordiv, operator.truediv, torch.matmul, operator.pow, torch.pow ] CONV_MODULE_OP = [ torch.nn.Conv1d, torch.nn.Conv2d, torch.nn.Conv3d, torch.nn.ConvTranspose1d, torch.nn.ConvTranspose2d, torch.nn.ConvTranspose3d ] CONV_FUNC_OP = [ torch.conv1d, torch.conv2d, torch.conv3d, torch.conv_transpose1d, torch.conv_transpose2d, torch.conv_transpose3d ] EMBEDDING_MODULE_OP = [torch.nn.modules.sparse.Embedding] LINEAR_MODULE_OP = [torch.nn.Linear] LINEAR_FUNC_OP = [torch.nn.functional.linear, torch.matmul, torch.bmm] BATCHNORM_MODULE_OP = [torch.nn.BatchNorm1d, torch.nn.BatchNorm2d, torch.nn.BatchNorm3d, torch.nn.SyncBatchNorm] LAYERNORM_MODULE_OP = [torch.nn.LayerNorm] POOL_MODULE_OP = [torch.nn.MaxPool1d, torch.nn.MaxPool2d, torch.nn.MaxPool3d, torch.nn.AdaptiveAvgPool2d] NON_PARAM_FUNC_OP = [ torch.flatten, torch.reshape, torch.abs, torch.cos, torch.exp, operator.neg, torch.multiply, torch.nn.functional.relu, torch.nn.functional.dropout, torch.flatten, torch.where, operator.pow, torch.pow, torch.tanh, torch.add, torch.sub, torch.mul, torch.div, torch.floor_divide, torch.true_divide, operator.add, operator.sub, operator.mul, operator.floordiv, operator.truediv, # softmax should not be here torch.nn.functional.softmax ] INFINITY_COST = 1e13
import operator from copy import deepcopy from functools import reduce from typing import Dict import torch from colossalai.tensor.sharding_spec import ShardingSpec __all__ = [ 'transpose_partition_dim', 'update_partition_dim', 'enumerate_all_possible_1d_sharding', 'enumerate_all_possible_2d_sharding', 'generate_sharding_size' ] def transpose_partition_dim(sharding_spec: ShardingSpec, dim1: int, dim2: int) -> ShardingSpec: """ Switch the sharding mesh dimensions for two tensor dimensions. This operation is in-place. Args: sharding_spec (ShardingSpec): the sharding spec for which partition dim are switched dim1 (int): the tensor dimension to switch dim2 (int): the tensor dimension to switch """ assert len(sharding_spec.entire_shape) >= 2, \ 'The entire_shape of the sharding spec must have at least 2 dimensions' dim_partition_dict = sharding_spec.dim_partition_dict # transpose the dim partition dim1_partition = dim_partition_dict.pop(dim1, None) dim2_partition = dim_partition_dict.pop(dim2, None) if dim1_partition: dim_partition_dict[dim2] = dim1_partition if dim2_partition: dim_partition_dict[dim1] = dim2_partition # get the transposed shape new_shape = list(sharding_spec.entire_shape[:]) new_shape[dim2], new_shape[dim1] = new_shape[dim1], new_shape[dim2] new_shape = torch.Size(new_shape) # re-init the sharding spec sharding_spec.__init__(sharding_spec.device_mesh, new_shape, dim_partition_dict) return sharding_spec def update_partition_dim(sharding_spec: ShardingSpec, dim_mapping: Dict[int, int], physical_shape: torch.Size, inplace: bool = False): """ This method is used to update the partition dim dict from the logical one to the physical one. Args: sharding_spec (ShardingSpec): the sharding spec for which partition dims are updated dim_mapping (Dict[int, int]): the mapping from the logical tensor dimension to the physical tensor dimension physical_shape (torch.Size): the physical shape for the tensor """ if inplace: current_sharding_spec = sharding_spec else: current_sharding_spec = deepcopy(sharding_spec) old_dim_partition_dict = current_sharding_spec.dim_partition_dict new_dim_partition_dict = {} # assign new dim for old_dim, new_dim in dim_mapping.items(): mesh_dims = old_dim_partition_dict.pop(old_dim) new_dim_partition_dict[new_dim] = mesh_dims for tensor_dim, mesh_dims in old_dim_partition_dict.items(): if tensor_dim in new_dim_partition_dict: raise KeyError(f"There are duplicated entries for the tensor sharding dimension {tensor_dim}") else: new_dim_partition_dict[tensor_dim] = mesh_dims # update sharding spec current_sharding_spec.__init__(device_mesh=sharding_spec.device_mesh, entire_shape=physical_shape, dim_partition_dict=new_dim_partition_dict) return current_sharding_spec def enumerate_all_possible_2d_sharding(mesh_dim_0, mesh_dim_1, dim_size): dim_partition_list = [] # enumerate all the 2D sharding cases for i in range(dim_size): for j in range(i + 1, dim_size): dim_partition_dict_0 = {i: [mesh_dim_0], j: [mesh_dim_1]} dim_partition_dict_1 = {i: [mesh_dim_1], j: [mesh_dim_0]} dim_partition_list.append(dim_partition_dict_0) dim_partition_list.append(dim_partition_dict_1) for i in range(dim_size): dim_partition_dict_flatten = {i: [mesh_dim_0, mesh_dim_1]} dim_partition_list.append(dim_partition_dict_flatten) return dim_partition_list def enumerate_all_possible_1d_sharding(mesh_dim_0, dim_size): dim_partition_list = [] # enumerate all the 1D sharding cases for i in range(dim_size): dim_partition_dict_0 = {i: [mesh_dim_0]} dim_partition_list.append(dim_partition_dict_0) return dim_partition_list def generate_sharding_size(dim_partition_dict, device_mesh): total_sharding_size = 1 for mesh_dim_list in dim_partition_dict.values(): mesh_dim_sharding_size = [device_mesh.shape[mesh_dim] for mesh_dim in mesh_dim_list] sharding_size = reduce(operator.mul, mesh_dim_sharding_size) total_sharding_size *= sharding_size return total_sharding_size
import functools from typing import Any, Callable, Dict, List, Tuple, Type, Union import torch from colossalai.logging import get_dist_logger from colossalai.tensor.sharding_spec import ShardingSpec, ShardingSpecException __all__ = ['ignore_sharding_exception', 'pytree_map'] def ignore_sharding_exception(func): """ A function wrapper to handle the ShardingSpecException in the function. If ShardingSpecException occurs, this function will return None. Usage: # mute the assertion error in the function @ignore_sharding_exception def do_something(): ... """ @functools.wraps(func) def wrapper(*args, **kwargs): try: logger = get_dist_logger() rst = func(*args, **kwargs) return rst except ShardingSpecException as e: logger.debug(e) return None return wrapper def check_sharding_spec_validity(sharding_spec: ShardingSpec, tensor: torch.Tensor): """ This function checks whether the ShardingSpec is valid for the physical tensor. This check includes 3 items: 1. the sharding spec covers all dimensions of the physical tensor 2. the sharding spec for each dimension is divisible by the number of devices. 3. the sharding spec's entire shape must match the tensor shape # """ # make sure all dims are covered in sharding spec sharding_len = len(sharding_spec.sharding_sequence) tensor_num_dim = tensor.dim() num_devices_in_col = sharding_spec.device_mesh.mesh_shape[0] num_devices_in_row = sharding_spec.device_mesh.mesh_shape[1] assert sharding_len == tensor_num_dim, \ f'The ShardingSpec ({sharding_spec.sharding_sequence}) is created for {sharding_len}-dimension tensor, but the given tensor is {tensor_num_dim}-dimension ({tensor.shape}).' # make sure the sharding is valid for each dim for i in range(tensor_num_dim): dim_size = tensor.shape[i] dim_spec = sharding_spec.sharding_sequence[i] if str(dim_spec).startswith('S'): devices_str = str(dim_spec).lstrip('S') num_devices = 1 if '0' in devices_str: num_devices *= num_devices_in_col if '1' in devices_str: num_devices *= num_devices_in_row assert dim_size >= num_devices and dim_size % num_devices == 0, \ f'The dimension at index {i} has value {dim_size}, but it is sharded over {num_devices} devices.' # make sure the entire shape matches the physical tensor shape assert sharding_spec.entire_shape == tensor.shape, \ f'The entire_shape of the sharding spec {sharding_spec.entire_shape} does not match the tensor shape {tensor.shape}' def pytree_map(obj: Any, fn: Callable, process_types: Union[Type, Tuple[Type]] = (), map_all: bool = False) -> Any: """process object recursively, like pytree Args: obj (:class:`Any`): object to process fn (:class:`Callable`): a function to process subobject in obj process_types (:class: `type | tuple[type]`): types to determine the type to process map_all (:class: `bool`): if map_all is True, then any type of element will use fn Returns: :class:`Any`: returns have the same structure of `obj` and type in process_types after map of `fn` """ if isinstance(obj, dict): return {k: pytree_map(obj[k], fn, process_types, map_all) for k in obj} elif isinstance(obj, tuple): return tuple(pytree_map(o, fn, process_types, map_all) for o in obj) elif isinstance(obj, list): return list(pytree_map(o, fn, process_types, map_all) for o in obj) elif isinstance(obj, process_types): return fn(obj) else: return fn(obj) if map_all else obj
from enum import Enum from typing import Dict, List, Tuple import torch class PreviousStatus(Enum): """ This class shows the status of previous comparision. """ RESET = 0 # ORIGIN means the dimension size of original tensor is larger in the previous comparision. ORIGIN = 1 # TGT means the dimension size of target tensor is larger in the previous comparision. TGT = 2 def detect_reshape_mapping(origin_shape: torch.Size, tgt_shape: torch.Size) -> Dict[Tuple[int], Tuple[int]]: """ This method is used to detect the reshape mapping between original tensor and target tensor. Returns: reshape_mapping_dict: The dictionary shows how a tuple of origin dims(keys) mapping to the related target dims(values) during reshaping operation. Examples: import torch origin_shape = torch.Size([4, 4, 4]) tgt_shape = torch.Size([2, 8, 2, 2]) reshape_mapping_dict = detect_reshape_mapping(origin_shape, tgt_shape) print(reshape_mapping_dict) Output: {(2,): (3, 2), (1, 0): (1,), (0,): (0, 1)} """ # reverse the shape object origin_shape = list(origin_shape) tgt_shape = list(tgt_shape) origin_shape.reverse() tgt_shape.reverse() # initialize arguments reshape_mapping_dict = {} origin_len = len(origin_shape) tgt_len = len(tgt_shape) origin_index = 0 tgt_index = 0 original_dimension_size = origin_shape[origin_index] tgt_dimension_size = tgt_shape[tgt_index] tgt_dims = [tgt_len - tgt_index - 1] origin_dims = [origin_len - origin_index - 1] previous_label = PreviousStatus.RESET while origin_index != len(origin_shape) or tgt_index != len(tgt_shape): if original_dimension_size == tgt_dimension_size: reshape_mapping_dict[tuple(origin_dims)] = tuple(tgt_dims) # if the origin_dims has no element, it means the original tensor has been fully matched. # Therefore, we do not have to increase the origin_index for that case. if len(origin_dims) > 0: origin_index += 1 # if the tgt_dims has no element, it means the original tensor has been fully matched. # Therefore, we do not have to increase the tgt_index for that case. if len(tgt_dims) > 0: tgt_index += 1 # the last step of loop should always end with condition # so we need to manually skip the preparation for next step # in the last step. if origin_index == len(origin_shape) and tgt_index == len(tgt_shape): continue # If origin_index equals to origin_len, we just need to set the original_dimension_size # to 1 to match the remaining '1's in the target tensor shape. if origin_index == len(origin_shape): original_dimension_size = 1 origin_dims = [] else: original_dimension_size = origin_shape[origin_index] origin_dims = [origin_len - origin_index - 1] # If tgt_index equals to tgt_len, we just need to set the tgt_dimension_size # to 1 to match the remaining '1's in the original tensor shape. if tgt_index == len(tgt_shape): tgt_dimension_size = 1 tgt_dims = [] else: tgt_dimension_size = tgt_shape[tgt_index] tgt_dims = [tgt_len - tgt_index - 1] previous_label = PreviousStatus.RESET elif original_dimension_size > tgt_dimension_size: tgt_index += 1 if previous_label == PreviousStatus.TGT: # if the target dimension size is larger in the previous comparision, which means # the origin dimension size has already accumulated larger than target dimension size, so # we need to offload the origin dims and tgt dims into the reshape_mapping_dict. reshape_mapping_dict[tuple(origin_dims)] = tuple(tgt_dims) original_dimension_size = original_dimension_size // tgt_dimension_size origin_dims = [origin_len - origin_index - 1] tgt_dimension_size = tgt_shape[tgt_index] tgt_dims = [tgt_len - tgt_index - 1, tgt_len - tgt_index] # reset the previous_label after offloading the origin dims and tgt dims previous_label = PreviousStatus.RESET else: # accumulate the tgt_dimension_size until tgt_dimension_size larger than original_dimension_size tgt_dimension_size *= tgt_shape[tgt_index] tgt_dims.append(tgt_len - tgt_index - 1) previous_label = PreviousStatus.ORIGIN else: origin_index += 1 if previous_label == PreviousStatus.ORIGIN: # if the origin element is larger in the previous comparision, which means # the target element has already accumulated larger than origin element, so # we need to offload the origin dims and tgt dims into the reshape_mapping_dict. reshape_mapping_dict[tuple(origin_dims)] = tuple(tgt_dims) tgt_dimension_size = tgt_dimension_size // original_dimension_size tgt_dims = [tgt_len - tgt_index - 1] original_dimension_size = origin_shape[origin_index] origin_dims = [origin_len - origin_index - 1, origin_len - origin_index] # reset the previous_label after offloading the origin dims and tgt dims previous_label = PreviousStatus.RESET else: # accumulate the original_dimension_size until original_dimension_size larger than tgt_dimension_size original_dimension_size *= origin_shape[origin_index] origin_dims.append(origin_len - origin_index - 1) previous_label = PreviousStatus.TGT return reshape_mapping_dict def check_keep_sharding_status(input_dim_partition_dict: Dict[int, List[int]], reshape_mapping_dict: Dict[Tuple[int], Tuple[int]]) -> bool: """ This method is used to check whether the reshape operation could implement without converting the input to fully replicated status. Rule: For a sharded dimension of input tensor, if it is not the minimum element of the input tuple, the function will return false. To illustrate this issue, there are two cases to analyse: 1. no sharded dims in the input tuple: we could do the reshape operation safely just as the normal operation without distributed tensor. 2. sharded dims in the input tuple: the sharded dim must be the minimum element, then during shape consistency process, torch.cat will be implemented on the sharded dim, and everything after the sharded dim get recovered. Examples: # the second dimension of the input has been sharded. input_dim_partition_dict = {1: [1]} origin_shape = torch.Size([8, 4, 2]) tgt_shape = torch.Size([2, 4, 8]) reshape_mapping_dict = detect_reshape_mapping(origin_shape, tgt_shape) # {(2, 1): (2,), (0,): (1, 0)} # the sharded dim of input is 1, which is the minimum element of the tuple (2, 1), # so we do not have to convert the input to fully replicated status. print(check_keep_sharding_status(input_dim_partition_dict, reshape_mapping_dict)) Output: True """ sharded_dims = list(input_dim_partition_dict.keys()) for input_dims in reshape_mapping_dict.keys(): # if input_dims has no element, we could just skip this iteration. if len(input_dims) == 0: continue min_element = min(input_dims) for dim in input_dims: if dim in sharded_dims and dim is not min_element: return False return True def infer_output_dim_partition_dict(input_dim_partition_dict: Dict[int, List[int]], reshape_mapping_dict: Dict[Tuple[int], Tuple[int]]) -> Dict[Tuple[int], Tuple[int]]: """ This method is used to infer the output dim partition dict for a reshape operation, given the input dim partition dict and reshape mapping dict. """ assert check_keep_sharding_status(input_dim_partition_dict, reshape_mapping_dict), \ 'we only infer output dim partition dict for the reshape operation could keep sharding spec.' sharded_dims = list(input_dim_partition_dict.keys()) output_dim_partition_dict = {} for input_dims, output_dims in reshape_mapping_dict.items(): for dim in input_dims: if dim in sharded_dims: output_dim_partition_dict[min(output_dims)] = input_dim_partition_dict[dim] # we could break because input dims cannot contain two sharded dims, otherwise # the keep sharding status check will fail. break return output_dim_partition_dict
from .broadcast import ( BroadcastType, comm_actions_for_oprands, get_broadcast_shape, is_broadcastable, recover_sharding_spec_for_broadcast_shape, ) from .factory import generate_resharding_costs, generate_sharding_spec from .misc import check_sharding_spec_validity, ignore_sharding_exception, pytree_map from .reshape import check_keep_sharding_status, detect_reshape_mapping, infer_output_dim_partition_dict from .sharding import ( enumerate_all_possible_1d_sharding, enumerate_all_possible_2d_sharding, generate_sharding_size, transpose_partition_dim, update_partition_dim, ) __all__ = [ 'BroadcastType', 'get_broadcast_shape', 'is_broadcastable', 'recover_sharding_spec_for_broadcast_shape', 'generate_resharding_costs', 'generate_sharding_spec', 'ignore_sharding_exception', 'check_sharding_spec_validity' 'transpose_partition_dim', 'update_partition_dim', 'enumerate_all_possible_1d_sharding', 'enumerate_all_possible_2d_sharding', 'generate_sharding_size', 'comm_actions_for_oprands', 'pytree_map', 'detect_reshape_mapping', 'check_keep_sharding_status', 'infer_output_dim_partition_dict' ]
import operator import warnings from functools import reduce from typing import Dict, List, Optional, Union import torch from colossalai.device.device_mesh import DeviceMesh from colossalai.tensor.shape_consistency import ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec from torch.fx.node import Node from ..constants import INFINITY_COST __all__ = ['generate_sharding_spec', 'generate_resharding_costs'] def generate_sharding_spec(input_: Union[Node, torch.Tensor], device_mesh: DeviceMesh, dim_partition_dict: Dict[int, List[int]]) -> ShardingSpec: """ Generate the sharding spec of the tensor based on the given dim_partition_dict. Args: input_ (Union[Node, torch.Tensor]): the input can be a Node object or a PyTorch tensor. If a node is used, it will look for its meta data associated with this node. device_mesh (DeviceMesh): a DeviceMesh object which contains the meta information about the cluster. dim_partition_dict (Dict[int, List[int]]): a dictionary to specify the sharding specs, the key is the tensor dimension and the value is the mesh dimension for sharding. """ if isinstance(input_, Node): assert hasattr(input_, '_meta_data'), f'The given node has no attribte _meta_data' meta_tensor = input_._meta_data assert meta_tensor is not None, "The given node's _meta_data attribute is None" shape = meta_tensor.shape elif isinstance(input_, torch.Tensor): shape = input_.shape else: raise TypeError( f'We cannot generate sharding spec for {type(input_)} type, only torch.fx.Node or torch.Tensor is expected.' ) for dim_index, sharding_index_list in dim_partition_dict.items(): sharding_list = [device_mesh.mesh_shape[sharding_index] for sharding_index in sharding_index_list] sharding_size = reduce(operator.mul, sharding_list, 1) assert shape[ dim_index] % sharding_size == 0, f'we cannot shard the {dim_index} dimension of tensor into {sharding_size} partitions.' sharding_spec = ShardingSpec(device_mesh=device_mesh, entire_shape=shape, dim_partition_dict=dim_partition_dict) return sharding_spec def generate_resharding_costs(nodes: List[Node], sharding_specs: List[ShardingSpec], count_backward: Optional[bool] = True, dtype: Optional[torch.dtype] = None, index=None): ''' Compute the resharding costs with this specific strategy. Argument: nodes (List[Node]): a list of nodes sharding_spec_for_input(ShardingSpec): a list of ShardingSpec for the nodes. count_backward (Optional[bool]): whether to include the cost of resharding in the backward pass, default is True. False can be used for inference. dtype (Optional[torch.dtype]): the data type for cost calculation, default is None. ''' # The resharding_cost of weight is counted due to sharing weight cases. resharding_costs = {} size_per_elem_bytes = torch.tensor([], dtype=dtype).element_size() # shape consistency manager is a singleton class shape_consistency_manager = ShapeConsistencyManager() for input_node, input_spec in zip(nodes, sharding_specs): resharding_costs[input_node] = [] for strategy in input_node.strategies_vector: input_sharding_spec = strategy.output_sharding_spec if not isinstance(input_sharding_spec, ShardingSpec): assert isinstance(input_sharding_spec, list), 'only ShardingSpec or List[ShardingSpec] is expected.' input_sharding_spec = input_sharding_spec[index] assert isinstance(input_sharding_spec, ShardingSpec), f'The input node should NOT be a tuple of tensor.' try: # compute the resharding cost _, _, total_resharding_cost = shape_consistency_manager.shape_consistency( input_sharding_spec, input_spec) # we need multiply the size of elem dtype to get correct communication cost resharding_cost = total_resharding_cost["total"] * size_per_elem_bytes except AssertionError as e: warnings.warn(f'{e}') resharding_cost = INFINITY_COST resharding_costs[input_node].append(resharding_cost) return resharding_costs
from enum import Enum, auto from typing import List import torch from torch.fx.node import Node from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, OperationData, OperationDataType, ) from colossalai.tensor.comm_spec import CollectiveCommPattern, CommSpec from colossalai.tensor.sharding_spec import ShardingSpec __all__ = [ 'BroadcastType', 'is_broadcastable', 'get_broadcast_shape', 'recover_sharding_spec_for_broadcast_shape', 'comm_actions_for_oprands' ] class BroadcastType(Enum): EQUAL = auto() PADDDING = auto() MULTIPLE = auto() def is_broadcastable(shape1: torch.Size, shape2: torch.Size) -> bool: """ Check if two shapes are broadcastable to each other. """ for s1, s2 in zip(shape1[::-1], shape2[::-1]): if s1 == 1 or s2 == 1 or s1 == s2: pass else: return False return True def get_broadcast_shape(shape1: torch.Size, shape2: torch.Size) -> List[int]: """ Compute the broadcast shape given two shapes. """ assert is_broadcastable(shape1, shape2), f'{shape1} and {shape2} are not broadcastable' shape1_reverse = shape1[::-1] shape2_reverse = shape2[::-1] min_common_dim = min(len(shape1), len(shape2)) dims = [] for s1, s2 in zip(shape1_reverse, shape2_reverse): dims.append(max(s1, s2)) # append the remaining dims dims.extend(shape1_reverse[min_common_dim:]) dims.extend(shape2_reverse[min_common_dim:]) return dims[::-1] def get_broadcast_dim_info(logical_shape, physical_shape): # get the number of dimensions logical_num_dims = len(logical_shape) physical_num_dims = len(physical_shape) assert logical_num_dims >= physical_num_dims, \ 'The number of dimensions in the logical shape is smaller than that of the physical shape, this tensor is not broadcast!' # track the dim and its broadcasting type logical_dim_broadcast_info = {} for i in range(logical_num_dims): # get the trailing dim size logical_dim_idx = logical_num_dims - i - 1 phyiscal_dim_idx = physical_num_dims - i - 1 logical_dim_size = logical_shape[logical_dim_idx] if phyiscal_dim_idx >= 0: physical_dim_size = physical_shape[phyiscal_dim_idx] if physical_dim_size == logical_dim_size: logical_dim_broadcast_info[logical_dim_idx] = BroadcastType.EQUAL elif physical_dim_size == 1 and physical_dim_size != logical_dim_size: logical_dim_broadcast_info[logical_dim_idx] = BroadcastType.MULTIPLE else: logical_dim_broadcast_info[logical_dim_idx] = BroadcastType.PADDDING return logical_dim_broadcast_info def recover_sharding_spec_for_broadcast_shape(logical_sharding_spec: ShardingSpec, logical_shape: torch.Size, physical_shape: torch.Size) -> ShardingSpec: """ This function computes the sharding spec for the physical shape of a broadcast tensor. Args: logical_sharding_spec (ShardingSpec): the sharding spec for the broadcast tensor logical_shape (torch.Size): logical shape is the broadcast shape of a tensor physical_shape (torch.Size): the shape of the tensor before broadcasting """ # if the two shapes are the same, no broadcast occurs # we directly return the current sharding spec # recording the sharding dimensions removed during logical shape converting to physical one removed_dims = [] if list(logical_shape) == list(physical_shape): return logical_sharding_spec, removed_dims # get the number of dimensions logical_num_dims = len(logical_shape) physical_num_dims = len(physical_shape) # get the broadcast info logical_dim_broadcast_info = get_broadcast_dim_info(logical_shape, physical_shape) # generate the sharding spec for the physical shape physical_dim_partition = {} logical_dim_partition = logical_sharding_spec.dim_partition_dict for shape_dim, mesh_dim in logical_dim_partition.items(): logical_broadcast_type = logical_dim_broadcast_info[shape_dim] if logical_broadcast_type == BroadcastType.PADDDING or logical_broadcast_type == BroadcastType.MULTIPLE: removed_dims.extend(mesh_dim) else: # get the corresponding physical dim physical_dim = physical_num_dims - (logical_num_dims - shape_dim) physical_dim_partition[physical_dim] = mesh_dim physical_sharding_spec = ShardingSpec(device_mesh=logical_sharding_spec.device_mesh, entire_shape=physical_shape, dim_partition_dict=physical_dim_partition) return physical_sharding_spec, removed_dims def comm_actions_for_oprands(node: Node, removed_dims: List[int], op_data: OperationData, sharding_spec: ShardingSpec) -> CommAction: """ This method is used to generate communication actions for oprands which lose information during convert logical shape to physical shape. """ if len(removed_dims) == 1: # if list length is 1, extract element from list to avoid using flatten device mesh removed_dims = removed_dims[0] comm_spec = CommSpec(comm_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, sharding_spec=sharding_spec, logical_process_axis=removed_dims) if op_data.type == OperationDataType.PARAM: comm_type = CommType.HOOK else: comm_type = CommType.BEFORE arg_index = -1 for index, arg in enumerate(node.args): if op_data.name == str(arg): arg_index = index assert arg_index >= 0, f'op_data should be an argument of node.' comm_action = CommAction( comm_spec=comm_spec, comm_type=comm_type, arg_index=arg_index, ) return comm_action
import builtins import math import operator from copy import deepcopy from typing import Dict, List import torch from torch.fx import Graph, Node from colossalai.auto_parallel.tensor_shard.node_handler import ( GetattrHandler, OutputHandler, PlaceholderHandler, operator_registry, ) from colossalai.auto_parallel.tensor_shard.sharding_strategy import StrategiesVector from colossalai.auto_parallel.tensor_shard.utils import generate_resharding_costs, generate_sharding_spec from colossalai.device.device_mesh import DeviceMesh from ..options import DataloaderOption, SolverOptions __all__ = ['StrategiesConstructor'] class StrategiesConstructor: """ StrategiesConstructor is used to construct the parallelization plan for the model execution. Args: graph (Graph): a Graph object used for analysis and strategy generation. device_mesh (DeviceMesh): a DeviceMesh object which contains the meta information about the cluster. solver_options (SolverOptions): a SolverOptions object which specifies the preferences for plan searching. """ def __init__(self, graph: Graph, device_mesh: DeviceMesh, solver_options: SolverOptions): self.graph = graph assert graph.owning_module is not None, 'The given graph is not associated with a owning_module' self.root_module = self.graph.owning_module self.nodes = list(graph.nodes) self.device_mesh = device_mesh self.leaf_strategies = [] self.strategy_map = {} self.solver_options = solver_options self.no_strategy_nodes = [] def remove_duplicated_strategy(self, strategies_vector): ''' In build_strategies_and_cost method, we may produce some duplicated strategies. In this method, we will remove the duplicated strategies depending on the strategies name. Note that this operation is in-place. ''' name_checklist = [] remove_list = [] for strategy in strategies_vector: if strategy.name not in name_checklist: name_checklist.append(strategy.name) else: remove_list.append(strategy) for strategy in remove_list: strategies_vector.remove(strategy) def build_strategies_and_cost(self): """ This method is to build the strategy vector for each node in the computation graph. """ def _check_no_strategy_for_node(node): if node.op in ('placeholder', 'get_attr', 'output'): return False def _check_no_strategy_for_data(data): label = True if isinstance(data, torch.Tensor): return False elif isinstance(data, (tuple, list)): for d in data: label = label and _check_no_strategy_for_data(d) return label return _check_no_strategy_for_data(node._meta_data) for node in self.nodes: strategies_vector = StrategiesVector(node) if _check_no_strategy_for_node(node): self.no_strategy_nodes.append(node) pass # placeholder node elif node.op == 'placeholder': if self.solver_options.dataloader_option == DataloaderOption.DISTRIBUTED: placeholder_option = 'distributed' else: assert self.solver_options.dataloader_option == DataloaderOption.REPLICATED, f'placeholder_option {self.solver_options.dataloader_option} is not supported' placeholder_option = 'replicated' placeholder_handler = PlaceholderHandler(node, self.device_mesh, strategies_vector, placeholder_option=placeholder_option) placeholder_handler.register_strategy() # get_attr node elif node.op == 'get_attr': getattr_handler = GetattrHandler(node, self.device_mesh, strategies_vector, shard_option=self.solver_options.shard_option, solver_perference=self.solver_options.solver_perference) getattr_handler.register_strategy() # call_module node elif node.op == 'call_module': target = node.target submod = self.root_module.get_submodule(target) submod_type = type(submod) handler = operator_registry.get(submod_type)(node, self.device_mesh, strategies_vector, shard_option=self.solver_options.shard_option, solver_perference=self.solver_options.solver_perference) handler.register_strategy() # attach metainfo_vector to node if hasattr(handler, 'metainfo_vector'): setattr(node, 'metainfo_vector', handler.metainfo_vector) # call_function node elif node.op == 'call_function': target = node.target handler = operator_registry.get(target)(node, self.device_mesh, strategies_vector, shard_option=self.solver_options.shard_option, solver_perference=self.solver_options.solver_perference) handler.register_strategy() # attach metainfo_vector to node if hasattr(handler, 'metainfo_vector'): setattr(node, 'metainfo_vector', handler.metainfo_vector) # call_method node elif node.op == 'call_method': method = getattr(node.args[0]._meta_data.__class__, node.target) handler = operator_registry.get(method)(node, self.device_mesh, strategies_vector, shard_option=self.solver_options.shard_option, solver_perference=self.solver_options.solver_perference) handler.register_strategy() # attach metainfo_vector to node if hasattr(handler, 'metainfo_vector'): setattr(node, 'metainfo_vector', handler.metainfo_vector) # output node elif node.op == 'output': if self.solver_options.dataloader_option == DataloaderOption.DISTRIBUTED: output_option = 'distributed' else: assert self.solver_options.dataloader_option == DataloaderOption.REPLICATED, f'placeholder_option {self.solver_options.dataloader_option} is not supported' output_option = 'replicated' output_handler = OutputHandler(node, self.device_mesh, strategies_vector, output_option=output_option) output_handler.register_strategy() self.remove_duplicated_strategy(strategies_vector) setattr(node, 'strategies_vector', strategies_vector) self.leaf_strategies.append(strategies_vector) self.strategy_map[node] = strategies_vector # remove no strategy nodes remove_list = [] for strategies_vector in self.leaf_strategies: if len(strategies_vector) == 0: remove_list.append(strategies_vector.node) for node in remove_list: if node.strategies_vector in self.leaf_strategies: self.leaf_strategies.remove(node.strategies_vector) if node in self.strategy_map: self.strategy_map.pop(node)
import torch from colossalai.auto_parallel.tensor_shard.constants import INFINITY_COST class CostGraph: ''' A graph data structure to simplify the edge cost graph. It has two main functions: 1. To feed the quadratic resharding costs into solver, we need to linearize it. We build edge_cost in CostGraph, and it stored every combinations of strategies for a src-dst node pair in an 1D list. 2. To reduce the searching space, we merge computationally-trivial operators, such as element-wise operators, transpose, and reduction, into their following nodes. The merging infomation will be given by the StrategiesVector depending on the type of target node and following nodes. Argument: leaf_strategies(List[StrategiesVector]): It stores StrategiesVector of every nodes on the graph. simplify(bool, optional): The generated cost graph will be simplified if it is true. (default to True) ''' def __init__(self, leaf_strategies, simplify=True, forward_only=False): self.leaf_strategies = leaf_strategies self.nodes = [strategies_vector.node for strategies_vector in self.leaf_strategies] # stores number of strategies in each node self.node_lens = {strategies_vector.node: len(strategies_vector) for strategies_vector in self.leaf_strategies} # extra_node_costs will store the extra costs introduced by merging nodes self.extra_node_costs = {} self.following_dict = {} self.simplify = simplify self.forward_only = forward_only self._build_cost_graph() def _remove_invalid_node(self, node, attr_name): remove_list = [] target_node_list = getattr(node, attr_name, []) for target_node in target_node_list: if target_node not in self.nodes: remove_list.append(target_node) for element in remove_list: target_node_list.remove(element) def _build_cost_graph(self): ''' This method will generate edge_cost for adjacent node pair. Additionally, 'parents' and 'children' attribute will be set to node. ''' self.edge_costs = {} if self.simplify: self.merge_pair = [] for strategies_vector in self.leaf_strategies: # build edge_cost dst_node = strategies_vector.node for src_node in strategies_vector.predecessor_nodes: if src_node not in self.nodes: continue node_pair = (src_node, dst_node) edge_cost = {} for i in range(len(strategies_vector)): for j in range(len(src_node.strategies_vector)): resharding_cost_item = strategies_vector[i].resharding_costs[src_node][j] if self.forward_only: edge_cost[(j, i)] = resharding_cost_item.fwd else: edge_cost[(j, i)] = resharding_cost_item.total self.edge_costs[node_pair] = edge_cost parent_nodes = [] children_nodes = [] def _check_tensor_in_node(data): """ This method is used to check whether the data has a tensor inside or not. """ has_tensor_flag = False if isinstance(data, torch.Tensor): return True elif isinstance(data, (tuple, list)): for d in data: has_tensor_flag = has_tensor_flag or _check_tensor_in_node(d) return has_tensor_flag for node in strategies_vector.predecessor_nodes: if _check_tensor_in_node(node._meta_data): parent_nodes.append(node) for node in strategies_vector.successor_nodes: if _check_tensor_in_node(node._meta_data): children_nodes.append(node) setattr(dst_node, 'parents', parent_nodes) setattr(dst_node, 'children', children_nodes) if self.simplify and strategies_vector.check_merge(): for followed_node in strategies_vector.predecessor_nodes: # we only merge node pairs which src node has a tensor element inside. # This is necessay because the node without a tensor element inside will not # be assigned any strategy. if _check_tensor_in_node(followed_node._meta_data): self.merge_pair.append((followed_node, dst_node)) def get_edge_cost(self, src_node, dst_node): return self.edge_costs[(src_node, dst_node)] def merge_node(self, src_node, dst_node): ''' To merge dst_node into src_node, we need to do it in following steps: 1. For each strategy in dst_node, we need to pick an appropriate strategy of src_node to merge, it is important because the logical resharding costs between the parents node of src_node and merged node depend on the src_node strategies dispatching. For example, for the graph 0->1->2, after merging node 1 into node 2, edge_costs[(node 0, node 2)][(0, 0)] = edge_costs[(node 0, node 1)][(0, x)] x represents the picking strategy of node 1 merged into node 2 strategy 0. 2. We need to accumulate the extra costs introduced by merging nodes, the extra costs contains two parts, one is resharding costs between src_node strategy and dst_node strategy, another is the origin extra costs in src_node strategy. 3. Build connections between new node pairs, and remove the src_node after all consumer nodes detached from it. Argument: src_node(Node): The node will be merged into dst_node. dst_node(Node): The node to integrate src_node. ''' # build merge_map merge_map = {} for src_index, _ in enumerate(src_node.strategies_vector): min_cost = INFINITY_COST lowest_cost_index = -1 for dst_index, dst_strategy in enumerate(dst_node.strategies_vector): resharding_cost_item = dst_strategy.resharding_costs[src_node][src_index] if self.forward_only: resharding_cost = resharding_cost_item.fwd else: resharding_cost = resharding_cost_item.total if resharding_cost <= min_cost: min_cost = resharding_cost lowest_cost_index = dst_index merge_map[src_index] = lowest_cost_index # extra_node_cost for src node self.extra_node_costs[src_node] = [0.0] * self.node_lens[src_node] for src_index, strategy in enumerate(src_node.strategies_vector): target_strate_index = merge_map[src_index] target_strategy = dst_node.strategies_vector[target_strate_index] resharding_cost_item = target_strategy.resharding_costs[src_node][src_index] if self.forward_only: resharding_cost_to_add = resharding_cost_item.fwd else: resharding_cost_to_add = resharding_cost_item.total self.extra_node_costs[src_node][src_index] += resharding_cost_to_add if dst_node in self.extra_node_costs: self.extra_node_costs[src_node][src_index] += self.extra_node_costs[dst_node][target_strate_index] # add new node pair to cost graph for child_node in dst_node.children: new_node_pair = (src_node, child_node) old_node_pair = (dst_node, child_node) if new_node_pair in self.edge_costs: continue edge_cost = {} for i in range(self.node_lens[src_node]): for j in range(self.node_lens[child_node]): dst_strate_index = merge_map[i] edge_cost[(i, j)] = self.edge_costs[old_node_pair][(dst_strate_index, j)] if new_node_pair not in self.edge_costs: self.edge_costs[new_node_pair] = edge_cost else: # we should accumulate the resharding costs if args of child node contain # both src node and dst node. for index_pair, resharding_cost in self.edge_costs[new_node_pair]: self.edge_costs[new_node_pair][index_pair] += edge_cost[index_pair] # connect src node and children of dst node dst_node.parents.remove(src_node) src_node.children.remove(dst_node) self.edge_costs.pop((src_node, dst_node)) for child_node in dst_node.children: if child_node not in src_node.children: src_node.children.append(child_node) if src_node not in child_node.parents: child_node.parents.append(src_node) # remove dst node from cost graph when dst node has no producer. if len(dst_node.parents) == 0: child_node.parents.remove(dst_node) node_pair = (dst_node, child_node) self.edge_costs.pop(node_pair) if len(dst_node.parents) == 0: self.following_dict[dst_node] = src_node dst_node.children = [] def _reindexing_src(self, src): if src not in self.following_dict: return src return self._reindexing_src(self.following_dict[src]) def simplify_graph(self): if not self.simplify: return self.merge_pair.reverse() for (src_node, dst_node) in self.merge_pair: self.merge_node(src_node, dst_node) self.merge_pair.reverse() reindexing_following_dict = {} for dst, src in self.following_dict.items(): reindexing_following_dict[dst] = self._reindexing_src(src) self.following_dict = reindexing_following_dict
import multiprocessing import time import warnings from typing import Dict import numpy as np from torch.fx.graph import Graph from torch.fx.node import Node from colossalai.auto_parallel.tensor_shard.constants import INFINITY_COST from .cost_graph import CostGraph from .graph_analysis import GraphAnalyser from .strategies_constructor import StrategiesConstructor try: import pulp from pulp import LpMinimize, LpProblem, LpStatus, LpVariable, lpDot, lpSum except: warnings.warn(f'please install the pulp') __all___ = ['Solver'] class Solver: def __init__(self, graph: Graph, strategies_constructor: StrategiesConstructor, cost_graph: CostGraph, graph_analyser: GraphAnalyser, memory_budget: float = -1.0, solution_numbers: int = 1, forward_only: bool = False, memory_increasing_coefficient: float = 1.3, verbose=False): ''' Solver class will integrate information provided by the components and use ILP solver to find a possible optimal strategies combination for target computing graph. Argument: graph: The computing graph to be optimized. strategies_constructor: It will provide all the possible strategies for each node in the computing graph. cost_graph: A graph data structure to simplify the edge cost graph. graph_analyser: graph_analyser will analyse the graph to obtain the variable liveness information, which will be used to generate memory constraints. memory_budget: Memory constraint for the solution. solution_numbers: If solution_numbers is larger than one, solver will us a serious of solutions based on different memory budget. memory_increasing_coefficient: If solution_numbers is larger than one, we will use this coefficient to generate new memory budget. ''' self.graph = graph self.strategies_constructor = strategies_constructor self.cost_graph = cost_graph self.graph_analyser = graph_analyser self.leaf_strategies = self.strategies_constructor.leaf_strategies self.nodes = [strategies_vector.node for strategies_vector in self.leaf_strategies] self.strategy_map = self.strategies_constructor.strategy_map self.memory_budget = memory_budget self.solution_numbers = solution_numbers self.forward_only = forward_only if self.solution_numbers > 1: self.memory_increasing_coefficient = memory_increasing_coefficient else: self.memory_increasing_coefficient = 1 self.liveness_list = self.graph_analyser.liveness_analysis() self.node_index_dict = self._generate_node_index_dict() # The last solution vector of auto sharding. self.last_s_val = None # The last objective value of the best ILP solution. self.last_objective = None self.verbose = verbose def _recover_merged_node_strategy(self): ''' During cost graph constructing, some nodes, such as unary element-wise node or ReshapeOp, were merged into the previous node. Therefore, the index of those strategies are copied from the previous node. This method is used to recover the strategy index of those merged node. ''' for node_index, node in enumerate(self.nodes): if node.strategies_vector.check_merge(): # the merged node has only one input, and its strategies follow the input sharding strategy input_strategies_vector = node.args[0].strategies_vector input_best_strategy_index = self.last_s_val[node_index - 1] input_sharding_spec = input_strategies_vector[input_best_strategy_index].output_sharding_spec for strategy_index, strategy in enumerate(node.strategies_vector): if strategy.input_shardings[0].sharding_sequence == input_sharding_spec.sharding_sequence: self.last_s_val[node_index] = strategy_index break def _generate_node_index_dict(self) -> Dict[Node, int]: node_index_dict = {} for index, strategies_vector in enumerate(self.leaf_strategies): node_index_dict[strategies_vector.node] = index return node_index_dict def _prepare_data_for_solver(self): ''' Extract information from components for solver. ''' node_nums = len(self.leaf_strategies) memory_budget = self.memory_budget # prepare strategies_len strategies_len = [] for node in self.nodes: strategies_len.append(self.cost_graph.node_lens[node]) strategies_len = np.array(strategies_len) # prepare following_nodes following_nodes = self.cost_graph.following_dict index_following_nodes = {} for src, target in following_nodes.items(): src_index = self.node_index_dict[src] target_index = self.node_index_dict[target] index_following_nodes[src_index] = target_index following_nodes = index_following_nodes for index in range(node_nums): if index not in following_nodes: following_nodes[index] = -1 # prepare edge_pairs and resharding costs edge_pairs = [] resharding_costs = [] for pairs, edge_cost in self.cost_graph.edge_costs.items(): src_node = pairs[0] dst_node = pairs[1] src_node_index = self.node_index_dict[src_node] dst_node_index = self.node_index_dict[dst_node] edge_pairs.append(src_node_index) edge_pairs.append(dst_node_index) for i in range(strategies_len[src_node_index]): for j in range(strategies_len[dst_node_index]): resharding_costs.append(edge_cost[(i, j)]) edge_pairs = np.array(edge_pairs) resharding_costs = np.array(resharding_costs) # prepare liveness_set liveness_set = self.liveness_list # omit alias_set now alias_set = None alias_convert_costs = None # prepare compute_costs, communication_costs and memory_costs compute_costs = [] communication_costs = [] memory_costs = [] extra_node_costs = self.cost_graph.extra_node_costs for strategies_vector in self.leaf_strategies: node = strategies_vector.node for index, strategy in enumerate(strategies_vector): compute_cost_item = strategy.compute_cost communication_cost_item = strategy.communication_cost memory_cost_item = strategy.memory_cost if self.forward_only: origin_communication_cost = communication_cost_item.fwd compute_cost = compute_cost_item.fwd # extract MemoryCost item from the memory TrainCycleItem memory_cost = memory_cost_item.fwd else: origin_communication_cost = communication_cost_item.total compute_cost = compute_cost_item.total # extract MemoryCost item from the memory TrainCycleItem memory_cost = memory_cost_item.total # extract the memory cost in float from MemoryCost item and sum them up memory_cost = memory_cost.parameter + memory_cost.activation + memory_cost.buffer compute_costs.append(compute_cost) # node in extra_node_costs means it has some extra communication # cost from node merging, so we need to add those extra communication # cost into if node in extra_node_costs: extra_node_cost = extra_node_costs[node][index] communication_cost = origin_communication_cost + extra_node_cost communication_costs.append(communication_cost) else: communication_costs.append(origin_communication_cost) memory_costs.append(memory_cost) compute_costs = np.array(compute_costs) communication_costs = np.array(communication_costs) memory_costs = np.array(memory_costs) # omit initial value for nodes s_init_np = None return node_nums, memory_budget, strategies_len, following_nodes, edge_pairs, alias_set, liveness_set, compute_costs, communication_costs, memory_costs, resharding_costs, alias_convert_costs, s_init_np, self.verbose def _call_solver_serialized_args(self, node_nums, memory_budget, strategies_len, following_nodes, edge_pairs, alias_set, liveness_set, compute_costs, communication_costs, memory_costs, resharding_costs, alias_convert_costs, s_init_np=None, verbose=True): """ Call the solver with serialized arguments. """ tic = time.time() for x in [strategies_len, edge_pairs, compute_costs, communication_costs, memory_costs, resharding_costs]: assert isinstance(x, np.ndarray) assert len(strategies_len) == node_nums, "strategies_len" def get_non_zero_index(binary_vector): """ Get the index of non-zero item in a vector. """ ct = 0 ret = None for i, elem in enumerate(binary_vector): if pulp.value(elem): ret = i ct += 1 assert ct == 1 return ret # 0. Unpack flatten numpy arrays s_follow = following_nodes E = edge_pairs.reshape((-1, 2)) # noqa r = [] pt = 0 edge_set = set() for (i, j) in E: prod_length = strategies_len[i] * strategies_len[j] if (i, j) in edge_set: raise ValueError(f"Duplicated edges: {(i, j)}") edge_set.add((i, j)) r.append(resharding_costs[pt:pt + prod_length]) pt += prod_length assert pt == len(resharding_costs) ###################### # omit alias set now # ###################### # A = alias_set.reshape((-1, 2)) # noqa # for (i, j) in A: # prod_length = strategies_len[i] * strategies_len[j] # v.append(alias_convert_costs[pt:pt + prod_length]) # pt += prod_length # assert pt == len(alias_convert_costs) # L = [] # noqa # pt = node_nums # for i in range(node_nums): # length = liveness_set[i] # L.append(liveness_set[pt:pt + length]) # pt += length # assert pt == len(liveness_set) v = [] pt = 0 c = [] d = [] m = [] pt = 0 for i in range(node_nums): length = strategies_len[i] c.append(compute_costs[pt:pt + length]) d.append(communication_costs[pt:pt + length]) m.append(memory_costs[pt:pt + length]) pt += length assert pt == len(compute_costs), f"{pt} == {len(compute_costs)}" assert pt == len(communication_costs), f"{pt} == {len(communication_costs)}" assert pt == len(memory_costs), f"{pt} == {len(memory_costs)}" # 1. Create variables ############################# # create variables for node # ############################# s = [] num_nodes = 0 reverse_follow_backpatch = [] for i in range(node_nums): if s_follow[i] < 0: if strategies_len[i] == 1: s.append([1]) else: num_nodes += 1 s.append(LpVariable.matrix(f"s[{i}]", (range(strategies_len[i]),), cat="Binary")) else: if s_follow[i] < len(s): s.append(s[s_follow[i]]) else: s.append(None) reverse_follow_backpatch.append(i) for i in reverse_follow_backpatch: s[i] = s[s_follow[i]] ############################# # create variables for edge # ############################# e = [] num_edges = 0 for (idx, (i, j)) in enumerate(E): if len(s[i]) == 1: e.append(s[j]) elif len(s[j]) == 1: e.append(s[i]) else: num_edges += 1 e.append(LpVariable.matrix(f"e[{i},{j}]", (range(len(s[i]) * len(s[j])),), cat="Binary")) assert len(e[idx]) == len(r[idx]) for element in s: assert len(element) > 0 # 2. Set initial value ###################################### # set a initial value for warm start # ###################################### if s_init_np is not None: s_init = s_init_np.reshape((-1, 3)) for (idx, value, fix) in s_init: for i in range(len(s[idx])): s[idx][i].setInitialValue(i == value) if fix: s[idx][i].fixValue() # 3. Objective prob = LpProblem("myProblem", LpMinimize) ################################################################### # computing the node cost(computing cost and communication cost) # ################################################################### obj = 0 for i in range(node_nums): assert len(s[i]) == len(c[i]) assert len(s[i]) == len(d[i]) obj += lpDot(s[i], c[i]) + lpDot(s[i], d[i]) ############################################# # computing the edge cost(resharding cost) # ############################################# for i in range(len(E)): assert len(e[i]) == len(r[i]) obj += lpDot(e[i], r[i]) prob += obj # 4. Constraints # (a). specified by `cat="Binary"` # (b) ################################################# # make sure each node only choose one strategy # ################################################# for i in range(node_nums): if s_follow[i] < 0: prob += lpSum(s[i]) == 1 # (c) ################################################# # compute memory consumption with liveness set # ################################################# if memory_budget > 0: for liveness_stage in liveness_set: mem = 0 for live_variable in liveness_stage.unique_live_vars: if live_variable.node not in self.node_index_dict: continue node_index = self.node_index_dict[live_variable.node] mem += lpSum(s[node_index][j] * m[node_index][j] for j in range(len(s[node_index]))) prob += mem <= memory_budget # (d). specified by `cat="Binary"` for (idx, (i, j)) in enumerate(E): if strategies_len[i] == 1 or strategies_len[j] == 1: continue # (e) prob += lpSum(e[idx]) == 1 # (f) for row in range(len(s[i])): C = len(s[j]) # noqa prob += lpSum(e[idx][row * C + col] for col in range(0, C)) <= s[i][row] # (g) for col in range(len(s[j])): R = len(s[i]) # noqa C = len(s[j]) # noqa prob += lpSum(e[idx][row * C + col] for row in range(0, R)) <= s[j][col] # (h) ###################### # omit alias set now # ###################### # alias_set = set() # for (idx, (i, j)) in enumerate(A): # R = len(s[i]) # noqa # C = len(s[j]) # noqa # if (i, j) in alias_set: # raise ValueError(f"Duplicated edges: {(i, j)}") # alias_set.add((i, j)) # alias_set.add((j, i)) # for row in range(len(s[i])): # for col in range(len(s[j])): # if v[idx][row * C + col] > 0.5: # prob += s[i][row] + s[j][col] <= 1 msg = verbose time_limit = 600 assert "COIN_CMD" in pulp.listSolvers( onlyAvailable=True), ("Please install ILP solvers by 'sudo apt install coinor-cbc'") solver = pulp.COIN_CMD(mip=True, msg=msg, timeLimit=time_limit, threads=multiprocessing.cpu_count()) # solver = pulp.GLPK_CMD(mip=True, msg=msg, timeLimit=time_limit) prob.solve(solver) status = prob.status objective = pulp.value(prob.objective) objective = float(objective) if objective is not None else -1.0 if verbose: print(f"ILP Status: {LpStatus[status]}\tObjective: {objective}\t" f"Time: {time.time() - tic}") print(f"#nodes: {num_nodes}, #edges: {num_edges}") if prob.status in [pulp.LpStatusInfeasible]: raise RuntimeError("Cannot run the function under the given memory budget. " "Please increase the memory budget.") # Get and check results s_val = np.full((node_nums,), -1, dtype=np.int32) for i in range(node_nums): s_val[i] = get_non_zero_index(s[i]) e_val = np.full((len(E),), -1, dtype=np.int32) for (idx, (i, j)) in enumerate(E): e_val[idx] = get_non_zero_index(e[idx]) i_spec_index = e_val[idx] // len(s[j]) j_spec_index = e_val[idx] % len(s[j]) assert i_spec_index == s_val[i], f"e_val[{i}][{j}]" assert j_spec_index == s_val[j], f"e_val[{i}][{j}]" if verbose and r[idx][e_val[idx]] > 0: print(f"Edge cost {(i, j)} : {r[idx][e_val[idx]]}") self.last_s_val = list(s_val) # self._recover_merged_node_strategy() self.last_objective = objective if objective > INFINITY_COST: warnings.warn("Detect unexpected behaviors in the auto-sharding pass.") return self.last_s_val, e_val, self.last_objective, status def call_solver_serialized_args(self): """ Call the solver with serialized arguments and handle python errors. Additionally, we could give a serious of solutions with different memory budget. """ if self.solution_numbers == 1: args = self._prepare_data_for_solver() ret = self._call_solver_serialized_args(*args) return ret origin_memory_budget = self.memory_budget memory_budget_list = [ origin_memory_budget * self.memory_increasing_coefficient**i for i in range(self.solution_numbers) ] ret_list = [] for memory_budget in memory_budget_list: self.memory_budget = memory_budget args = self._prepare_data_for_solver() ret = self._call_solver_serialized_args(*args) ret_list.append(ret) return ret_list
from .cost_graph import CostGraph from .graph_analysis import GraphAnalyser from .solver import Solver from .strategies_constructor import StrategiesConstructor __all__ = ['GraphAnalyser', 'Solver', 'StrategiesConstructor', 'CostGraph']
from dataclasses import dataclass from typing import List from torch.fx.graph import Graph from torch.fx.graph_module import GraphModule from torch.fx.node import Node from colossalai.fx.passes.utils import get_node_module __all__ = ['LiveVariable', 'LiveVariableVector', 'LiveStage', 'GraphAnalyser'] @dataclass class LiveVariable: """ LiveVariable is a data structure to store the meta information of a variable for liveness analysis. """ name: str node: Node is_inplace: bool class LiveVariableVector(list): """ LiveVariableVector is a data structure to store the list of LiveVariable objects. """ def exists(self, name) -> bool: """ Check if a variable has already existed in the current list by name. """ for var in self: if name == var.name: return True return False def get(self, name) -> LiveVariable: for var in self: if name == var.name: return var raise KeyError(f"Variable {name} is not found") def copy(self) -> "LiveVariableVector": """ Create a copy of this vector """ vector = LiveVariableVector() for var in self: vector.append(var) return vector @dataclass class LiveStage: """ LiveStage is a data structure to record the living variables at this current node. """ name: str node: Node all_live_vars: LiveVariableVector unique_live_vars: LiveVariableVector class GraphAnalyser: def __init__(self, gm: GraphModule): self._gm = gm self._graph = gm.graph @property def gm(self) -> GraphModule: """ Return the GraphModule object associated with this analyser. """ return self._gm @property def graph(self) -> Graph: """ Return the Graph object associated with this analyser. """ return self._graph def liveness_analysis(self) -> List[LiveStage]: """ Analyse the graph to obtain the variable liveness information. This function returns an ordered dictionary where the key is the compute stage ID and the value is a LivenessStage object. """ compute_nodes = self.graph.nodes liveness_list = [] # checked: record all variables created since the first stage # all: record the live variables only exist until the current stage. # this can be different from the `checked list`` as some varialbes may be destroyed prior to this stage. # unique: record the unique live variables only exist until the current stage. # this is different from `all list` as some variables are duplicated. checked_variables = LiveVariableVector() all_live_variables = LiveVariableVector() unique_live_vars = LiveVariableVector() for idx, node in enumerate(compute_nodes): ############################# # find new living variables # ############################# # detect whether the current op is an in-place op # if it is an in-place op, we would deem it as a duplciate var is_inplace = False if node.op == 'call_function': # check if this is an inplace op such as torch.nn.functional.relu(x, inplace=True) if node.kwargs.get('inplace', False): is_inplace = True elif node.op == 'call_module': # to check if this is an inplace op such as torch.nn.Relu(inplace=True) module = get_node_module(node) if getattr(module, 'inplace', False): is_inplace = True # add the output var meta = getattr(node, '_meta_data', None) live_var = LiveVariable(name=node.name, node=node, is_inplace=is_inplace) if not is_inplace: unique_live_vars.append(live_var) checked_variables.append(live_var) all_live_variables.append(live_var) # check if any input is not checked yet for arg in node.args: if not isinstance(arg, Node): continue arg_name = arg.name if not checked_variables.exists(arg_name): live_var_from_arg = LiveVariable(name=arg_name, node=node, is_inplace=False) all_live_variables.append(live_var_from_arg) checked_variables.append(live_var_from_arg) unique_live_vars.append(live_var_from_arg) # TODO: add the logic to remove live variables # this should be completed if we are able to trace the backward compute graph # add this stage to liveness dict stage = LiveStage(name=node.name, node=node, all_live_vars=all_live_variables.copy(), unique_live_vars=unique_live_vars.copy()) # if a LiveStage is covered by another LiveStage, we just keep the larger one. replace = False for index, prev_stage in enumerate(liveness_list): all_covered = True for ele in prev_stage.unique_live_vars: if ele not in stage.unique_live_vars: all_covered = False break if all_covered: replace = True break if replace: liveness_list[index] = stage else: liveness_list.append(stage) return liveness_list def get_alias_set(self): pass
from abc import ABC, abstractmethod from typing import Dict, List, Tuple, Union import torch from torch.fx.node import Node from colossalai.auto_parallel.meta_profiler.metainfo import MetaInfo, meta_register from colossalai.auto_parallel.tensor_shard.options import ShardOption, SolverPerference from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( OperationData, OperationDataType, ShardingSpec, ShardingStrategy, StrategiesVector, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import check_sharding_spec_validity from colossalai.device.device_mesh import DeviceMesh from colossalai.logging import get_dist_logger from colossalai.tensor.shape_consistency import ShapeConsistencyManager from .strategy import StrategyGenerator class NodeHandler(ABC): ''' The NodeHandler is an abstract class used to generate every possible strategies for an operator node. Args: node (Node): the input node in node argument list. device_mesh (DeviceMesh): A logical view of a physical mesh. strategies_vector (StrategiesVector): all the strategies generated in this handler will be recorded into the strategies_vector. ''' def __init__(self, node: Node, device_mesh: DeviceMesh, strategies_vector: StrategiesVector, shard_option: ShardOption = ShardOption.STANDARD, solver_perference: SolverPerference = SolverPerference.STANDARD) -> None: self.node = node self.predecessor_node = list(node._input_nodes.keys()) self.successor_node = list(node.users.keys()) self.device_mesh = device_mesh self.strategies_vector = strategies_vector self.shard_option = shard_option self.solver_perference = solver_perference def update_resharding_cost(self, strategy: ShardingStrategy) -> None: """ Compute the resharding costs and save the costs in the ShardingStrategy object. """ # TODO: test this function when other handlers are ready resharding_costs = {} shape_consistency_manager = ShapeConsistencyManager() for node in self.predecessor_node: node_name = str(node) # get the current sharding spec generated by this node handler # we will not compute the resharding costs for the node not counted in the strategy. # And the node with tuple or list output need to be handled below. node_in_strategy = [op_data.name for op_data in strategy.sharding_specs.keys()] if str(node) not in node_in_strategy: continue op_data = strategy.get_op_data_by_name(node_name) current_sharding_spec = strategy.sharding_specs[op_data] # get the sharding specs for this node generated # in its own node handler assert hasattr(node, 'strategies_vector'), \ f'The predecessor node {node_name} has no strategy vector to compute the resharding cost.' prev_strategy_vector = node.strategies_vector prev_sharding_specs = [ prev_strategy.get_sharding_spec_by_name(node_name) for prev_strategy in prev_strategy_vector ] # create data structrure to store costs if node not in resharding_costs: resharding_costs[node] = [] def _compute_resharding_cost( prev_sharding_spec: Union[ShardingSpec, List[ShardingSpec]], current_sharding_spec: Union[ShardingSpec, List[ShardingSpec]], data: Union[torch.Tensor, List[torch.Tensor], Tuple[torch.Tensor]]) -> TrainCycleItem: """ This is a helper function to compute the resharding cost for a specific strategy of a node. """ if prev_sharding_spec is None: return TrainCycleItem(fwd=0, bwd=0, total=0) elif isinstance(prev_sharding_spec, ShardingSpec): if isinstance(data, torch.Tensor): dtype = data.dtype size_per_elem_bytes = torch.tensor([], dtype=dtype).element_size() _, _, consistency_cost = shape_consistency_manager.shape_consistency( prev_sharding_spec, current_sharding_spec) resharding_cost = TrainCycleItem(fwd=consistency_cost["forward"] * size_per_elem_bytes, bwd=consistency_cost["backward"] * size_per_elem_bytes, total=consistency_cost["total"] * size_per_elem_bytes) return resharding_cost else: # This raise is used to check if we have missed any type of data. # It could be merged into Parameter branch, which means we won't handle # non-tensor arguments. raise ValueError(f'Unsupported data type {type(data)}') else: assert isinstance(prev_sharding_spec, (tuple, list)), \ f'prev_sharding_spec should be in type of ShardingSpec, List[ShardingSpec], \ or Tuple[ShardingSpec], but got {type(prev_sharding_spec)}' fwd_cost = 0 bwd_cost = 0 total_cost = 0 for index, (prev_sharding_spec_item, current_sharding_spec_item) in enumerate(zip(prev_sharding_spec, current_sharding_spec)): item_cost = _compute_resharding_cost(prev_sharding_spec_item, current_sharding_spec_item, data[index]) fwd_cost += item_cost.fwd bwd_cost += item_cost.bwd total_cost += item_cost.total resharding_cost = TrainCycleItem(fwd=fwd_cost, bwd=bwd_cost, total=total_cost) return resharding_cost # for each sharding spec generated by the predecessor's node handler # compute the resharding cost to switch to the sharding spec generated # by the current node handler for prev_sharding_spec in prev_sharding_specs: resharding_cost = _compute_resharding_cost(prev_sharding_spec, current_sharding_spec, op_data.data) resharding_costs[node].append(resharding_cost) strategy.resharding_costs = resharding_costs return strategy def get_target_function(self) -> callable: """ This function is used to get the target function for the node handler. The target function is used to analyze the costs of strategies. """ if self.node.op in ('placeholder', 'get_attr', 'output'): return None if self.node.op == 'call_module': target = self.node.graph.owning_module.get_submodule(self.node.target) elif self.node.op == 'call_function': target = self.node.target elif self.node.op == 'call_method': target = getattr(self.node.args[0]._meta_data.__class__, self.node.target) else: raise ValueError(f'Unsupported node type: {self.node.op}') return target def register_strategy(self, compute_resharding_cost: bool = True) -> StrategiesVector: """ Register different sharding strategies for the current node. """ strategy_generators = self.get_strategy_generator() for generator in strategy_generators: strategies = generator.generate() # postprocess a strategy # postprocess can produce one strategy or multiple strategies post_processed_strategies_map = map(self.post_process, strategies) post_processed_strategies = [] for strategy in post_processed_strategies_map: if isinstance(strategy, (list, tuple)): post_processed_strategies.extend(strategy) else: post_processed_strategies.append(strategy) # compute the resharding costs based on the previous node # strategies if specified if compute_resharding_cost: updated_strategies = map(self.update_resharding_cost, post_processed_strategies) post_processed_strategies = list(updated_strategies) self.strategies_vector.extend(post_processed_strategies) # validating the correctness of the sharding strategy for strategy in self.strategies_vector: for op_data, sharding_spec in strategy.sharding_specs.items(): if op_data.data is not None and isinstance(op_data.data, torch.Tensor): check_sharding_spec_validity(sharding_spec, op_data.data) remove_strategy_list = [] for strategy in self.strategies_vector: shard_axis_list = [] last_axis = len(self.device_mesh.mesh_shape) - 1 for op_data, sharding_spec in strategy.sharding_specs.items(): if op_data.data is not None and isinstance(op_data.data, torch.Tensor): for dim, shard_axes in sharding_spec.dim_partition_dict.items(): for shard_axis in shard_axes: if shard_axis not in shard_axis_list: shard_axis_list.append(shard_axis) shard_level = len(shard_axis_list) using_last_axis = last_axis in shard_axis_list or -1 in shard_axis_list if self.shard_option == ShardOption.SHARD and shard_level == 0: remove_strategy_list.append(strategy) if self.shard_option == ShardOption.FULL_SHARD and shard_level <= 1: remove_strategy_list.append(strategy) if self.shard_option == ShardOption.SHARD_LAST_AXIS: if shard_level != 1 or using_last_axis == False: remove_strategy_list.append(strategy) for strategy in remove_strategy_list: self.strategies_vector.remove(strategy) return self.strategies_vector def post_process(self, strategy: ShardingStrategy) -> Union[ShardingStrategy, List[ShardingStrategy]]: # tranform the strategy generated # e.g. to process the sharding strategy for the transposed weights return strategy @abstractmethod def get_strategy_generator(self) -> List[StrategyGenerator]: """ Define which generators should be used by this NodeHandler object. """ pass @abstractmethod def get_operation_data_mapping(self) -> Dict[str, OperationData]: """ Returns the mapping between the logical operation data to its physical data. A logical operation data is a data associated with an operation, which can be input and output. It is defined by the strategy generator, for example, a matrix multiplication operation has two operands "input" and "other" and one result "output". For a nn.Linear module, the physical operand for "input" is the module input, the physical operand for "other" is the module weight, and the physical result for "output" is the module output. Note that the operand name is specified by the StrategyGenerator object. For example: # for a linear layer mapping = { "input": Operand(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data), "other": Operand(name="weight", type=OperationDataType.PARAM, data=self.named_parameters['weight']), "bias": Operand(name="bias", type=OperationDataType.PARAM, data=self.named_parameters['bias']), "output": Operand(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data), } """ pass class MetaInfoNodeHandler(NodeHandler): """ This is a base class to handle the nodes patched in the meta profiler. Note: this class will be integrated into the NodeHandler class in the future, after all the functions are patched. """ def register_strategy(self, compute_resharding_cost: bool = True) -> StrategiesVector: """ This method is inherited from NodeHandler. It will register the strategies first, and rewrite the memory_cost and compute_cost of the strategy using the MetaInfo class. """ super().register_strategy(compute_resharding_cost=compute_resharding_cost) target = self.get_target_function() # Currently we haven't patched all the torch functions and modules, so if the target # is not patched, we will use the default cost model to compute the cost. # TODO: patch all torch functions and modules to make it clean if meta_register.has(target.__class__) or meta_register.has(target): metainfo_vector = [] for strategy in self.strategies_vector: metainfo = MetaInfo(strategy, target) strategy.compute_cost = metainfo.compute_cost strategy.memory_cost = metainfo.memory_cost metainfo_vector.append(metainfo) # attach metainfos to the handler setattr(self, "metainfo_vector", metainfo_vector) else: logger = get_dist_logger() logger.warning(f'The target function {target} is not patched yet, ') return self.strategies_vector class ModuleHandler(NodeHandler): def __init__(self, *args, **kwargs) -> None: super().__init__(*args, **kwargs) # set attributes to access module parameters for convenience assert self.node.graph.owning_module is not None, \ f'The graph is not associated with a module, please make sure it can be used to instantiate a GraphModule object.' module = self.node.graph.owning_module.get_submodule(self.node.target) named_parameters = list(module.named_parameters(recurse=False)) named_buffers = list(module.named_buffers(recurse=False)) # convert named parameters from list to dict named_parameters = {k: v for k, v in named_parameters} named_buffers = {k: v for k, v in named_buffers} self.module = module self.named_parameters = named_parameters self.named_buffers = named_buffers class MetaInfoModuleHandler(ModuleHandler): """ This is a base class to handle the module patched in the meta profiler. Note: this class will be integrated into the ModuleHandler class in the future, after all the modules are patched. """ def register_strategy(self, compute_resharding_cost: bool = True) -> StrategiesVector: """ This method is inherited from NodeHandler. It will register the strategies first, and rewrite the memory_cost and compute_cost of the strategy using the MetaInfo class. """ super().register_strategy(compute_resharding_cost=compute_resharding_cost) target = self.get_target_function() # Currently we haven't patched all the torch functions and modules, so if the target # is not patched, we will use the default cost model to compute the cost. # TODO: patch all torch functions and modules to make it clean if meta_register.has(target.__class__) or meta_register.has(target): metainfo_vector = [] for strategy in self.strategies_vector: metainfo = MetaInfo(strategy, target) strategy.compute_cost = metainfo.compute_cost strategy.memory_cost = metainfo.memory_cost metainfo_vector.append(metainfo) # attach metainfos to the handler setattr(self, "metainfo_vector", metainfo_vector) else: logger = get_dist_logger() logger.warning(f'The target function {target} is not patched yet') return self.strategies_vector
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import MetaInfoNodeHandler, NodeHandler from .registry import operator_registry from .strategy import StrategyGenerator, UnaryElementwiseGenerator __all__ = ['UnaryElementwiseHandler'] @operator_registry.register(torch.Tensor.to) @operator_registry.register(torch.Tensor.type) @operator_registry.register(torch.abs) @operator_registry.register(torch.nn.ReLU) @operator_registry.register(torch.nn.Tanh) @operator_registry.register(torch.tanh) @operator_registry.register(torch.nn.modules.dropout.Dropout) @operator_registry.register(torch.Tensor.contiguous) @operator_registry.register(torch.nn.functional.dropout) class UnaryElementwiseHandler(MetaInfoNodeHandler): """ A UnaryElementwiseHandler which deals with the sharding strategies for UnaryElementwise Op. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(UnaryElementwiseGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"input": physical_input_operand, "output": physical_output} return mapping
from typing import Dict, List import torch from colossalai.auto_parallel.meta_profiler.metainfo import MetaInfo from ..sharding_strategy import OperationData, OperationDataType, StrategiesVector from .node_handler import MetaInfoModuleHandler, ModuleHandler from .registry import operator_registry from .strategy import BatchNormStrategyGenerator, StrategyGenerator __all__ = ['BatchNormModuleHandler'] @operator_registry.register(torch.nn.BatchNorm1d) @operator_registry.register(torch.nn.BatchNorm2d) @operator_registry.register(torch.nn.BatchNorm3d) class BatchNormModuleHandler(MetaInfoModuleHandler): """ A BatchNormModuleHandler which deals with the sharding strategies for nn.BatchNormXd module. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(BatchNormStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) physical_other_operand = OperationData(name="weight", type=OperationDataType.PARAM, data=self.named_parameters['weight'], logical_shape=self.named_parameters['weight'].shape) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) physical_running_mean_operand = OperationData(name="running_mean", type=OperationDataType.BUFFER, data=self.named_buffers['running_mean'], logical_shape=self.named_buffers['running_mean'].shape) physical_running_var_operand = OperationData(name="running_var", type=OperationDataType.BUFFER, data=self.named_buffers['running_var'], logical_shape=self.named_buffers['running_var'].shape) physical_num_batches_tracked_operand = OperationData( name="num_batches_tracked", type=OperationDataType.BUFFER, data=self.named_buffers['num_batches_tracked'], logical_shape=self.named_buffers['num_batches_tracked'].shape) mapping = { "input": physical_input_operand, "other": physical_other_operand, "output": physical_output, "running_mean": physical_running_mean_operand, "running_var": physical_running_var_operand, "num_batches_tracked": physical_num_batches_tracked_operand } if self.named_parameters['bias'] is not None: physical_bias_operand = OperationData(name="bias", type=OperationDataType.PARAM, data=self.named_parameters['bias']) mapping['bias'] = physical_bias_operand return mapping
from typing import Dict, List, Union import torch from torch.fx.node import Node from colossalai.auto_parallel.tensor_shard.sharding_strategy import OperationData, OperationDataType, ShardingStrategy from colossalai.tensor.shape_consistency import CollectiveCommPattern, CommSpec, ShapeConsistencyManager from ..constants import BCAST_FUNC_OP from ..utils import comm_actions_for_oprands, recover_sharding_spec_for_broadcast_shape from .node_handler import MetaInfoNodeHandler, NodeHandler from .registry import operator_registry from .strategy import BinaryElementwiseStrategyGenerator, StrategyGenerator __all__ = ['BinaryElementwiseHandler'] @operator_registry.register(BCAST_FUNC_OP) class BinaryElementwiseHandler(MetaInfoNodeHandler): """ An BinaryBcastOpHandler is a node handler which deals with operations which have two operands and broadcasting occurs such as torch.add. """ def get_operation_data_mapping(self) -> Dict[str, OperationData]: bcast_shape = self.node._meta_data.shape def _get_op_data_type(tensor): if isinstance(tensor, torch.nn.parameter.Parameter): return OperationDataType.PARAM else: return OperationDataType.ARG def _get_arg_value(idx): non_tensor = False if isinstance(self.node.args[idx], Node): meta_data = self.node.args[idx]._meta_data # The meta_data of node type argument could also possibly be a non-tensor object. if not isinstance(meta_data, torch.Tensor): assert isinstance(meta_data, (int, float)) meta_data = torch.Tensor([meta_data]).to('meta') non_tensor = True else: # this is in fact a real data like int 1 # but we can deem it as meta data # as it won't affect the strategy generation assert isinstance(self.node.args[idx], (int, float)) meta_data = torch.Tensor([self.node.args[idx]]).to('meta') non_tensor = True return meta_data, non_tensor input_meta_data, non_tensor_input = _get_arg_value(0) other_meta_data, non_tensor_other = _get_arg_value(1) output_meta_data = self.node._meta_data # we need record op_data with non-tensor data in this list, # and filter the non-tensor op_data in post_process. self.non_tensor_list = [] # assert False input_op_data = OperationData(name=str(self.node.args[0]), type=_get_op_data_type(input_meta_data), data=input_meta_data, logical_shape=bcast_shape) other_op_data = OperationData(name=str(self.node.args[1]), type=_get_op_data_type(other_meta_data), data=other_meta_data, logical_shape=bcast_shape) output_op_data = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_meta_data, logical_shape=bcast_shape) if non_tensor_input: self.non_tensor_list.append(input_op_data) if non_tensor_other: self.non_tensor_list.append(other_op_data) mapping = {'input': input_op_data, 'other': other_op_data, 'output': output_op_data} return mapping def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(BinaryElementwiseStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def post_process(self, strategy: ShardingStrategy) -> Union[ShardingStrategy, List[ShardingStrategy]]: # convert bias from its logical sharding spec to its physical sharding spec op_data_mapping = self.get_operation_data_mapping() for op_name, op_data in op_data_mapping.items(): if op_data in self.non_tensor_list: # remove the sharding spec if the op_data is not a tensor, e.g. torch.pow(tensor, 2) strategy.sharding_specs.pop(op_data) else: # convert the logical sharding spec to physical sharding spec if broadcast # e.g. torch.rand(4, 4) + torch.rand(4) physical_shape = op_data.data.shape logical_shape = op_data.logical_shape sharding_spec = strategy.get_sharding_spec_by_name(op_data.name) sharding_spec, removed_dims = recover_sharding_spec_for_broadcast_shape( sharding_spec, logical_shape, physical_shape) strategy.sharding_specs[op_data] = sharding_spec if len(removed_dims) > 0: comm_action = comm_actions_for_oprands(node=self.node, removed_dims=removed_dims, op_data=op_data, sharding_spec=sharding_spec) strategy.communication_actions[op_data] = comm_action return strategy
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import MetaInfoModuleHandler, ModuleHandler from .registry import operator_registry from .strategy import LayerNormGenerator, StrategyGenerator __all__ = ['LayerNormModuleHandler'] @operator_registry.register(torch.nn.LayerNorm) class LayerNormModuleHandler(MetaInfoModuleHandler): """ A LayerNormModuleHandler which deals with the sharding strategies for nn.LayerNorm module. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(LayerNormGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) physical_other_operand = OperationData(name="weight", type=OperationDataType.PARAM, data=self.named_parameters['weight'], logical_shape=self.named_parameters['weight'].shape) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} if self.named_parameters['bias'] is not None: physical_bias_operand = OperationData(name="bias", type=OperationDataType.PARAM, data=self.named_parameters['bias']) mapping['bias'] = physical_bias_operand return mapping
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import SplitGenerator, StrategyGenerator __all__ = ['SplitHandler'] @operator_registry.register(torch.Tensor.split) @operator_registry.register(torch.split) class SplitHandler(NodeHandler): """ A SplitHandler which deals with the sharding strategies for torch.permute or torch.split. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(SplitGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # check if the input operand is a parameter if isinstance(self.node.args[0]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG input_data = self.node.args[0]._meta_data physical_input_operand = OperationData(name=str(self.node.args[0]), type=data_type, data=input_data) split_size = self.node.args[1] if len(self.node.args) == 3: # (input, split_size, split_dim) split_dim = self.node.args[2] else: if self.node.kwargs: split_dim = self.node.kwargs['dim'] else: split_dim = 0 num_dims = self.node.args[0]._meta_data.dim() # recover negative value to positive if split_dim < 0: split_dim += num_dims split_info = (split_size, split_dim) physical_shape_operand = OperationData(name='split_info', type=OperationDataType.ARG, data=split_info) output_data = self.node._meta_data physical_output_operand = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data) mapping = { "input": physical_input_operand, "split_info": physical_shape_operand, "output": physical_output_operand } return mapping
from typing import Dict, List import torch import torch.nn.functional as F from ..sharding_strategy import OperationData, OperationDataType, ShardingStrategy, StrategiesVector from ..utils import transpose_partition_dim from .node_handler import MetaInfoModuleHandler, MetaInfoNodeHandler, ModuleHandler, NodeHandler from .registry import operator_registry from .strategy import ConvStrategyGenerator, StrategyGenerator __all__ = ['ConvModuleHandler', 'ConvFunctionHandler'] @operator_registry.register(torch.nn.Conv1d) @operator_registry.register(torch.nn.Conv2d) @operator_registry.register(torch.nn.Conv3d) class ConvModuleHandler(MetaInfoModuleHandler): """ A ConvModuleHandler which deals with the sharding strategies for nn.Convxd module. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(ConvStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) logical_shape_for_weight = list(self.named_parameters["weight"].shape) logical_shape_for_weight[0], logical_shape_for_weight[1] = logical_shape_for_weight[ 1], logical_shape_for_weight[0] physical_other_operand = OperationData(name="weight", type=OperationDataType.PARAM, data=self.named_parameters['weight'], logical_shape=torch.Size(logical_shape_for_weight)) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} if "bias" in self.named_parameters: physical_bias_operand = OperationData(name="bias", type=OperationDataType.PARAM, data=self.named_parameters['bias']) mapping['bias'] = physical_bias_operand return mapping def post_process(self, strategy: ShardingStrategy): """ Convert the sharding spec of the weight parameter back to its original shape. """ for op_data, sharding_spec in strategy.input_sharding_specs.items(): if op_data.name == "weight": transpose_partition_dim(sharding_spec, 0, 1) return strategy @operator_registry.register(F.conv1d) @operator_registry.register(F.conv2d) @operator_registry.register(F.conv3d) class ConvFunctionHandler(MetaInfoNodeHandler): """ A ConvFunctionHandler which deals with the sharding strategies for nn.functional.ConvXd functions. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(ConvStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) # check if the other operand is a parameter if isinstance(self.node.args[1]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG logical_shape_for_weight = list(self.node.args[1]._meta_data.shape) logical_shape_for_weight[0], logical_shape_for_weight[1] = logical_shape_for_weight[ 1], logical_shape_for_weight[0] physical_other_operand = OperationData(name=str(self.node.args[1]), type=data_type, data=self.node.args[1]._meta_data, logical_shape=torch.Size(logical_shape_for_weight)) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} if "bias" in self.node.kwargs and self.node.kwargs['bias'] is not None: # check if the other operand is a parameter if isinstance(self.node.kwargs["bias"]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG physical_bias_operand = OperationData(name=str(self.node.kwargs["bias"]), type=data_type, data=self.node.kwargs["bias"]._meta_data) mapping['bias'] = physical_bias_operand return mapping def post_process(self, strategy: ShardingStrategy): """ Convert the sharding spec of the weight parameter back to its original shape. """ for op_data, sharding_spec in strategy.input_sharding_specs.items(): if op_data.name == str(self.node.args[1]): transpose_partition_dim(sharding_spec, 0, 1) return strategy
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import MetaInfoModuleHandler, ModuleHandler from .registry import operator_registry from .strategy import NormalPoolStrategyGenerator, StrategyGenerator __all__ = ['NormPoolingHandler'] @operator_registry.register(torch.nn.MaxPool1d) @operator_registry.register(torch.nn.MaxPool2d) @operator_registry.register(torch.nn.MaxPool1d) @operator_registry.register(torch.nn.AvgPool1d) @operator_registry.register(torch.nn.AvgPool2d) @operator_registry.register(torch.nn.AvgPool3d) class NormPoolingHandler(MetaInfoModuleHandler): """ A NormPoolingHandler which deals with the sharding strategies for nn.MaxPoolxd module. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(NormalPoolStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) physical_weight_operand = OperationData(name="kernel", type=OperationDataType.ARG, data=self.module.kernel_size) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"input": physical_input_operand, "other": physical_weight_operand, "output": physical_output} return mapping
import operator from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import StrategyGenerator, TensorStrategyGenerator, TensorTupleStrategyGenerator __all__ = ['GetItemHandler'] @operator_registry.register(operator.getitem) class GetItemHandler(NodeHandler): """ A GetItemHandler which deals with the sharding strategies for operator.getitem. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] if isinstance(op_data_mapping["input"].data, torch.Tensor): generators.append(TensorStrategyGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) else: generators.append(TensorTupleStrategyGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) physical_other_operand = OperationData(name="index", type=OperationDataType.ARG, data=self.node.args[1]) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"input": physical_input_operand, "index": physical_other_operand, "output": physical_output} return mapping
class Registry: # TODO: refactor the registry classes used in colossalai.registry, colossalai.fx and here def __init__(self, name): self.name = name self.store = {} def register(self, source): def wrapper(func): if isinstance(source, (list, tuple)): # support register a list of items for this func for element in source: self.store[element] = func else: self.store[source] = func return func return wrapper def get(self, source): assert source in self.store, f'{source} not found in the {self.name} registry' target = self.store[source] return target def has(self, source): return source in self.store operator_registry = Registry('operator')
import copy import operator from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType, ShardingStrategy, StrategiesVector from ..utils import recover_sharding_spec_for_broadcast_shape from .node_handler import NodeHandler from .registry import operator_registry from .strategy import StrategyGenerator, WhereGenerator __all__ = ['WhereHandler'] @operator_registry.register(torch.where) class WhereHandler(NodeHandler): """ A WhereHandler which deals with the sharding strategies for torch.where. """ def get_strategy_generator(self) -> List[StrategyGenerator]: logical_op_data_mapping, _ = self.get_operation_data_mapping() generators = [] generators.append(WhereGenerator(logical_op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_condition_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data) physical_x_operand = OperationData(name=str(self.node.args[1]), type=OperationDataType.ARG, data=self.node.args[1]._meta_data) physical_y_operand = OperationData(name=str(self.node.args[2]), type=OperationDataType.ARG, data=self.node.args[2]._meta_data) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) physical_mapping = { "condition": physical_condition_operand, "x": physical_x_operand, "y": physical_y_operand, "output": physical_output } logical_shape_for_all = self.node._meta_data.shape logical_mapping = {} for key, physical_operand in physical_mapping.items(): logical_mapping[key] = self.convert_physical_operand_to_logical_operand(physical_operand, logical_shape_for_all) return logical_mapping, physical_mapping def convert_physical_operand_to_logical_operand(self, physical_operand, target_shape): logical_operand = copy.deepcopy(physical_operand) logical_operand.logical_shape = target_shape return logical_operand def post_process(self, strategy: ShardingStrategy): logical_op_data_mapping, physical_op_data_mapping = self.get_operation_data_mapping() for key in logical_op_data_mapping.keys(): logical_sharding_spec = strategy.sharding_specs[logical_op_data_mapping[key]] logical_shape = logical_op_data_mapping[key].logical_shape physical_shape = physical_op_data_mapping[key].logical_shape physical_sharding_spec, removed_dims = recover_sharding_spec_for_broadcast_shape( logical_sharding_spec, logical_shape, physical_shape) strategy.sharding_specs.pop(logical_op_data_mapping[key]) strategy.sharding_specs[physical_op_data_mapping[key]] = physical_sharding_spec strategy.name = f"{strategy.sharding_specs[physical_op_data_mapping['output']].sharding_sequence} = {strategy.sharding_specs[physical_op_data_mapping['condition']].sharding_sequence} x {strategy.sharding_specs[physical_op_data_mapping['x']].sharding_sequence} x {strategy.sharding_specs[physical_op_data_mapping['y']].sharding_sequence}" return strategy
from typing import Dict, List, Union import torch import torch.nn.functional as F from colossalai.auto_parallel.tensor_shard.utils import update_partition_dim from colossalai.logging import get_dist_logger from colossalai.tensor.sharding_spec import ShardingNotDivisibleError from ..sharding_strategy import OperationData, OperationDataType, ShardingStrategy from .node_handler import ModuleHandler, NodeHandler from .registry import operator_registry from .strategy import EmbeddingStrategyGenerator, StrategyGenerator __all__ = ['EmbeddingModuleHandler', 'EmbeddingFunctionHandler'] def _convert_logical_sharding_to_physical_sharding_spec_for_embedding(strategy: ShardingStrategy, input_name: str, output_name: str) -> List[ShardingStrategy]: """ This function converts the logical sharding spec to the physical sharding spec for both the input and output of the embedding operation. Args: strategy (ShardingStrategy): the logical strategy generated by the strategy generator. input_name (str): the name of the OperationData object for the input. output_name (str): the name of the OperationData object for the output. """ # the result will be a list of strategies sharding_strategies = [] # get operation data input_op_data = strategy.get_op_data_by_name(input_name) output_op_data = strategy.get_op_data_by_name(output_name) input_sharding_spec = strategy.get_sharding_spec_by_name(input_op_data.name) output_sharding_spec = strategy.get_sharding_spec_by_name(output_op_data.name) # recover the last logical dimension to physical dimension last_logical_output_dims = len(output_op_data.logical_shape) - 1 last_physical_output_dims = output_op_data.data.dim() - 1 # get logger for debug message logger = get_dist_logger() # For the input of the embedding operation, it can be multi-dimensional. The sharding spec is only generated for # logical 1D non-matrix dimension, the logical non-matrix dimension can belong to the 0th to Nth dimension of the # physical input shape. Thus, we enumerate to get all possible cases. if input_sharding_spec.dim_partition_dict: # if bool(input_sharding_spec.dim_partition_dict), it means that the # the generated sharding strategy does shard the non-matrix dimension, # in this case, we need to do enumeration num_input_dims = input_op_data.data.dim() for i in range(num_input_dims): strategy_copy = strategy.clone() input_sharding_spec = strategy_copy.get_sharding_spec_by_name(input_op_data.name) output_sharding_spec = strategy_copy.get_sharding_spec_by_name(output_op_data.name) try: # replace the 0th dimension in the logical sharding with ith dimension in the physical sharding update_partition_dim(sharding_spec=input_sharding_spec, dim_mapping={0: i}, physical_shape=input_op_data.data.shape, inplace=True) if last_logical_output_dims in output_sharding_spec.dim_partition_dict: dim_mapping = {0: i, last_logical_output_dims: last_physical_output_dims} else: dim_mapping = {0: i} update_partition_dim(sharding_spec=output_sharding_spec, dim_mapping=dim_mapping, physical_shape=output_op_data.data.shape, inplace=True) strategy_copy.name = f'{strategy.name}_{i}' sharding_strategies.append(strategy_copy) except ShardingNotDivisibleError as e: logger.debug( f'Errored occurred when converting the logical sharding spec to the physical one. Error details: {e}' ) else: # the generated sharding strategy does not shard the non-matrix dimension, # in this case, we don't need to do enumeration # but instead, we still need to convert the logical shape to physical shape strategy_copy = strategy.clone() input_sharding_spec = strategy_copy.get_sharding_spec_by_name(input_op_data.name) output_sharding_spec = strategy_copy.get_sharding_spec_by_name(output_op_data.name) # after updating, the logical shape will be replaced by the physical shape update_partition_dim(sharding_spec=input_sharding_spec, dim_mapping={}, physical_shape=input_op_data.data.shape, inplace=True) if last_logical_output_dims in output_sharding_spec.dim_partition_dict: dim_mapping = {last_logical_output_dims: last_physical_output_dims} else: dim_mapping = {} update_partition_dim(sharding_spec=output_sharding_spec, dim_mapping=dim_mapping, physical_shape=output_op_data.data.shape, inplace=True) sharding_strategies.append(strategy_copy) return sharding_strategies @operator_registry.register(torch.nn.Embedding) class EmbeddingModuleHandler(ModuleHandler): """ A EmbeddingModuleHandler which deals with the sharding strategies for nn.Embedding module. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(EmbeddingStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # In nn.Embedding operation, all the dimensions of input will be treated as the batch dimension, # and then the sharding spec will be generated based on the logical 1D tensor. # After that, the logical sharding info will be enumerated among all the physical dimensions. # Finally, the input will be transformed back to its original shape in self.post_process input_meta_data = self.node.args[0]._meta_data input_logical_shape = input_meta_data.view(-1).shape physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=input_meta_data, logical_shape=input_logical_shape) physical_other_operand = OperationData(name="weight", type=OperationDataType.PARAM, data=self.named_parameters['weight']) # Same as input, in nn.Embedding operation, all the dimensions of output will be treated as # (batch dimension, embedding dimension), and then the sharding spec will be generated based # on the logical 2D tensor. # After that, the logical sharding info of batch dimension will be enumerated among all the physical dimensions. # Finally, the output will be transformed back to its original shape in self.post_process output_meta_data = self.node._meta_data output_logical_shape = output_meta_data.view(-1, output_meta_data.shape[-1]).shape physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_meta_data, logical_shape=output_logical_shape) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} return mapping def post_process(self, strategy: ShardingStrategy) -> Union[ShardingStrategy, List[ShardingStrategy]]: """ Convert the sharding spec from the logical shape to the physical shape. """ # create multiple sharding strategies for the inputs # as input can be multi-dimensinal and the partition dim is only 2D, # we need to map the partition at logical dim 0 to one of the first few dimensions of the input and output strategies = _convert_logical_sharding_to_physical_sharding_spec_for_embedding(strategy=strategy, input_name=str( self.node.args[0]), output_name=str(self.node)) return strategies @operator_registry.register(F.embedding) class EmbeddingFunctionHandler(NodeHandler): """ A EmbeddingFunctionHandler which deals with the sharding strategies for F.embedding. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(EmbeddingStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # In F.embedding operation, all the dimensions of input will be treated as the batch dimension, # and then the sharding spec will be generated based on the logical 1D tensor. # After that, the logical sharding info will be enumerated among all the physical dimensions. # Finally, the input will be transformed back to its original shape in self.post_process input_meta_data = self.node.args[0]._meta_data input_logical_shape = input_meta_data.view(-1).shape physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data, logical_shape=input_logical_shape) # check if the other operand is a parameter if isinstance(self.node.args[1]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG physical_other_operand = OperationData(name=str(self.node.args[1]), type=data_type, data=self.node.args[1]._meta_data) # Same as input, in F.embedding operation, all the dimensions of output will be treated as # (batch dimension, embedding dimension), and then the sharding spec will be generated based # on the logical 2D tensor. # After that, the logical sharding info of batch dimension will be enumerated among all the physical dimensions. # Finally, the output will be transformed back to its original shape in self.post_process output_meta_data = self.node._meta_data output_logical_shape = output_meta_data.view(-1, output_meta_data.shape[-1]).shape physical_output = OperationData( name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data, logical_shape=output_logical_shape, ) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} return mapping def post_process(self, strategy: ShardingStrategy): """ Convert the sharding spec from the logical shape to the physical shape. """ # create multiple sharding strategies for the inputs # as input can be multi-dimensinal and the partition dim is only 2D, # we need to map the partition at logical dim 0 to one of the first few dimensions of the input and output strategies = _convert_logical_sharding_to_physical_sharding_spec_for_embedding(strategy=strategy, input_name=str( self.node.args[0]), output_name=str(self.node)) return strategies
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import PermuteGenerator, StrategyGenerator __all__ = ['PermuteHandler'] @operator_registry.register(torch.Tensor.permute) @operator_registry.register(torch.permute) class PermuteHandler(NodeHandler): """ A PermuteHandler which deals with the sharding strategies for torch.permute or torch.transpose. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(PermuteGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # check if the input operand is a parameter if isinstance(self.node.args[0]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG input_data = self.node.args[0]._meta_data physical_input_operand = OperationData(name=str(self.node.args[0]), type=data_type, data=input_data) permute_dims = [] if self.node.op == 'call_method': # torch.Tensor.permute (input, *dims) for arg in self.node.args: if isinstance(arg, torch.fx.Node): if isinstance(arg._meta_data, int): permute_dims.append(arg._meta_data) else: assert isinstance(arg, int), 'The argument in permute node should be either type of Node or int.' permute_dims.append(arg) else: # torch.permute (input, dims) for arg in self.node.args: if isinstance(arg, torch.fx.Node): if isinstance(arg._meta_data, (tuple, list)): permute_dims.extend(arg._meta_data) else: assert isinstance( arg, (tuple, list)), 'The argument in permute node should be type of Node, Tuple[int] or List[int].' permute_dims.extend(arg) num_dims = self.node._meta_data.dim() for i in range(num_dims): # recover negative value to positive if permute_dims[i] < 0: permute_dims[i] += num_dims physical_shape_operand = OperationData(name='permute_dims', type=OperationDataType.ARG, data=list(permute_dims)) output_data = self.node._meta_data physical_output_operand = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data) mapping = { "input": physical_input_operand, "permute_dims": physical_shape_operand, "output": physical_output_operand } return mapping
from .addmm_handler import ADDMMFunctionHandler from .batch_norm_handler import BatchNormModuleHandler from .binary_elementwise_handler import BinaryElementwiseHandler from .bmm_handler import AddBMMFunctionHandler, BMMFunctionHandler from .conv_handler import ConvFunctionHandler, ConvModuleHandler from .default_reshape_handler import DefaultReshapeHandler from .embedding_handler import EmbeddingFunctionHandler, EmbeddingModuleHandler from .getattr_handler import GetattrHandler from .getitem_handler import GetItemHandler from .layer_norm_handler import LayerNormModuleHandler from .linear_handler import LinearFunctionHandler, LinearModuleHandler from .matmul_handler import MatMulHandler from .normal_pooling_handler import NormPoolingHandler from .output_handler import OutputHandler from .permute_handler import PermuteHandler from .placeholder_handler import PlaceholderHandler from .registry import operator_registry from .softmax_handler import SoftmaxHandler from .split_handler import SplitHandler from .sum_handler import SumHandler from .tensor_constructor_handler import TensorConstructorHandler from .transpose_handler import TransposeHandler from .unary_elementwise_handler import UnaryElementwiseHandler from .view_handler import ViewHandler from .where_handler import WhereHandler __all__ = [ 'LinearFunctionHandler', 'LinearModuleHandler', 'BMMFunctionHandler', 'AddBMMFunctionHandler', 'LayerNormModuleHandler', 'BatchNormModuleHandler', 'ConvModuleHandler', 'ConvFunctionHandler', 'UnaryElementwiseHandler', 'DefaultReshapeHandler', 'PlaceholderHandler', 'OutputHandler', 'WhereHandler', 'NormPoolingHandler', 'BinaryElementwiseHandler', 'MatMulHandler', 'operator_registry', 'ADDMMFunctionHandler', 'GetItemHandler', 'GetattrHandler', 'ViewHandler', 'PermuteHandler', 'TensorConstructorHandler', 'EmbeddingModuleHandler', 'EmbeddingFunctionHandler', 'SumHandler', 'SoftmaxHandler', 'TransposeHandler', 'SplitHandler' ]
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import StrategyGenerator, ViewGenerator __all__ = ['ViewHandler'] @operator_registry.register(torch.Tensor.reshape) @operator_registry.register(torch.reshape) @operator_registry.register(torch.Tensor.view) class ViewHandler(NodeHandler): """ A ViewHandler which deals with the sharding strategies for Reshape Op, such as torch.reshape. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(ViewGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process # check if the input operand is a parameter if isinstance(self.node.args[0]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG input_data = self.node.args[0]._meta_data physical_input_operand = OperationData(name=str(self.node.args[0]), type=data_type, data=input_data) target_shape = self.node._meta_data.shape physical_shape_operand = OperationData(name='tgt_shape', type=OperationDataType.ARG, data=target_shape) output_data = self.node._meta_data physical_output_operand = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data) mapping = { "input": physical_input_operand, "tgt_shape": physical_shape_operand, "output": physical_output_operand } return mapping
from typing import Dict, List, Union import torch import torch.nn.functional as F from colossalai.auto_parallel.tensor_shard.utils import ( check_sharding_spec_validity, transpose_partition_dim, update_partition_dim, ) from colossalai.logging import get_dist_logger from colossalai.tensor.sharding_spec import ShardingNotDivisibleError from ..sharding_strategy import OperationData, OperationDataType, ShardingStrategy, StrategiesVector from .node_handler import MetaInfoModuleHandler, MetaInfoNodeHandler, ModuleHandler, NodeHandler from .registry import operator_registry from .strategy import LinearProjectionStrategyGenerator, StrategyGenerator __all__ = ['LinearModuleHandler', 'LinearFunctionHandler'] def _update_sharding_spec_for_transposed_weight_for_linear(strategy: ShardingStrategy, weight_name: str) -> ShardingStrategy: """ This function is a helper function used by both module node handler and function node handler. This function will convert the sharding spec for the transposed weight to the correct partititon spec. Args: strategy (ShardingStrategy): the strategy generated by the strategy generator. weight_name (str): the name of the OperationData object for the weight. """ # switch the dimensions of the transposed weight sharding_spec = strategy.get_sharding_spec_by_name(weight_name) op_data = strategy.get_op_data_by_name(weight_name) assert op_data.logical_shape[0] == op_data.data.shape[1] and \ op_data.logical_shape[1] == op_data.data.shape[0], \ "Expected the logical shape of the linear operator's weight is equal to transposed physical shape" dim_size = len(op_data.logical_shape) transpose_partition_dim(sharding_spec, 0, dim_size - 1) return strategy def _convert_logical_sharding_to_physical_sharding_spec_for_linear(strategy: ShardingStrategy, input_name: str, output_name: str) -> List[ShardingStrategy]: """ This function converts the logical sharding spec to the physical sharding spec for both the input and output of the linear operation. The input and output should have the same sharding spec. Args: strategy (ShardingStrategy): the logical strategy generated by the strategy generator. input_name (str): the name of the OperationData object for the input. output_name (str): the name of the OperationData object for the output. """ # the result will be a list of strategies sharding_strategies = [] # get operation data input_op_data = strategy.get_op_data_by_name(input_name) output_op_data = strategy.get_op_data_by_name(output_name) input_sharding_spec = strategy.get_sharding_spec_by_name(input_op_data.name) output_sharding_spec = strategy.get_sharding_spec_by_name(output_op_data.name) # recover the last logical dimension to physical dimension last_logical_input_dims = len(input_op_data.logical_shape) - 1 last_logical_output_dims = len(output_op_data.logical_shape) - 1 last_physical_input_dims = input_op_data.data.dim() - 1 last_physical_output_dims = output_op_data.data.dim() - 1 if last_logical_input_dims in input_sharding_spec.dim_partition_dict: input_last_dim_mapping = {last_logical_input_dims: last_physical_input_dims} else: input_last_dim_mapping = {} if last_logical_output_dims in output_sharding_spec.dim_partition_dict: output_last_dim_mapping = {last_logical_output_dims: last_physical_output_dims} else: output_last_dim_mapping = {} # get logger for debug message logger = get_dist_logger() # for the input of the linear operation, it can be multi-dimensional. The sharding spec generated is only # 2D, where the first dimension is non-matrix dimension and the last dimension is the matrix dimension. # the logical non-matrix dimension can belong to the 0th to (N-1)th dimension of the physical input shape. # Thus, we enumerate to get all possible cases. if 0 in input_sharding_spec.dim_partition_dict: # if 0 is in the dim_partition_dict, it means that the # the generated sharding strategy does shard the non-matrix dimension, # in this case, we need to do enumeration num_input_dims = input_op_data.data.dim() for i in range(num_input_dims - 1): strategy_copy = strategy.clone() input_sharding_spec = strategy_copy.get_sharding_spec_by_name(input_op_data.name) output_sharding_spec = strategy_copy.get_sharding_spec_by_name(output_op_data.name) try: # replace the 0th dimension in the logical sharding with ith dimension in the physical sharding input_dim_mapping = {0: i} input_dim_mapping.update(input_last_dim_mapping) update_partition_dim(sharding_spec=input_sharding_spec, dim_mapping=input_dim_mapping, physical_shape=input_op_data.data.shape, inplace=True) output_dim_mapping = {0: i} output_dim_mapping.update(output_last_dim_mapping) update_partition_dim(sharding_spec=output_sharding_spec, dim_mapping=output_dim_mapping, physical_shape=output_op_data.data.shape, inplace=True) strategy_copy.name = f'{strategy.name}_{i}' sharding_strategies.append(strategy_copy) except ShardingNotDivisibleError as e: logger.debug( f'Errored occurred when converting the logical sharding spec to the physical one. Error details: {e}' ) else: # the generated sharding strategy does not shard the non-matrix dimension, # in this case, we don't need to do enumeration # but instead, we still need to convert the logical shape to physical shape strategy_copy = strategy.clone() input_sharding_spec = strategy_copy.get_sharding_spec_by_name(input_op_data.name) output_sharding_spec = strategy_copy.get_sharding_spec_by_name(output_op_data.name) # after updating, the logical shape will be replaced by the physical shape input_dim_mapping = {} input_dim_mapping.update(input_last_dim_mapping) update_partition_dim(sharding_spec=input_sharding_spec, dim_mapping=input_dim_mapping, physical_shape=input_op_data.data.shape, inplace=True) output_dim_mapping = {} output_dim_mapping.update(output_last_dim_mapping) update_partition_dim(sharding_spec=output_sharding_spec, dim_mapping=output_dim_mapping, physical_shape=output_op_data.data.shape, inplace=True) sharding_strategies.append(strategy_copy) return sharding_strategies @operator_registry.register(torch.nn.Linear) class LinearModuleHandler(MetaInfoModuleHandler): """ A LinearModuleHandler which deals with the sharding strategies for nn.Linear module. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append( LinearProjectionStrategyGenerator(op_data_mapping, self.device_mesh, linear_projection_type='linear', solver_perference=self.solver_perference)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process input_meta_data = self.node.args[0]._meta_data input_logical_shape = input_meta_data.view(-1, input_meta_data.shape[-1]).shape physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=input_meta_data, logical_shape=input_logical_shape) physical_other_operand = OperationData(name="weight", type=OperationDataType.PARAM, data=self.named_parameters['weight'], logical_shape=self.named_parameters['weight'].shape[::-1]) output_meta_data = self.node._meta_data output_logical_shape = output_meta_data.view(-1, output_meta_data.shape[-1]).shape physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_meta_data, logical_shape=output_logical_shape) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} if 'bias' in self.named_parameters is not None: physical_bias_operand = OperationData(name="bias", type=OperationDataType.PARAM, data=self.named_parameters['bias']) mapping['bias'] = physical_bias_operand return mapping def post_process(self, strategy: ShardingStrategy) -> Union[ShardingStrategy, List[ShardingStrategy]]: """ Convert the sharding spec from the logical shape to the physical shape. In this function, two tasks are completed: 1. the sharding spec is updated for the transposed weight 2. the input and output sharding specs are updated to physical shape. """ # switch the dimensions of the transposed weight strategy = _update_sharding_spec_for_transposed_weight_for_linear(strategy=strategy, weight_name='weight') # create multiple sharding strategies for the inputs # as input can be multi-dimensinal and the partition dim is only 2D, # we need to map the partition at dim 0 to one of the first few dimensions of the input strategies = _convert_logical_sharding_to_physical_sharding_spec_for_linear(strategy=strategy, input_name=str(self.node.args[0]), output_name=str(self.node)) return strategies @operator_registry.register(F.linear) class LinearFunctionHandler(MetaInfoNodeHandler): """ A LinearFunctionHandler which deals with the sharding strategies for F.Linear. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append( LinearProjectionStrategyGenerator(op_data_mapping, self.device_mesh, linear_projection_type='linear')) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process input_meta_data = self.node.args[0]._meta_data input_logical_shape = input_meta_data.view(-1, input_meta_data.shape[-1]).shape physical_input_operand = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.node.args[0]._meta_data, logical_shape=input_logical_shape) # check if the other operand is a parameter if isinstance(self.node.args[1]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG physical_other_operand = OperationData(name=str(self.node.args[1]), type=data_type, data=self.node.args[1]._meta_data, logical_shape=self.node.args[1]._meta_data.shape[::-1]) output_meta_data = self.node._meta_data output_logical_shape = output_meta_data.view(-1, output_meta_data.shape[-1]).shape physical_output = OperationData( name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data, logical_shape=output_logical_shape, ) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} if 'bias' in self.node.kwargs and self.node.kwargs['bias'] is not None: # check if the other operand is a parameter if isinstance(self.node.kwargs["bias"]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG physical_bias_operand = OperationData(name=str(self.node.kwargs["bias"]), type=data_type, data=self.node.kwargs["bias"]._meta_data) mapping['bias'] = physical_bias_operand return mapping def post_process(self, strategy: ShardingStrategy): # switch the dimensions of the transposed weight strategy = _update_sharding_spec_for_transposed_weight_for_linear(strategy=strategy, weight_name=str(self.node.args[1])) # create multiple sharding strategies for the inputs # as input can be multi-dimensinal and the partition dim is only 2D, # we need to map the partition at dim 0 to one of the first few dimensions of the input strategies = _convert_logical_sharding_to_physical_sharding_spec_for_linear(strategy=strategy, input_name=str(self.node.args[0]), output_name=str(self.node)) return strategies
from typing import Dict, List, Union import torch from colossalai.tensor.shape_consistency import CollectiveCommPattern, CommSpec, ShapeConsistencyManager from ..sharding_strategy import CommAction, CommType, OperationData, OperationDataType, ShardingStrategy from ..utils import comm_actions_for_oprands, recover_sharding_spec_for_broadcast_shape from .node_handler import NodeHandler from .registry import operator_registry from .strategy import LinearProjectionStrategyGenerator, StrategyGenerator __all__ = ['ADDMMFunctionHandler'] @operator_registry.register(torch.addmm) @operator_registry.register(torch.Tensor.addmm) class ADDMMFunctionHandler(NodeHandler): """ This is a NodeHandler class which deals with the batched matrix multiplication operation in PyTorch. Such operations including `torch.bmm` and `torch.Tensor.bmm` require the tensor to be 3D, thus, there is no logical-physical shape conversion in this handler. """ def _infer_op_data_type(self, tensor: torch.Tensor) -> OperationDataType: if isinstance(tensor, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG return data_type def get_operation_data_mapping(self) -> Dict[str, OperationData]: # input operand input_data = self.node.args[1]._meta_data physical_input_operand = OperationData(name=str(self.node.args[1]), type=self._infer_op_data_type(input_data), data=input_data) # other operand other_data = self.node.args[2]._meta_data physical_other_operand = OperationData(name=str(self.node.args[2]), type=self._infer_op_data_type(other_data), data=other_data) # bias physical shape bias_logical_shape = self.node._meta_data.shape bias_data = self.node.args[0]._meta_data physical_bias_operand = OperationData(name=str(self.node.args[0]), type=self._infer_op_data_type(bias_data), data=bias_data, logical_shape=bias_logical_shape) # output physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = { "input": physical_input_operand, "other": physical_other_operand, "output": physical_output, 'bias': physical_bias_operand } return mapping def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append( LinearProjectionStrategyGenerator(op_data_mapping, self.device_mesh, linear_projection_type='addmm')) return generators def post_process(self, strategy: ShardingStrategy) -> Union[ShardingStrategy, List[ShardingStrategy]]: # convert bias from its logical sharding spec to its physical sharding spec op_data_mapping = self.get_operation_data_mapping() bias_op_data = op_data_mapping['bias'] bias_physical_shape = bias_op_data.data.shape bias_logical_shape = bias_op_data.logical_shape bias_sharding_spec = strategy.get_sharding_spec_by_name(bias_op_data.name) bias_sharding_spec, removed_dims = recover_sharding_spec_for_broadcast_shape( bias_sharding_spec, bias_logical_shape, bias_physical_shape) strategy.sharding_specs[bias_op_data] = bias_sharding_spec if len(removed_dims) > 0: comm_action = comm_actions_for_oprands(node=self.node, removed_dims=removed_dims, op_data=bias_op_data, sharding_spec=bias_sharding_spec) strategy.communication_actions[bias_op_data] = comm_action return strategy
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import MetaInfoNodeHandler, NodeHandler from .registry import operator_registry from .strategy import DefaultReshapeGenerator, StrategyGenerator __all__ = ['DefaultReshapeHandler'] @operator_registry.register(torch.flatten) @operator_registry.register(torch.Tensor.unsqueeze) @operator_registry.register(torch.nn.AdaptiveAvgPool2d) class DefaultReshapeHandler(MetaInfoNodeHandler): """ A DefaultReshapeHandler which deals with the sharding strategies for Reshape Op, such as torch.reshape. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(DefaultReshapeGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def infer_logical_shape(self, data): """ This function is used to infer logical shape for operands. Notes: This function is only used for the operands whose data are not only in type of tensor, such as tuple of tensor. """ if isinstance(data, torch.Tensor): return data.shape else: assert isinstance(data, tuple), "input_data should be a tuple of tensor or a tensor." logical_shape = [] for tensor in data: assert isinstance(tensor, torch.Tensor), "input_data should be a tuple of tensor or a tensor." logical_shape.append(tensor.shape) logical_shape = tuple(logical_shape) return logical_shape def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process # check if the input operand is a parameter if isinstance(self.node.args[0]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG input_data = self.node.args[0]._meta_data input_logical_shape = self.infer_logical_shape(input_data) physical_input_operand = OperationData(name=str(self.node.args[0]), type=data_type, data=input_data, logical_shape=input_logical_shape) output_data = self.node._meta_data output_logical_shape = self.infer_logical_shape(output_data) physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data, logical_shape=output_logical_shape) mapping = {"input": physical_input_operand, "output": physical_output} return mapping
from typing import Dict, List from torch.fx.node import Node from colossalai.device.device_mesh import DeviceMesh from ..sharding_strategy import OperationData, OperationDataType, StrategiesVector from .node_handler import NodeHandler from .strategy import PlaceholderGenerator, StrategyGenerator __all__ = ['PlaceholderHandler'] class PlaceholderHandler(NodeHandler): """ A PlaceholderHandler which deals with the sharding strategies for Placeholder Node. """ def __init__(self, node: Node, device_mesh: DeviceMesh, strategies_vector: StrategiesVector, placeholder_option: str) -> None: super().__init__(node, device_mesh, strategies_vector) self.placeholder_option = placeholder_option def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append( PlaceholderGenerator(op_data_mapping, self.device_mesh, placeholder_option=self.placeholder_option)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"output": physical_output} return mapping
from typing import Dict, List from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .strategy import GetattrGenerator, StrategyGenerator __all__ = ['GetattrHandler'] class GetattrHandler(NodeHandler): """ A GetattrHandler which deals with the sharding strategies for Getattr Node. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(GetattrGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process # There are only two possible types for get_attr node: # 1. torch.Tensor(torch.nn.Parameters or torch.nn.Buffers) # 2. torch.nn.Module # temporarily, we just support first case in Tracer, so we don't have to worry about # issue related to the node._meta_data type. physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping = {"output": physical_output} return mapping
from typing import Dict, List, Union import torch from colossalai.tensor.shape_consistency import CollectiveCommPattern, CommSpec, ShapeConsistencyManager from ..sharding_strategy import CommAction, CommType, OperationData, OperationDataType, ShardingStrategy from ..utils import comm_actions_for_oprands, recover_sharding_spec_for_broadcast_shape from .node_handler import NodeHandler from .registry import operator_registry from .strategy import BatchedMatMulStrategyGenerator, StrategyGenerator __all__ = ['BMMFunctionHandler', 'AddBMMFunctionHandler'] def _get_data_mapping_for_bmm_op(node, input_idx, other_idx, bias_idx=None): """ This function is a helper function which extracts the common logic for both `bmm` and `addbmm` node handler to reduce code redundancy. """ # input operand physical_input_operand = OperationData(name=str(node.args[input_idx]), type=OperationDataType.ARG, data=node.args[input_idx]._meta_data) # other operand physical_other_operand = OperationData(name=str(node.args[other_idx]), type=OperationDataType.ARG, data=node.args[other_idx]._meta_data) # output physical_output = OperationData(name=str(node), type=OperationDataType.OUTPUT, data=node._meta_data) mapping = {"input": physical_input_operand, "other": physical_other_operand, "output": physical_output} if bias_idx is not None: # bias physical shape bias_logical_shape = node._meta_data.shape physical_bias_operand = OperationData(name=str(node.args[bias_idx]), type=OperationDataType.ARG, data=node.args[bias_idx]._meta_data, logical_shape=bias_logical_shape) mapping['bias'] = physical_bias_operand return mapping @operator_registry.register(torch.bmm) @operator_registry.register(torch.Tensor.bmm) class BMMFunctionHandler(NodeHandler): """ This is a NodeHandler class which deals with the batched matrix multiplication operation in PyTorch. Such operations including `torch.bmm` and `torch.Tensor.bmm` require the tensor to be 3D, thus, there is no logical-physical shape conversion in this handler. """ def get_operation_data_mapping(self) -> Dict[str, OperationData]: mapping = _get_data_mapping_for_bmm_op(node=self.node, input_idx=0, other_idx=1) return mapping def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(BatchedMatMulStrategyGenerator(op_data_mapping, self.device_mesh)) return generators @operator_registry.register(torch.addbmm) @operator_registry.register(torch.Tensor.addbmm) class AddBMMFunctionHandler(NodeHandler): """ This is a NodeHandler class which deals with the addition + batched matrix multiplication operation in PyTorch. Such operations including `torch.addbmm` and `torch.Tensor.addbmm` require the two matmul tensor to be 3D. However, due to the addition, logical-physical shape conversion is required for the bias term. As the addbmm operation will reduce the batch dimension, the bias is maximum 2D. """ def get_operation_data_mapping(self) -> Dict[str, OperationData]: mapping = _get_data_mapping_for_bmm_op(node=self.node, input_idx=1, other_idx=2, bias_idx=0) return mapping def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(BatchedMatMulStrategyGenerator(op_data_mapping, self.device_mesh)) return generators def post_process(self, strategy: ShardingStrategy) -> Union[ShardingStrategy, List[ShardingStrategy]]: # convert bias from its logical sharding spec to its physical sharding spec op_data_mapping = self.get_operation_data_mapping() if 'bias' in op_data_mapping: bias_op_data = op_data_mapping['bias'] bias_physical_shape = bias_op_data.data.shape bias_logical_shape = bias_op_data.logical_shape bias_sharding_spec = strategy.get_sharding_spec_by_name(bias_op_data.name) bias_sharding_spec, removed_dims = recover_sharding_spec_for_broadcast_shape( bias_sharding_spec, bias_logical_shape, bias_physical_shape) strategy.sharding_specs[bias_op_data] = bias_sharding_spec if len(removed_dims) > 0: comm_action = comm_actions_for_oprands(node=self.node, removed_dims=removed_dims, op_data=bias_op_data, sharding_spec=bias_sharding_spec) strategy.communication_actions[bias_op_data] = comm_action return strategy
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import StrategyGenerator, TransposeGenerator __all__ = ['TransposeHandler'] @operator_registry.register(torch.Tensor.transpose) @operator_registry.register(torch.transpose) class TransposeHandler(NodeHandler): """ A TransposeHandler which deals with the sharding strategies for torch.permute or torch.transpose. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(TransposeGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # check if the input operand is a parameter if isinstance(self.node.args[0]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG input_data = self.node.args[0]._meta_data physical_input_operand = OperationData(name=str(self.node.args[0]), type=data_type, data=input_data) transpose_dims = [] # torch.transpose (input, dim0, dim1) for arg in self.node.args: if isinstance(arg, torch.fx.Node): if isinstance(arg._meta_data, int): transpose_dims.append(arg._meta_data) else: transpose_dims.append(arg) num_dims = self.node._meta_data.dim() for i in range(2): # recover negative value to positive if transpose_dims[i] < 0: transpose_dims[i] += num_dims physical_shape_operand = OperationData(name='transpose_dims', type=OperationDataType.ARG, data=list(transpose_dims)) output_data = self.node._meta_data physical_output_operand = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data) mapping = { "input": physical_input_operand, "transpose_dims": physical_shape_operand, "output": physical_output_operand } return mapping
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import SoftmaxGenerator, StrategyGenerator __all__ = ['SoftmaxHandler'] @operator_registry.register(torch.nn.Softmax) @operator_registry.register(torch.nn.functional.softmax) class SoftmaxHandler(NodeHandler): """ A SoftmaxHandler which deals with the sharding strategies for torch.nn.Softmax or torch.nn.functional.softmax. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(SoftmaxGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # check if the input operand is a parameter if isinstance(self.node.args[0]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG input_data = self.node.args[0]._meta_data physical_input_operand = OperationData(name=str(self.node.args[0]), type=data_type, data=input_data) softmax_dim = self.node.kwargs['dim'] num_dims = self.node.args[0]._meta_data.dim() # recover negative value to positive if softmax_dim < 0: softmax_dim += num_dims physical_dim_operand = OperationData(name='softmax_dim', type=OperationDataType.ARG, data=softmax_dim) output_data = self.node._meta_data physical_output_operand = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data) mapping = { "input": physical_input_operand, "softmax_dim": physical_dim_operand, "output": physical_output_operand } return mapping
from typing import Dict, List import torch from colossalai.device.device_mesh import DeviceMesh from ..sharding_strategy import OperationData, OperationDataType, StrategiesVector from .node_handler import NodeHandler from .strategy import OutputGenerator, StrategyGenerator __all__ = ['OutputHandler'] class OutputHandler(NodeHandler): """ A OutputHandler which deals with the sharding strategies for Output Node. """ def __init__(self, node: torch.fx.node.Node, device_mesh: DeviceMesh, strategies_vector: StrategiesVector, output_option: str) -> None: super().__init__(node, device_mesh, strategies_vector) self.output_option = output_option def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(OutputGenerator(op_data_mapping, self.device_mesh, self.predecessor_node, self.output_option)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # use transposed shape for strategies # the strategies will be transformed back to its original shape in self.post_process mapping = {} output_meta_data = [] for index, input_node in enumerate(self.predecessor_node): input_meta_data = input_node._meta_data physical_inputs = OperationData(name=str(input_node), type=OperationDataType.ARG, data=input_meta_data) name_key = f'input_{index}' mapping[name_key] = physical_inputs output_meta_data.append(input_meta_data) assert len(output_meta_data) > 0, f'Output node {self.node} has no input node.' if len(output_meta_data) == 1: output_meta_data = output_meta_data[0] else: output_meta_data = tuple(output_meta_data) self.node._meta_data = output_meta_data physical_output = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.node._meta_data) mapping["output"] = physical_output return mapping
import operator from abc import ABC, abstractmethod from copy import deepcopy from enum import Enum from functools import reduce from typing import Dict, List, Union import torch from colossalai.auto_parallel.tensor_shard.utils.broadcast import ( BroadcastType, get_broadcast_dim_info, get_broadcast_shape, ) from colossalai.tensor.sharding_spec import ShardingSpecException from ..sharding_strategy import OperationData, OperationDataType, ShardingStrategy from ..utils import recover_sharding_spec_for_broadcast_shape from .node_handler import MetaInfoNodeHandler, NodeHandler from .registry import operator_registry from .strategy import ( BatchedMatMulStrategyGenerator, DotProductStrategyGenerator, LinearProjectionStrategyGenerator, MatVecStrategyGenerator, StrategyGenerator, ) class MatMulType(Enum): """ The MatMulType is categorized into 4 types based on the reference of torch.matmul in https://pytorch.org/docs/stable/generated/torch.matmul.html. DOT: dot product, both tensors are 1D, these two tensors need to have the same number of elements MM: matrix-matrix product, both tensors are 2D or the 1st tensor is 1D and the 2nd tensor is 2D MV: matrix-vector product: the 1st tensor is 2D and the 2nd tensor is 1D BMM: batched matrix-matrix multiplication, one tensor is at least 1D and the other is at least 3D """ DOT = 0 MM = 1 MV = 2 BMM = 3 def get_matmul_type(input_dim: int, other_dim: int): """ Determine which type of matmul operation should be executed for the given tensor dimensions. Args: input_dim (int): the number of dimensions for the input tenosr other_dim (int): the number of dimensions for the other tenosr """ if input_dim == 1 and other_dim == 1: matmul_type = MatMulType.DOT elif input_dim in [1, 2] and other_dim == 2: matmul_type = MatMulType.MM elif input_dim == 2 and other_dim == 1: matmul_type = MatMulType.MV elif input_dim >= 1 and other_dim >= 1 and (input_dim > 2 or other_dim > 2): matmul_type = MatMulType.BMM else: raise ValueError( f"The input and other tensors are of {input_dim} and {other_dim} which cannot used to execute matmul operation" ) return matmul_type class BmmTransform(ABC): """ BmmTransform is an abstraction of the shape conversion between logical and physical operation data during the strategy generation. """ @abstractmethod def apply(self, shape_mapping: Dict[str, List[int]]): pass @abstractmethod def recover(self, op_data_mapping: Dict[str, OperationData], strategy: ShardingStrategy): pass class Padder(BmmTransform): """ Add padding to the matrix dimensions for batched matrix multiplication. """ def __init__(self) -> None: # keep the padding dim, op_name -> padded_dim self.padded_dim_mapping = {} def apply(self, shape_mapping: Dict[str, List[int]]): mapping_copy = deepcopy(shape_mapping) input_shape = mapping_copy['input'] other_shape = mapping_copy['other'] if len(input_shape) == 1: # if the input is a 1D tensor, 1 is prepended to its shape # and it will be removed afterwards input_shape.insert(0, 1) self.padded_dim_mapping['input'] = -2 self.padded_dim_mapping['output'] = -2 elif len(other_shape) == 1: # if the other is a 1D tensor, 1 is appended to its shape # and it will be removed afterwards other_shape = other_shape.append(1) self.padded_dim_mapping['other'] = -1 self.padded_dim_mapping['output'] = -1 return mapping_copy def recover(self, op_data_mapping: Dict[str, OperationData], strategy: ShardingStrategy): input_op_data = op_data_mapping['input'] other_op_data = op_data_mapping['other'] def _remove_padded_dim(key, strategy): op_data = op_data_mapping[key] sharding_spec = strategy.get_sharding_spec_by_name(op_data.name) tensor_shape = list(sharding_spec.entire_shape) dim_partition_list = [None] * len(tensor_shape) # padded dim is a negative number as the padded dim must be a matrix dim padded_dim = self.padded_dim_mapping[key] # compute the new dim partition for tensor_dim, mesh_dims in sharding_spec.dim_partition_dict.items(): dim_partition_list[tensor_dim] = mesh_dims dim_partition_list.pop(padded_dim) unpadded_dim_partition_list = {k: v for k, v in enumerate(dim_partition_list) if v is not None} # compute unpadded tensor shape tensor_shape.pop(padded_dim) assert tensor_shape == list(op_data.data.shape), f'{tensor_shape} vs {list(op_data.data.shape)}' # update sharding spec sharding_spec.__init__(sharding_spec.device_mesh, tensor_shape, unpadded_dim_partition_list) # enumerate all sharding strategies strategies = [] try: strategy_copy = strategy.clone() # only one of input and other will be padded if 'input' in self.padded_dim_mapping: _remove_padded_dim('input', strategy_copy) _remove_padded_dim('output', strategy_copy) elif 'other' in self.padded_dim_mapping: _remove_padded_dim('other', strategy_copy) _remove_padded_dim('output', strategy_copy) strategies.append(strategy_copy) except ShardingSpecException as e: pass return strategies class Broadcaster(BmmTransform): """ Broadcast the non-matrix dimensions for batched matrix multiplication. """ def __init__(self) -> None: self.broadcast_dim_info = {} def apply(self, shape_mapping: Dict[str, List[int]]): mapping_copy = shape_mapping.copy() # get shapes input_shape = mapping_copy['input'] other_shape = mapping_copy['other'] # sanity check assert len(input_shape) > 1 and len(other_shape) > 1 # broadcast the batch dim and record bcast_non_matrix_dims = get_broadcast_shape(input_shape[:-2], other_shape[:-2]) # store the broadcast dim info input_broadcast_dim_info = get_broadcast_dim_info(bcast_non_matrix_dims, input_shape[:-2]) other_broadcast_dim_info = get_broadcast_dim_info(bcast_non_matrix_dims, other_shape[:-2]) self.broadcast_dim_info['input'] = input_broadcast_dim_info self.broadcast_dim_info['other'] = other_broadcast_dim_info # create the full logical shape input_shape = bcast_non_matrix_dims + input_shape[-2:] other_shape = bcast_non_matrix_dims + other_shape[-2:] assert len(input_shape) == len(other_shape) mapping_copy['input'] = input_shape mapping_copy['other'] = other_shape return mapping_copy def recover(self, op_data_mapping: Dict[str, OperationData], strategy: ShardingStrategy): # remove sharding on the broadcast dim def _remove_sharding_on_broadcast_dim(key, strategy): op_data = op_data_mapping[key] sharding_spec = strategy.get_sharding_spec_by_name(op_data.name) tensor_shape = list(sharding_spec.entire_shape) for dim_idx, broadcast_type in self.broadcast_dim_info[key].items(): if broadcast_type == BroadcastType.MULTIPLE: # if the dim is originally 1 and multiplied during broadcast # we set its sharding to R # e.g. [1, 2, 4] x [4, 4, 8] -> [4, 2, 8] # the dim 0 of [1, 2, 4] is multiplied to 4 tensor_shape[dim_idx] = 1 elif broadcast_type == BroadcastType.PADDDING: # if the dim is padded # we remove its sharding tensor_shape[dim_idx] = None tensor_shape_before_broadcast = [dim for dim in tensor_shape if dim is not None] physical_sharding_spec, removed_dims = recover_sharding_spec_for_broadcast_shape( logical_sharding_spec=sharding_spec, logical_shape=sharding_spec.entire_shape, physical_shape=tensor_shape_before_broadcast) strategy.sharding_specs[op_data] = physical_sharding_spec # enumerate all sharding strategies strategies = [] try: strategy_copy = strategy.clone() _remove_sharding_on_broadcast_dim('input', strategy_copy) _remove_sharding_on_broadcast_dim('other', strategy_copy) strategies.append(strategy_copy) except ShardingSpecException as e: pass return strategies class Viewer(BmmTransform): """ Change the shape of the tensor from N-D to 3D """ def __init__(self) -> None: self.batch_dims_before_view = None def apply(self, shape_mapping: Dict[str, List[int]]): mapping_copy = shape_mapping.copy() self.batch_dims_before_view = list(mapping_copy['input'][:-2]) # get shapes input_shape = shape_mapping['input'] other_shape = shape_mapping['other'] # view to 3d tensor assert len(input_shape) >= 3 and len(other_shape) >= 3 input_shape = [reduce(operator.mul, input_shape[:-2])] + input_shape[-2:] other_shape = [reduce(operator.mul, other_shape[:-2])] + other_shape[-2:] output_shape = input_shape[:2] + other_shape[2:] mapping_copy['input'] = input_shape mapping_copy['other'] = other_shape mapping_copy['output'] = output_shape return mapping_copy def recover(self, op_data_mapping: Dict[str, OperationData], strategy: ShardingStrategy): # get operation data def _update_sharding_spec(key, strategy, physical_batch_dim): """ Map the logical batch dim to the physical batch dim """ op_data = op_data_mapping[key] sharding_spec = strategy.get_sharding_spec_by_name(op_data.name) dim_partition_dict = sharding_spec.dim_partition_dict entire_shape = sharding_spec.entire_shape # upddate the dimension index for the matrix dimensions if 2 in dim_partition_dict: dim_partition_dict[len(self.batch_dims_before_view) + 1] = dim_partition_dict.pop(2) if 1 in dim_partition_dict: dim_partition_dict[len(self.batch_dims_before_view)] = dim_partition_dict.pop(1) # map the logical batch dim to phyiscal batch dim if 0 in dim_partition_dict: batch_dim_shard = dim_partition_dict.pop(0) dim_partition_dict[physical_batch_dim] = batch_dim_shard # the new shape will be the batch dims + the last 2 matrix dims shape_before_view = self.batch_dims_before_view + list(entire_shape[-2:]) sharding_spec.__init__(sharding_spec.device_mesh, shape_before_view, dim_partition_dict) num_batch_dim_before_view = len(self.batch_dims_before_view) # enumerate all sharding strategies strategies = [] for i in range(num_batch_dim_before_view): # create a new strategy strategy_copy = strategy.clone() try: _update_sharding_spec('input', strategy_copy, i) _update_sharding_spec('other', strategy_copy, i) _update_sharding_spec('output', strategy_copy, i) strategies.append(strategy_copy) except ShardingSpecException as e: continue return strategies def _get_bmm_logical_shape(input_shape, other_shape, transforms): """ Compute the logical shapes for BMM operation. BMM has a general representation [b, i, k] = [b, i, j] x [b, j, k] The dimension b is called non-matrix (batch) dimension and the remaining dimensions are called matrix dimensions The logical shape for the bmm operands will undergo three stages 1. append/prepend the 1 to the 1D tensor if there is any 2. broadcast the non-matrix dimensions 3. reshape to 3 dimensions """ shape_mapping = {'input': input_shape, 'other': other_shape} for transform in transforms: shape_mapping = transform.apply(shape_mapping) input_shape = shape_mapping.get('input', None) other_shape = shape_mapping.get('other', None) output_shape = shape_mapping.get('output', None) return input_shape, other_shape, output_shape @operator_registry.register(torch.matmul) @operator_registry.register(torch.Tensor.matmul) class MatMulHandler(MetaInfoNodeHandler): """ The MatMulHandler is a node handler which handles the sharding strategy generation for the matmul operation. According to https://pytorch.org/docs/stable/generated/torch.matmul.html, the operations will vary depending on the operands. """ def __init__(self, *args, **kwargs) -> None: super().__init__(*args, **kwargs) # check which type of operation this matmul will call self.input_meta_data = self.node.args[0]._meta_data self.other_meta_data = self.node.args[1]._meta_data self.output_meta_data = self.node._meta_data input_dim = self.input_meta_data.dim() other_dim = self.other_meta_data.dim() self.matmul_type = get_matmul_type(input_dim, other_dim) if self.matmul_type == MatMulType.BMM: # bmm operation can possibly involve padding, broadcasting and view # these transforms will be used to create logical shape and # recover physical sharding spec self.transforms = [Padder(), Broadcaster(), Viewer()] else: self.transforms = None def get_strategy_generator(self) -> List[StrategyGenerator]: generators = [] op_data_mapping = self.get_operation_data_mapping() if self.matmul_type == MatMulType.BMM: generators.append(BatchedMatMulStrategyGenerator(op_data_mapping, self.device_mesh)) elif self.matmul_type == MatMulType.DOT: generators.append(DotProductStrategyGenerator(op_data_mapping, self.device_mesh)) elif self.matmul_type == MatMulType.MV: generators.append(MatVecStrategyGenerator(op_data_mapping, self.device_mesh)) elif self.matmul_type == MatMulType.MM: generators.append( LinearProjectionStrategyGenerator(op_data_mapping, self.device_mesh, linear_projection_type='linear')) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: logical_shape_func = { MatMulType.DOT: self._get_logical_shape_for_dot, MatMulType.MM: self._get_logical_shape_for_mm, MatMulType.MV: self._get_logical_shape_for_mv, MatMulType.BMM: self._get_logical_shape_for_bmm } logical_shapes = logical_shape_func[self.matmul_type]() op_data_mapping = self._get_op_data_mapping(*logical_shapes) return op_data_mapping def _get_op_data_mapping(self, input_logical_shape, other_logical_shape, output_logical_shape): # convert list to torch.Size if input_logical_shape: input_logical_shape = torch.Size(input_logical_shape) if other_logical_shape: other_logical_shape = torch.Size(other_logical_shape) if output_logical_shape: output_logical_shape = torch.Size(output_logical_shape) # create op data input_op_data = OperationData(name=str(self.node.args[0]), type=OperationDataType.ARG, data=self.input_meta_data, logical_shape=input_logical_shape) other_op_data = OperationData(name=str(self.node.args[1]), type=OperationDataType.ARG, data=self.other_meta_data, logical_shape=other_logical_shape) output_op_data = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=self.output_meta_data, logical_shape=output_logical_shape) mapping = {'input': input_op_data, 'other': other_op_data, 'output': output_op_data} return mapping def _get_logical_shape_for_dot(self): """ The operands for the dot operation have the same logical shape as the physical shape """ return None, None, None def _get_logical_shape_for_mm(self): """ We need to handle the input tensor for a matrix-matrix multiplcation as the input tensor can be a 1D or 2D tensor. If it is a 1D tensor, 1 will be prepended to its shape (e.g. [4] -> [1, 4]). """ if self.input_meta_data.dim() == 1: input_logical_shape = [1] + list(self.input_meta_data.shape) input_logical_shape = torch.Size(input_logical_shape) else: input_logical_shape = None return input_logical_shape, None, None def _get_logical_shape_for_mv(self): """ No broadcasting or dim insertion occurs for matrix-vector operation. """ return None, None, None def _get_logical_shape_for_bmm(self): input_physical_shape = list(self.input_meta_data.shape) other_physical_shape = list(self.other_meta_data.shape) return _get_bmm_logical_shape(input_physical_shape, other_physical_shape, self.transforms) def post_process(self, strategy: ShardingStrategy) -> Union[ShardingStrategy, List[ShardingStrategy]]: if self.matmul_type in [MatMulType.DOT, MatMulType.MV]: return strategy elif self.matmul_type == MatMulType.MM: if self.input_meta_data.dim() == 1: # if a 1 is prepended to the input shape (this occurs when input is a 1D tensor) # we need to remove that dim input_sharding_spec = strategy.get_sharding_spec_by_name(str(self.node.args[0])) input_physical_shape = self.node.args[0]._meta_data.shape dim_partition_dict = input_sharding_spec.dim_partition_dict # remove the partitioning in the dim 0 if 0 in dim_partition_dict: dim_partition_dict.pop(0, None) # move the partitioning in dim 1 to dim 0 if -1 in dim_partition_dict: shard = dim_partition_dict.pop(-1) dim_partition_dict[0] = shard if 1 in dim_partition_dict: shard = dim_partition_dict.pop(1) dim_partition_dict[0] = shard # re-init the sharding spec input_sharding_spec.__init__(input_sharding_spec.device_mesh, entire_shape=input_physical_shape, dim_partition_dict=dim_partition_dict) return strategy else: return strategy elif self.matmul_type == MatMulType.BMM: op_data_mapping = self.get_operation_data_mapping() strategies = [strategy] # recover the physical sharding spec for transform in self.transforms[::-1]: recovered_stragies = [] for strategy_ in strategies: output = transform.recover(op_data_mapping, strategy_) if isinstance(output, ShardingStrategy): recovered_stragies.append(output) elif isinstance(output, (list, tuple)): recovered_stragies.extend(output) else: raise TypeError( f"Found unexpected output type {type(output)} from the recover method of BmmTransform") strategies = recovered_stragies for index, strategies in enumerate(strategies): strategies.name = f"{strategies.name}_{index}" return strategies
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import StrategyGenerator from .strategy.tensor_constructor_generator import TensorConstructorGenerator __all__ = ['TensorConstructorHandler'] @operator_registry.register(torch.arange) class TensorConstructorHandler(NodeHandler): """ A TensorConstructorHandler which deals with the sharding strategies for tensor constructor operations, such as torch.arange. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(TensorConstructorGenerator(op_data_mapping, self.device_mesh)) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: output_data = self.node._meta_data physical_output_operand = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data) mapping = {"output": physical_output_operand} return mapping
from typing import Dict, List import torch from ..sharding_strategy import OperationData, OperationDataType from .node_handler import NodeHandler from .registry import operator_registry from .strategy import StrategyGenerator, SumGenerator __all__ = ['SumHandler'] @operator_registry.register(torch.Tensor.sum) @operator_registry.register(torch.sum) class SumHandler(NodeHandler): """ A SumHandler which deals with the sharding strategies for torch.sum or torch.Tensor.sum. """ def get_strategy_generator(self) -> List[StrategyGenerator]: op_data_mapping = self.get_operation_data_mapping() generators = [] generators.append(SumGenerator(op_data_mapping, self.device_mesh, self.node.args[0])) return generators def get_operation_data_mapping(self) -> Dict[str, OperationData]: # check if the input operand is a parameter if isinstance(self.node.args[0]._meta_data, torch.nn.parameter.Parameter): data_type = OperationDataType.PARAM else: data_type = OperationDataType.ARG input_data = self.node.args[0]._meta_data physical_input_operand = OperationData(name=str(self.node.args[0]), type=data_type, data=input_data) if len(self.node.args) > 1: sum_dims = self.node.args[1] else: sum_dims = tuple(range(self.node.args[0]._meta_data.dim())) if isinstance(sum_dims, int): sum_dims = (sum_dims,) # recover negative value to positive num_dims = self.node.args[0]._meta_data.dim() for i in range(len(sum_dims)): if sum_dims[i] < 0: sum_dims[i] += num_dims # mapping the input dims to output dims # For examples: # input: torch.rand(2, 3, 4, 5) # output: torch.sum(input, (0, 2)) # sum_mapping_dict = {1: 0, 3: 1} # sum_mapping_dict[1] = 0 means the 0th dim of output is the 1st dim of input # sum_mapping_dict[3] = 1 means the 1st dim of output is the 3rd dim of input sum_mapping_dict = {} if 'keepdim' in self.node.kwargs and self.node.kwargs['keepdim']: for i in range(num_dims): sum_mapping_dict.update({i: i}) else: output_index = 0 for i in range(num_dims): if i not in sum_dims: sum_mapping_dict.update({i: output_index}) output_index += 1 assert output_index == self.node._meta_data.dim() sum_info = (sum_dims, sum_mapping_dict) physical_shape_operand = OperationData(name='sum_info', type=OperationDataType.ARG, data=sum_info) output_data = self.node._meta_data physical_output_operand = OperationData(name=str(self.node), type=OperationDataType.OUTPUT, data=output_data) mapping = { "input": physical_input_operand, "sum_info": physical_shape_operand, "output": physical_output_operand } return mapping
import copy from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import (MemoryCost, ShardingStrategy, TrainCycleItem) from .strategy_generator import FollowingStrategyGenerator __all__ = ['UnaryElementwiseGenerator'] class UnaryElementwiseGenerator(FollowingStrategyGenerator): """ UnaryElementwiseGenerator which deals with the sharding strategies of UnaryElementwiseOp. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): compute_cost = TrainCycleItem(fwd=10, bwd=10, total=20) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'output': self._compute_size_in_bytes(strategy, "output") } backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] # For element-wise function, we keep the sharding spec of output node same as # the input. Therefore, the different strategies of input node with same # output sharding spec will generate same strategy for element-wise function. for index, strategy in enumerate(self.predecessor_node.strategies_vector): dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] dim_partition_dict_for_input = input_sharding_spec.dim_partition_dict dim_partition_dict_for_output = copy.deepcopy(dim_partition_dict_for_input) dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. name = f'{sharding_spec_mapping["input"].sharding_sequence} -> {sharding_spec_mapping["output"].sharding_sequence}_{index}' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list
import copy import operator import warnings from functools import reduce from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ignore_sharding_exception from colossalai.tensor.shape_consistency import CollectiveCommPattern from .strategy_generator import StrategyGenerator class ConvStrategyGenerator(StrategyGenerator): """ ConvStrategyGenerator is a generic class to generate strategies. The operation data is defined as `output = input x other + bias`. """ def validate(self) -> bool: ''' In sanity check, we need make sure the input data having correct dimension size. For Conv1d, the dim of input data should be 3([N, C, L]). For Conv2d, the dim of input data should be 4([N, C, H, W]). For Conv3d, the dim of input data should be 5([N, C, H, W, D]). ''' input_op_data = self.op_data['input'] assert input_op_data.data.dim() in ( 3, 4, 5), f'We suppose the dim of input fed into conv op should in range of [3, 5].' def update_compute_cost(self, strategy: ShardingStrategy): ''' Compute the computation cost per device with this specific strategy. Note: compute_cost need to be devided by TFLOPS, now it just shows the computation size. ''' # TODO: compute_cost need to be devided by TFLOPS, now it just shows the computation size. # 1D: (L) * N * Cout * Cin * kernel # 2D: (H * W) * N * Cout * Cin * kernel # 3D: (H * W * D) * N * Cout * Cin * kernel sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() sharded_other_shape = strategy.sharding_specs[self.op_data['other']].get_sharded_shape_per_device() sharded_output_shape = strategy.sharding_specs[self.op_data['output']].get_sharded_shape_per_device() if self.has_bias: # bias add is an element wise operation, so the cost is equal to product of output shape. bias_compute_cost = reduce(operator.mul, sharded_output_shape) output_size = sharded_output_shape[2:] output_size_product = reduce(operator.mul, output_size) input_size = sharded_input_shape[2:] input_size_product = reduce(operator.mul, input_size, 1) kernel_size = sharded_other_shape[2:] kernel_size_product = reduce(operator.mul, kernel_size, 1) batch_size = sharded_input_shape[0] channel_in = sharded_input_shape[1] channel_out = sharded_other_shape[1] forward_compute_cost = output_size_product * batch_size * channel_in * channel_out * kernel_size_product backward_activation_cost = input_size_product * batch_size * channel_in * channel_out * kernel_size_product backward_weight_cost = output_size_product * batch_size * channel_in * channel_out * kernel_size_product backward_compute_cost = backward_weight_cost + backward_activation_cost if self.has_bias: forward_compute_cost += bias_compute_cost backward_compute_cost += bias_compute_cost total_compute_cost = forward_compute_cost + backward_compute_cost compute_cost = TrainCycleItem(fwd=forward_compute_cost, bwd=backward_compute_cost, total=total_compute_cost) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'other': self._compute_size_in_bytes(strategy, "other"), 'output': self._compute_size_in_bytes(strategy, "output") } if self.has_bias: bias_size = self._compute_size_in_bytes(strategy, "bias") forward_size_mapping['bias'] = bias_size backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + other + bias + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad + other_grad + bias_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost @ignore_sharding_exception def split_input_batch_weight_out_channel(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}S{mesh_dim_1} = S{mesh_dim_0}R x RS{mesh_dim_1}' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0] }, "other": { 1: [mesh_dim_1] }, "output": { 0: [mesh_dim_0], 1: [mesh_dim_1] }, } if self.has_bias: dim_partition_dict_mapping["bias"] = {0: [mesh_dim_1]} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping = {"input": input_comm_action} if self.is_param("other"): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping["other"] = other_comm_action if self.has_bias: if self.is_param('bias'): bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, key_for_kwarg='bias') communication_action_mapping["bias"] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_batch(self, mesh_dim_0): name = f'S{mesh_dim_0}R = S{mesh_dim_0}R x RR' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0] }, "other": {}, "output": { 0: [mesh_dim_0], }, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} if self.is_param("other"): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping["other"] = other_comm_action if self.has_bias: if self.is_param('bias'): bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, key_for_kwarg='bias') communication_action_mapping["bias"] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_both_dim_weight_in_channel(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}R = S{mesh_dim_0}S{mesh_dim_1} x S{mesh_dim_1}R' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0], 1: [mesh_dim_1], }, "other": { 0: [mesh_dim_1] }, "output": { 0: [mesh_dim_0], }, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action output_comm_action = self.get_communication_action( sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.AFTER) communication_action_mapping = {"output": output_comm_action} if self.is_param("other"): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping["other"] = other_comm_action if self.has_bias: if self.is_param("bias"): bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, key_for_kwarg='bias') communication_action_mapping["bias"] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_in_channel_weight_both_channel(self, mesh_dim_0, mesh_dim_1): name = f'RS{mesh_dim_1} = RS{mesh_dim_0} x S{mesh_dim_0}S{mesh_dim_1}' dim_partition_dict_mapping = { "input": { 1: [mesh_dim_0], }, "other": { 0: [mesh_dim_0], 1: [mesh_dim_1], }, "output": { 1: [mesh_dim_1], }, } if self.has_bias: dim_partition_dict_mapping["bias"] = { 0: [mesh_dim_1], } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action output_comm_action = self.get_communication_action( sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.AFTER) input_comm_action = self.get_communication_action( sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping = {"output": output_comm_action, "input": input_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_in_channel_weight_in_channel(self, mesh_dim_0): name = f'RR = RS{mesh_dim_0} x S{mesh_dim_0}R' dim_partition_dict_mapping = { "input": { 1: [mesh_dim_0], }, "other": { 0: [mesh_dim_0], }, "output": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action output_comm_action = self.get_communication_action( sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.AFTER) communication_action_mapping = {"output": output_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_weight_out_channel(self, mesh_dim_0): name = f'RS{mesh_dim_0} = RR x RS{mesh_dim_0}' dim_partition_dict_mapping = { "input": {}, "other": { 1: [mesh_dim_0], }, "output": { 1: [mesh_dim_0], }, } if self.has_bias: dim_partition_dict_mapping["bias"] = { 0: [mesh_dim_0], } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action input_comm_action = self.get_communication_action( sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping = {"input": input_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def non_split(self): name = f'RR = RR x RR' dim_partition_dict_mapping = { "input": {}, "other": {}, "output": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping={}) @ignore_sharding_exception def split_1d_parallel_on_input_batch(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}{mesh_dim_1}R = S{mesh_dim_0}{mesh_dim_1}R x RR' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0, mesh_dim_1], }, "other": {}, "output": { 0: [mesh_dim_0, mesh_dim_1], }, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} if self.is_param("other"): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping["other"] = other_comm_action if self.has_bias: if self.is_param("bias"): bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, key_for_kwarg='bias') communication_action_mapping["bias"] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_1d_parallel_on_in_channel(self, mesh_dim_0, mesh_dim_1): name = f'RR = RS{mesh_dim_0}{mesh_dim_1} x S{mesh_dim_0}{mesh_dim_1}R' dim_partition_dict_mapping = { "input": { 1: [mesh_dim_0, mesh_dim_1], }, "other": { 0: [mesh_dim_0, mesh_dim_1], }, "output": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action output_comm_action = self.get_communication_action( sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.AFTER) communication_action_mapping = {"output": output_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_1d_parallel_on_out_channel(self, mesh_dim_0, mesh_dim_1): name = f'RS{mesh_dim_0}{mesh_dim_1} = RR x RS{mesh_dim_0}{mesh_dim_1}' dim_partition_dict_mapping = { "input": {}, "other": { 1: [mesh_dim_0, mesh_dim_1], }, "output": { 1: [mesh_dim_0, mesh_dim_1], }, } if self.has_bias: dim_partition_dict_mapping["bias"] = { 0: [mesh_dim_0, mesh_dim_1], } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action input_comm_action = self.get_communication_action( sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping = {"input": input_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def collate_strategies(self) -> List[ShardingStrategy]: strategies = [] # SS = SR x RS strategies.append(self.split_input_batch_weight_out_channel(0, 1)) strategies.append(self.split_input_batch_weight_out_channel(1, 0)) # SR = SR x RR strategies.append(self.split_input_batch(0)) strategies.append(self.split_input_batch(1)) # SR = SS x SR strategies.append(self.split_input_both_dim_weight_in_channel(0, 1)) strategies.append(self.split_input_both_dim_weight_in_channel(1, 0)) # RS = RS x SS strategies.append(self.split_input_in_channel_weight_both_channel(0, 1)) strategies.append(self.split_input_in_channel_weight_both_channel(1, 0)) # RR = RS x SR strategies.append(self.split_input_in_channel_weight_in_channel(0)) strategies.append(self.split_input_in_channel_weight_in_channel(1)) # RS = RR x RS strategies.append(self.split_weight_out_channel(0)) strategies.append(self.split_weight_out_channel(1)) # RR= RR x RR strategies.append(self.non_split()) # S01R = S01R x RR strategies.append(self.split_1d_parallel_on_input_batch(0, 1)) # RR = RS01 x S01R strategies.append(self.split_1d_parallel_on_in_channel(0, 1)) # RS01 = RR x RS01 strategies.append(self.split_1d_parallel_on_out_channel(0, 1)) return strategies
import copy from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import MemoryCost, ShardingStrategy, TrainCycleItem from colossalai.auto_parallel.tensor_shard.utils import ( enumerate_all_possible_1d_sharding, enumerate_all_possible_2d_sharding, ignore_sharding_exception, ) from .strategy_generator import StrategyGenerator __all__ = ['WhereGenerator'] class WhereGenerator(StrategyGenerator): """ WhereGenerator is a generic class to generate strategies for Where operation. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): compute_cost = TrainCycleItem(fwd=10, bwd=10, total=20) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = { 'condition': self._compute_size_in_bytes(strategy, "condition"), 'x': self._compute_size_in_bytes(strategy, "x"), 'y': self._compute_size_in_bytes(strategy, "y"), 'output': self._compute_size_in_bytes(strategy, "output") } backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = condition + x + y + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items()]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=0) # compute bwd cost incurred # bwd_cost = condition_grad + x_grad + y_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items()]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=0) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=0) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost @ignore_sharding_exception def _generate_strategy_with_dim_partition(self, dim_partition): dim_partition_dict_mapping = { "condition": dim_partition, "x": dim_partition, "y": dim_partition, "output": dim_partition } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) name = f'{sharding_spec_mapping["output"].sharding_sequence} = {sharding_spec_mapping["condition"].sharding_sequence} x {sharding_spec_mapping["x"].sharding_sequence} x {sharding_spec_mapping["y"].sharding_sequence}' communication_action_mapping = {} strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) return strategy def enumerate_all_possible_output_spec(self, mesh_dim_0, mesh_dim_1, dimension_length): dim_partition_list = [] dim_partition_list.extend(enumerate_all_possible_1d_sharding(mesh_dim_0, dimension_length)) dim_partition_list.extend(enumerate_all_possible_1d_sharding(mesh_dim_1, dimension_length)) dim_partition_list.extend(enumerate_all_possible_2d_sharding(mesh_dim_0, mesh_dim_1, dimension_length)) # append {} for non_split case dim_partition_list.append({}) return dim_partition_list def collate_strategies(self) -> List[ShardingStrategy]: ''' Generate every possible strategies for a where node, and record all strategies into the strategies_vector. ''' strategy_list = [] dimension_length = len(self.op_data["output"].logical_shape) dim_partition_list = self.enumerate_all_possible_output_spec(0, 1, dimension_length) for dim_partition in dim_partition_list: strategy = self._generate_strategy_with_dim_partition(dim_partition) strategy_list.append(strategy) return strategy_list
from typing import Dict, List from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( MemoryCost, OperationData, ShardingStrategy, TrainCycleItem, ) from colossalai.device.device_mesh import DeviceMesh from .strategy_generator import StrategyGenerator __all__ = ['PlaceholderGenerator'] class PlaceholderGenerator(StrategyGenerator): """ PlaceholderGenerator is a generic class to generate strategies for placeholder node. """ def __init__(self, operation_data_mapping: Dict[str, OperationData], device_mesh: DeviceMesh, placeholder_option: str): super().__init__(operation_data_mapping, device_mesh) self.placeholder_option = placeholder_option def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): compute_cost = TrainCycleItem(fwd=10, bwd=10, total=20) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = {'output': self._compute_size_in_bytes(strategy, "output")} # compute fwd cost incurred # fwd_cost = output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items()]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=0) bwd_mem_cost = MemoryCost(activation=0, parameter=0) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=0) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost def replica_placeholder(self) -> ShardingStrategy: """ Generate replica strategy for placeholder node. """ dim_partition_dict_mapping = { "output": {}, } communication_action_mapping = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) name = 'Replica Placeholder' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) return strategy def distributed_placeholder(self, mesh_list) -> ShardingStrategy: """ Generate distributed strategy for placeholder node. """ dim_partition_dict_mapping = { "output": { 0: mesh_list }, } communication_action_mapping = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) name = 'Distributed Placeholder' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) return strategy def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] if self.placeholder_option == 'distributed': mesh_list = [0, 1] distributed_strategy = self.distributed_placeholder(mesh_list) strategy_list.append(distributed_strategy) else: assert self.placeholder_option == 'replicated', f'placeholder_option {self.placeholder_option} is not supported' replicated_strategy = self.replica_placeholder() strategy_list.append(replicated_strategy) return strategy_list
import copy import operator from functools import reduce from typing import List from colossalai.auto_parallel.tensor_shard.node_handler.strategy.strategy_generator import FollowingStrategyGenerator from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ( check_keep_sharding_status, detect_reshape_mapping, infer_output_dim_partition_dict, ) from colossalai.tensor.shape_consistency import CollectiveCommPattern from colossalai.tensor.sharding_spec import ShardingSpec __all__ = ['SumGenerator'] class SumGenerator(FollowingStrategyGenerator): """ SumGenerator deals with the sharding strategies of torch.sum op. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() sharded_output_shape = strategy.sharding_specs[self.op_data['output']].get_sharded_shape_per_device() input_size_product = reduce(operator.mul, sharded_input_shape) output_size_product = reduce(operator.mul, sharded_output_shape) compute_cost = TrainCycleItem(fwd=input_size_product, bwd=output_size_product, total=input_size_product + output_size_product) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'output': self._compute_size_in_bytes(strategy, "output") } backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] for index, strategy in enumerate(self.predecessor_node.strategies_vector): dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] dim_partition_dict_for_input = copy.deepcopy(input_sharding_spec.dim_partition_dict) sum_dims, sum_mapping_dict = self.op_data['sum_info'].data # TODO: a better way to handle the distributed sum is sum all the data on chip and then do all reduce # among all the shard groups recover_dims = [] dim_partition_dict_for_output = {} for dim in dim_partition_dict_for_input: if dim in sum_dims: recover_dims.append(dim) elif dim in sum_mapping_dict: dim_partition_dict_for_output[sum_mapping_dict[dim]] = dim_partition_dict_for_input[dim] else: raise RuntimeError(f'dim {dim} is not in sum_mapping_dict or sum_dims') for dim in recover_dims: dim_partition_dict_for_input.pop(dim) dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. name = f'{sharding_spec_mapping["input"].sharding_sequence} -> {sharding_spec_mapping["output"].sharding_sequence}_{index}' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list
from .batch_norm_generator import BatchNormStrategyGenerator from .binary_elementwise_generator import BinaryElementwiseStrategyGenerator from .conv_strategy_generator import ConvStrategyGenerator from .embedding_generator import EmbeddingStrategyGenerator from .getattr_generator import GetattrGenerator from .getitem_generator import GetItemStrategyGenerator, TensorStrategyGenerator, TensorTupleStrategyGenerator from .layer_norm_generator import LayerNormGenerator from .matmul_strategy_generator import ( BatchedMatMulStrategyGenerator, DotProductStrategyGenerator, LinearProjectionStrategyGenerator, MatVecStrategyGenerator, ) from .normal_pooling_generator import NormalPoolStrategyGenerator from .output_generator import OutputGenerator from .placeholder_generator import PlaceholderGenerator from .reshape_generator import ( DefaultReshapeGenerator, PermuteGenerator, SplitGenerator, TransposeGenerator, ViewGenerator, ) from .softmax_generator import SoftmaxGenerator from .strategy_generator import StrategyGenerator from .sum_generator import SumGenerator from .tensor_constructor_generator import TensorConstructorGenerator from .unary_elementwise_generator import UnaryElementwiseGenerator from .where_generator import WhereGenerator __all__ = [ 'StrategyGenerator', 'DotProductStrategyGenerator', 'MatVecStrategyGenerator', 'LinearProjectionStrategyGenerator', 'BatchedMatMulStrategyGenerator', 'ConvStrategyGenerator', 'UnaryElementwiseGenerator', 'BatchNormStrategyGenerator', 'GetItemStrategyGenerator', 'TensorStrategyGenerator', 'TensorTupleStrategyGenerator', 'LayerNormGenerator', 'PlaceholderGenerator', 'OutputGenerator', 'WhereGenerator', 'NormalPoolStrategyGenerator', 'BinaryElementwiseStrategyGenerator', 'GetattrGenerator', 'TensorConstructorGenerator', 'EmbeddingStrategyGenerator', 'SumGenerator', 'SoftmaxGenerator', 'ViewGenerator', 'PermuteGenerator', 'TransposeGenerator', 'SplitGenerator', 'DefaultReshapeGenerator' ]
import operator from abc import ABC, abstractmethod from functools import reduce from typing import Any, Dict, List, Union import torch from torch.fx import Node from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, OperationData, OperationDataType, ShardingStrategy, TrainCycleItem, ) from colossalai.device.device_mesh import DeviceMesh from colossalai.tensor.shape_consistency import CollectiveCommPattern, CommSpec, ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec from colossalai.tensor.utils import convert_dim_partition_dict class StrategyGenerator(ABC): """ StrategyGenerator is used to generate the same group of sharding strategies. TODO: remove the original strategy_generator.py after refactoring """ def __init__(self, operation_data_mapping: Dict[str, OperationData], device_mesh: DeviceMesh): self.op_data = operation_data_mapping self.device_mesh = device_mesh # validate the whether operation data is of desired value self.validate() @property def has_bias(self): """ A utility method to check for the existence of bias operand for convenience. """ return 'bias' in self.op_data def is_param(self, op_data_name): other_data = self.op_data[op_data_name] return other_data.type == OperationDataType.PARAM def is_buffer(self, op_data_name): other_data = self.op_data[op_data_name] return other_data.type == OperationDataType.BUFFER def get_sharding_strategy(self, name: str, sharding_spec_mapping: Dict[str, ShardingSpec], communication_action_mapping: Dict[str, CommSpec]): """ A factory method to produce a ShardingStrategy object. Args: sharding_spec_mapping (Dict[str, ShardingSpec]): the mapping between the operation data name and the ShardingSpec object. communication_action_mapping (Dict[str, CommSpec]): the mapping between the operation data name and the CommSpec object. """ sharding_specs = self.replace_op_name_with_op_data(sharding_spec_mapping) communication_actions = self.replace_op_name_with_op_data(communication_action_mapping) return ShardingStrategy(name=name, sharding_specs=sharding_specs, communication_actions=communication_actions) def to_sharding_spec_mapping(self, mapping: Dict[str, Dict[int, List[int]]]): """ A utility method to convert the the dim partition dict to a ShardingSpec object. Args: mapping (Dict[str, Dict[int, List[int]]]): the key of the mapping is the operation data name and the value is a dim partition dictionary. Notes: The op_data.data is commonly type of torch.Tensor, torch.nn.Parameter, so the sharding spec is easy to create from the shape of the data. However, if the op_data.data is of other non-iterative types, such as float or int, we should return None. If the op_data.data is of some iterative types, such as list or tuple, we should return a list of ShardingSpec objects follow the same rule as above mentioned. """ results = {} for op_data_name, dim_partition_dict in mapping.items(): if op_data_name in self.op_data: op_data = self.op_data[op_data_name] def _to_sharding_spec( data: any, logical_shape: any, dim_partition_dict: Dict[int, List[int]]) -> Union[ShardingSpec, List[ShardingSpec], None]: """ This is a recursive function to convert the dim partition dict to a ShardingSpec object. """ if isinstance(data, torch.Tensor): dim_size = len(logical_shape) dim_partition_dict = convert_dim_partition_dict(dim_size, dim_partition_dict) sharding_spec = ShardingSpec(device_mesh=self.device_mesh, entire_shape=logical_shape, dim_partition_dict=dim_partition_dict) return sharding_spec elif isinstance(data, (list, tuple)): sharding_spec = [] for data_element, logical_shape_element, dim_partition_dict_element in zip( data, logical_shape, dim_partition_dict): sharding_spec.append( _to_sharding_spec(data_element, logical_shape_element, dim_partition_dict_element)) return sharding_spec else: return None sharding_spec = _to_sharding_spec(op_data.data, op_data.logical_shape, dim_partition_dict) results[op_data_name] = sharding_spec return results def replace_op_name_with_op_data(self, mapping: Dict[str, Any]): """ Convert the key of the dictionary from the operation data name to an OperationData object. """ results = {} for k, v in mapping.items(): op_data = self.op_data[k] results[op_data] = v return results def get_communication_spec(self, sharding_spec: ShardingSpec, communication_pattern: CollectiveCommPattern, logical_process_axis: Union[int, List[int]]): """ A factory method to produce a CommSpec object. """ return CommSpec(comm_pattern=communication_pattern, sharding_spec=sharding_spec, logical_process_axis=logical_process_axis) def get_communication_action(self, sharding_spec: ShardingSpec, communication_pattern: CollectiveCommPattern, logical_process_axis: Union[int, List[int]], comm_type: CommType, arg_index: int = -1, key_for_kwarg: any = None) -> CommAction: """ A factory method to produce a CommAction object. """ return CommAction(comm_spec=self.get_communication_spec(sharding_spec=sharding_spec, communication_pattern=communication_pattern, logical_process_axis=logical_process_axis), comm_type=comm_type, arg_index=arg_index, key_for_kwarg=key_for_kwarg) def update_communication_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: """ Compute the communication cost involved in the forward and backward iteration. """ comm_cost = TrainCycleItem(fwd=0, bwd=0, total=0) def _compute_and_add(op_data: OperationData, comm_spec: CommSpec): num_ele_in_comm = comm_spec.get_comm_cost() dtype = op_data.data.dtype size_per_elem_bytes = torch.tensor([], dtype=dtype).element_size() for phase, cost in num_ele_in_comm.items(): num_ele_in_comm[phase] = num_ele_in_comm[phase] * size_per_elem_bytes comm_cost.fwd += num_ele_in_comm['forward'] comm_cost.bwd += num_ele_in_comm['backward'] comm_cost.total += num_ele_in_comm['total'] # check if communication action exists # if so, loop over each action and compute the cost of each action if strategy.communication_actions is not None: for operand, comm_action in strategy.communication_actions.items(): if isinstance(comm_action, CommAction): comm_spec = comm_action.comm_spec else: # this condition branch will be removed after all the handler updated. comm_spec = comm_action if isinstance(comm_spec, dict): src_spec = comm_spec['src_spec'] tgt_spec = comm_spec['tgt_spec'] shape_consistency_manager = ShapeConsistencyManager() _, comm_action_sequence, _ = shape_consistency_manager.shape_consistency(src_spec, tgt_spec) for comm_spec_ in comm_action_sequence: _compute_and_add(operand, comm_spec_) else: _compute_and_add(operand, comm_spec) # update the communication cost attribute in-place strategy.communication_cost = comm_cost return strategy @abstractmethod def update_compute_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: """ Customize this method to compute the computation flops. """ pass @abstractmethod def update_memory_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: """ Customize this method to compute the memory cost in bytes. """ pass def _compute_size_in_bytes(self, strategy: ShardingStrategy, key: str): """ Compute the size of a tensor in bytes. Args: strategy (ShardingStrategy): the ShardingStrategy generated. key (str): the name of the operation data defined by the generator. """ op_data = self.op_data[key] def _compute_size_in_bytes_helper(sharding_spec, meta_data): sharded_shape = sharding_spec.get_sharded_shape_per_device() if len(sharded_shape) == 0: num_elements = 1 else: num_elements = reduce(operator.mul, sharded_shape) dtype = getattr(meta_data, 'dtype') size_per_elem_bytes = torch.tensor([], dtype=dtype).element_size() return num_elements * size_per_elem_bytes if isinstance(op_data.data, tuple): assert isinstance(strategy.sharding_specs[op_data], list), \ 'sharding_spec of op_data should be a list of sharding specs if op_data.data is a tuple.' total_bytes = 0 for index, sharding_spec in enumerate(strategy.sharding_specs[op_data]): meta_data = op_data.data[index] if isinstance(meta_data, torch.Tensor): element_bytes = _compute_size_in_bytes_helper(sharding_spec, meta_data) else: # if meta_data is not a tensor, we count the memroy as 0 element_bytes = 0 total_bytes += element_bytes else: if isinstance(op_data.data, torch.Tensor): total_bytes = _compute_size_in_bytes_helper(strategy.sharding_specs[op_data], op_data.data) else: # if op_data.data is not a tensor, we count the memroy as 0 total_bytes = 0 return total_bytes def generate(self) -> List[ShardingStrategy]: """ Generate all possible sharding strategies for this operation. """ strategies = self.collate_strategies() # some strategies may be None as ignore_sharding_exception may return None # when ShardingSpecException occurs. # thus, remove those None values strategies = [strategy for strategy in strategies if strategy] # update the costs # update mete info on cost # these update methods are all in-place, the default method will do nothing # the cost info will only be added if the child class overrides these methods for strategy in strategies: self.update_communication_cost(strategy) self.update_compute_cost(strategy) self.update_memory_cost(strategy) return strategies @abstractmethod def collate_strategies(self) -> List[ShardingStrategy]: pass @abstractmethod def validate(self) -> bool: """ Validate if the operands are of desired shape. If True, means this generator can be used for the current operation. """ pass class FollowingStrategyGenerator(StrategyGenerator): """ FollowingStrategyGenerator is used to generate the sharding strategies which depends on its predecessor node. TODO: remove the original strategy_generator.py after refactoring """ def __init__(self, operation_data_mapping: Dict[str, OperationData], device_mesh: DeviceMesh, predecessor_node: Node): self.op_data = operation_data_mapping self.device_mesh = device_mesh self.predecessor_node = predecessor_node class OutputStrategyGenerator(StrategyGenerator): """ OutputStrategyGenerator is used to generate the sharding strategies for Output Node. """ def __init__(self, operation_data_mapping: Dict[str, OperationData], device_mesh: DeviceMesh, predecessor_nodes: List[Node]): super().__init__(operation_data_mapping, device_mesh) self.predecessor_nodes = predecessor_nodes
import copy import operator from functools import reduce from typing import List from colossalai.auto_parallel.tensor_shard.node_handler.strategy.strategy_generator import FollowingStrategyGenerator from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ( check_keep_sharding_status, detect_reshape_mapping, infer_output_dim_partition_dict, ) from colossalai.tensor.shape_consistency import CollectiveCommPattern __all__ = ['SoftmaxGenerator'] class SoftmaxGenerator(FollowingStrategyGenerator): """ SoftmaxGenerator is used to generate strategies for torch.nn.Softmax or F.softmax. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): ''' Compute the computation cost per device with this specific strategy. ''' sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() sharded_output_shape = strategy.sharding_specs[self.op_data['output']].get_sharded_shape_per_device() input_size_product = reduce(operator.mul, sharded_input_shape) output_size_product = reduce(operator.mul, sharded_output_shape) forward_compute_cost = output_size_product * 2 backward_compute_cost = input_size_product total_compute_cost = forward_compute_cost + backward_compute_cost compute_cost = TrainCycleItem(fwd=forward_compute_cost, bwd=backward_compute_cost, total=total_compute_cost) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'output': self._compute_size_in_bytes(strategy, "output") } backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] for index, strategy in enumerate(self.predecessor_node.strategies_vector): dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] dim_partition_dict_for_input = copy.deepcopy(input_sharding_spec.dim_partition_dict) softmax_dim = self.op_data['softmax_dim'].data if softmax_dim in dim_partition_dict_for_input: recover_dims = dim_partition_dict_for_input.pop(softmax_dim) dim_partition_dict_for_output = copy.deepcopy(dim_partition_dict_for_input) dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. name = f'{sharding_spec_mapping["input"].sharding_sequence} -> {sharding_spec_mapping["output"].sharding_sequence}_{index}' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list
from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import MemoryCost, ShardingStrategy, TrainCycleItem from colossalai.auto_parallel.tensor_shard.utils import ( enumerate_all_possible_1d_sharding, enumerate_all_possible_2d_sharding, ignore_sharding_exception, ) from colossalai.tensor.sharding_spec import ShardingSpecException from .strategy_generator import StrategyGenerator __all__ = ['GetattrGenerator'] class GetattrGenerator(StrategyGenerator): """ PlaceholderGenerator is a generic class to generate strategies for placeholder node. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): compute_cost = TrainCycleItem(fwd=10, bwd=10, total=20) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = {'output': self._compute_size_in_bytes(strategy, "output")} # compute fwd cost incurred # fwd_cost = output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items()]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=0) bwd_mem_cost = MemoryCost(activation=0, parameter=0) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=0) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost @ignore_sharding_exception def enumerate_all_possible_output(self, mesh_dim_0, mesh_dim_1): # we check for the output logical shape to get the number of dimensions dim_partition_list = [] dim_size = len(self.op_data['output'].logical_shape) # enumerate all the 2D sharding cases sharding_list_2d = enumerate_all_possible_2d_sharding(mesh_dim_0, mesh_dim_1, dim_size) dim_partition_list.extend(sharding_list_2d) # enumerate all the 1D sharding cases sharding_list_1d_on_dim_0 = enumerate_all_possible_1d_sharding(mesh_dim_0, dim_size) dim_partition_list.extend(sharding_list_1d_on_dim_0) sharding_list_1d_on_dim_1 = enumerate_all_possible_1d_sharding(mesh_dim_1, dim_size) dim_partition_list.extend(sharding_list_1d_on_dim_1) # add empty dict for fully replicated case dim_partition_list.append({}) # sharding strategy bookkeeping strategy_list = [] # convert these dim partition dict to sharding strategy for dim_partition_dict in dim_partition_list: dim_partition_dict_mapping = dict(output=dim_partition_dict) try: sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} # get name name = f"get_attr {sharding_spec_mapping['output'].sharding_sequence}" sharding_strategy = self.get_sharding_strategy( name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(sharding_strategy) except ShardingSpecException: continue return strategy_list def collate_strategies(self) -> List[ShardingStrategy]: return self.enumerate_all_possible_output(0, 1)
import copy import operator from functools import reduce from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ignore_sharding_exception from colossalai.tensor.shape_consistency import CollectiveCommPattern from .strategy_generator import StrategyGenerator __all__ = ['BatchNormStrategyGenerator'] class BatchNormStrategyGenerator(StrategyGenerator): """ A StrategyGenerator which deals with the sharding strategies of batch normalization. To keep the math consistency, there are two way to do BatchNorm if the input shards on batch dimension: 1. We gather the input partitions through batch dimension, then do the normal BatchNorm. 2. We do the SyncBatchNorm on the each input partition seperately, the SyncBN op will help us to keep the computing correctness. In this generator, both methods will be considered. """ def validate(self) -> bool: ''' In sanity check, we need make sure the input data having correct dimension size. For BatchNorm1d, the dim of input data should be 3([N, C, L]). For BatchNorm2d, the dim of input data should be 4([N, C, H, W]). For BatchNorm3d, the dim of input data should be 5([N, C, H, W, D]). ''' input_op_data = self.op_data['input'] assert input_op_data.data.dim() in ( 3, 4, 5), f'We suppose the dim of input fed into conv op should in range of [3, 5].' def update_compute_cost(self, strategy: ShardingStrategy): ''' Compute the computation cost per device with this specific strategy. Note: compute_cost need to be devided by TFLOPS, now it just shows the computation size. ''' # TODO: a constant coefficient need to be added. # 1D: (L) * N * Cin # 2D: (H * W) * N * Cin # 3D: (H * W * D) * N * Cin sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() sharded_output_shape = strategy.sharding_specs[self.op_data['output']].get_sharded_shape_per_device() if self.has_bias: # bias add is an element wise operation, so the cost is equal to product of output shape. bias_compute_cost = reduce(operator.mul, sharded_output_shape) input_product = reduce(operator.mul, sharded_input_shape, 1) forward_compute_cost = input_product backward_activation_compute_cost = input_product backward_weight_compute_cost = input_product backward_compute_cost = backward_weight_compute_cost + backward_activation_compute_cost if self.has_bias: forward_compute_cost += bias_compute_cost backward_compute_cost += bias_compute_cost total_compute_cost = forward_compute_cost + backward_compute_cost compute_cost = TrainCycleItem(fwd=forward_compute_cost, bwd=backward_compute_cost, total=total_compute_cost) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'other': self._compute_size_in_bytes(strategy, "other"), 'output': self._compute_size_in_bytes(strategy, "output"), 'running_mean': self._compute_size_in_bytes(strategy, "running_mean"), 'running_var': self._compute_size_in_bytes(strategy, "running_var"), } if self.has_bias: bias_size = self._compute_size_in_bytes(strategy, "bias") forward_size_mapping['bias'] = bias_size backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + other + bias + output fwd_activation_cost = sum( [v for k, v in forward_size_mapping.items() if not self.is_param(k) and not self.is_buffer(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_buffer_cost = sum([v for k, v in forward_size_mapping.items() if self.is_buffer(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost, buffer=fwd_buffer_cost) # compute bwd cost incurred # bwd_cost = input_grad + other_grad + bias_grad bwd_activation_cost = sum( [v for k, v in backward_size_mapping.items() if not self.is_param(k) and not self.is_buffer(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost, buffer=fwd_buffer_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost @ignore_sharding_exception def split_input_channel(self, mesh_dim_0): name = f'RS{mesh_dim_0} = RS{mesh_dim_0} x S{mesh_dim_0}' dim_partition_dict_mapping = { "input": { 1: [mesh_dim_0] }, "other": { 0: [mesh_dim_0] }, "output": { 1: [mesh_dim_0] }, "running_mean": { 0: [mesh_dim_0] }, "running_var": { 0: [mesh_dim_0] }, "num_batches_tracked": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {0: [mesh_dim_0]} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_channel_1d(self, mesh_dim_0, mesh_dim_1): name = f'RS{mesh_dim_0}{mesh_dim_1} = RS{mesh_dim_0}{mesh_dim_1} x S{mesh_dim_0}{mesh_dim_1}' dim_partition_dict_mapping = { "input": { 1: [mesh_dim_0, mesh_dim_1] }, "other": { 0: [mesh_dim_0, mesh_dim_1] }, "output": { 1: [mesh_dim_0, mesh_dim_1] }, "running_mean": { 0: [mesh_dim_0, mesh_dim_1] }, "running_var": { 0: [mesh_dim_0, mesh_dim_1] }, "num_batches_tracked": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {0: [mesh_dim_0, mesh_dim_1]} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def non_split(self): name = f'RR = RR x R' dim_partition_dict_mapping = { "input": {}, "other": {}, "output": {}, "running_mean": {}, "running_var": {}, "num_batches_tracked": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_batch(self, mesh_dim_0): name = f'S{mesh_dim_0}R = S{mesh_dim_0}R x R WITH SYNC_BN' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0] }, "other": {}, "output": { 0: [mesh_dim_0] }, "running_mean": {}, "running_var": {}, "num_batches_tracked": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action # For SyncBN case, we don't need to do communication for weight and bias. # TODO: the communication happens interally at SyncBN operation. We need to replace the BN operation # to SyncBN operation instead of inserting a communication node. output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.IMPLICIT) # TODO: Temporary solution has no communication cost, # above action should be added after the SyncBN replace pass completed. communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_batch_1d(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}{mesh_dim_1}R = S{mesh_dim_0}{mesh_dim_1}R x R WITH SYNC_BN' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0, mesh_dim_1] }, "other": {}, "output": { 0: [mesh_dim_0, mesh_dim_1] }, "running_mean": {}, "running_var": {}, "num_batches_tracked": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action # For SyncBN case, we don't need to do communication for gradients of weight and bias. # TODO: the communication happens interally at SyncBN operation. We need to replace the BN operation # to SyncBN operation instead of inserting a communication node. output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.IMPLICIT) # TODO: Temporary solution has no communication cost, # above action should be added after the SyncBN replace pass completed. communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_both_dim(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}S{mesh_dim_1} = S{mesh_dim_0}S{mesh_dim_1} x S{mesh_dim_1} WITH SYNC_BN' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0], 1: [mesh_dim_1], }, "other": { 0: [mesh_dim_1], }, "output": { 0: [mesh_dim_0], 1: [mesh_dim_1], }, "running_mean": { 0: [mesh_dim_1], }, "running_var": { 0: [mesh_dim_1], }, "num_batches_tracked": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = { 0: [mesh_dim_1], } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action # For SyncBN case, we don't need to do communication for gradients of weight and bias. # TODO: the communication happens interally at SyncBN operation. We need to replace the BN operation # to SyncBN operation instead of inserting a communication node. output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=[mesh_dim_0], comm_type=CommType.IMPLICIT) # TODO: Temporary solution has no communication cost, # above action should be added after the SyncBN replace pass completed. communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def collate_strategies(self) -> List[ShardingStrategy]: ''' Generate every possible strategies for a BatchNorm node, and record all strategies into the strategies_vector. ''' strategy_list = [] # RS = RS x S strategy_list.append(self.split_input_channel(0)) strategy_list.append(self.split_input_channel(1)) # RR = RR x R strategy_list.append(self.non_split()) # RS01 = RS01 x S01 strategy_list.append(self.split_input_channel_1d(0, 1)) # The strategies with SYNC_BN are temporarily commented, # because it requires some additional passes to keep runtime # computation correctness. # TODO: The strategies below should be uncommented after runtime # passes ready. # SR = SR x R WITH SYNC_BN strategy_list.append(self.split_input_batch(0)) strategy_list.append(self.split_input_batch(1)) # SS = SS x S WITH SYNC_BN strategy_list.append(self.split_input_both_dim(0, 1)) strategy_list.append(self.split_input_both_dim(1, 0)) # S01R = S01R x R WITH SYNC_BN strategy_list.append(self.split_input_batch_1d(0, 1)) return strategy_list
import copy from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.logging import get_dist_logger from colossalai.tensor.shape_consistency import CollectiveCommPattern from colossalai.tensor.sharding_spec import ShardingSpecException from .strategy_generator import FollowingStrategyGenerator __all__ = ['GetItemStrategyGenerator', 'TensorStrategyGenerator', 'TensorTupleStrategyGenerator'] class GetItemStrategyGenerator(FollowingStrategyGenerator): """ GetItemStrategyGenerator is a generic class to generate strategies for operator.getitem. The operation data is defined as `output = input[other]`. There are mainly three use cases: 1. args_0._meta_data: torch.Tensor, args_1._meta_data: int 2. args_0._meta_data: torch.Tensor, args_1._meta_data: slice 3. args_0._meta_data: Tuple[torch.Tensor], args_1._meta_data: int """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): compute_cost = TrainCycleItem(fwd=10, bwd=10, total=20) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'output': self._compute_size_in_bytes(strategy, "output") } backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost class TensorStrategyGenerator(GetItemStrategyGenerator): ''' Deal with case 1 and 2. ''' def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] getitem_index = self.op_data['index'].data for index, strategy in enumerate(self.predecessor_node.strategies_vector): try: logger = get_dist_logger() dim_partition_dict_mapping = {} communication_action_mapping = {} dim_partition_dict_for_input = copy.deepcopy( strategy.output_sharding_specs[self.op_data["input"]].dim_partition_dict) int_index = False if isinstance(getitem_index, int): int_index = True getitem_dims = [ 0, ] shift_length = 1 elif isinstance(getitem_index, slice): getitem_dims = [ 0, ] else: getitem_dims = [i for i in range(len(getitem_index))] if isinstance(getitem_index[0], int): int_index = True shift_length = len(getitem_index) gather_dims = [] for dim in getitem_dims: if dim in dim_partition_dict_for_input: gather_dims.append(dim) for dim in gather_dims: dim_partition_dict_for_input.pop(dim) dim_partition_dict_for_output = copy.deepcopy(dim_partition_dict_for_input) if int_index: shift_dim_partition_dict_for_output = {} for dim, mesh_dim_list in dim_partition_dict_for_output.items(): shift_dim_partition_dict_for_output[dim - shift_length] = mesh_dim_list dim_partition_dict_for_output = shift_dim_partition_dict_for_output dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) name = f'{sharding_spec_mapping["output"].sharding_sequence} = {sharding_spec_mapping["input"].sharding_sequence}_{index}' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) except ShardingSpecException as e: logger.debug(e) continue strategy_list.append(strategy) for strategy in strategy_list: self.update_communication_cost(strategy) self.update_compute_cost(strategy) self.update_memory_cost(strategy) return strategy_list class TensorTupleStrategyGenerator(GetItemStrategyGenerator): ''' Deal with case 3. ''' def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] index = self.op_data["index"].data for strategy_index, strategy in enumerate(self.predecessor_node.strategies_vector): # the sharding spec for input in this case is a tuple of ShardingSpec. sharding_spec_for_input = strategy.output_sharding_specs[self.op_data["input"]] dim_partition_dict_for_output = sharding_spec_for_input[index].dim_partition_dict dim_partition_dict_mapping = {} communication_action_mapping = {} dim_partition_dict_mapping = { "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) sharding_spec_mapping["input"] = sharding_spec_for_input input_sharding_info = f"get the {index} element from (" for sharding_spec in sharding_spec_for_input: input_sharding_info += f'{sharding_spec.sharding_sequence}, ' input_sharding_info += ")" name = f'{sharding_spec_mapping["output"].sharding_sequence} = {input_sharding_info}_{strategy_index}' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list
import copy from typing import List from colossalai.auto_parallel.tensor_shard.node_handler.strategy.strategy_generator import FollowingStrategyGenerator from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ( check_keep_sharding_status, detect_reshape_mapping, infer_output_dim_partition_dict, ) from colossalai.tensor.shape_consistency import CollectiveCommPattern from colossalai.tensor.sharding_spec import ShardingSpec __all__ = ['ReshapeGenerator', 'ViewGenerator', 'PermuteGenerator', 'TransposeGenerator', 'SplitGenerator'] class ReshapeGenerator(FollowingStrategyGenerator): """ ReshapeGenerator is the base class for all the reshape operation. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): compute_cost = TrainCycleItem(fwd=10, bwd=10, total=20) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'output': self._compute_size_in_bytes(strategy, "output") } backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost def collate_strategies(self) -> List[ShardingStrategy]: return super().collate_strategies() class ViewGenerator(ReshapeGenerator): """ ViewGenerator deals with the sharding strategies of view op. """ def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] for index, strategy in enumerate(self.predecessor_node.strategies_vector): dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] origin_shape = self.op_data['input'].data.shape tgt_shape = self.op_data['tgt_shape'].data reshape_mapping_dict = detect_reshape_mapping(origin_shape, tgt_shape) dim_partition_dict_for_input = input_sharding_spec.dim_partition_dict keep_sharding_status = check_keep_sharding_status(dim_partition_dict_for_input, reshape_mapping_dict) if keep_sharding_status: dim_partition_dict_for_output = infer_output_dim_partition_dict(dim_partition_dict_for_input, reshape_mapping_dict) else: dim_partition_dict_for_output = {} dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. if keep_sharding_status: name = f'{sharding_spec_mapping["input"].sharding_sequence} -> {sharding_spec_mapping["output"].sharding_sequence}_{index}' else: name = f'{sharding_spec_mapping["input"].sharding_sequence} -> FULLY REPLICATED_{index}' # add comm action for converting input to fully replicated total_mesh_dim_list = [] for mesh_dim_list in dim_partition_dict_for_input.values(): total_mesh_dim_list.extend(mesh_dim_list) # if there is only one sharding dimension, we should use the value instead of list as logical_process_axis. if len(total_mesh_dim_list) == 1: total_mesh_dim_list = total_mesh_dim_list[0] # the total mesh dim list only has one element, so the shard dim has only one element as well. shard_dim = list(dim_partition_dict_for_input.keys())[0] input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.GATHER_FWD_SPLIT_BWD, logical_process_axis=total_mesh_dim_list, comm_type=CommType.BEFORE, arg_index=0) # it will gather the input through gather_dim during forward phase. input_comm_action.comm_spec.gather_dim = shard_dim # it will split the input activation grad through shard_dim during backward phase. input_comm_action.comm_spec.shard_dim = shard_dim elif len(total_mesh_dim_list) >= 2: source_spec = sharding_spec_mapping["input"] target_spec = ShardingSpec(device_mesh=self.device_mesh, entire_shape=source_spec.entire_shape, dim_partition_dict={}) comm_spec = {'src_spec': source_spec, 'tgt_spec': target_spec} input_comm_action = CommAction(comm_spec=comm_spec, comm_type=CommType.BEFORE, arg_index=0) else: input_comm_action = None if input_comm_action is not None: communication_action_mapping["input"] = input_comm_action strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list class PermuteGenerator(ReshapeGenerator): """ PermuteGenerator deals with the sharding strategies of permute op. """ def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] for index, strategy in enumerate(self.predecessor_node.strategies_vector): dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] permute_dims = self.op_data['permute_dims'].data dim_partition_dict_for_input = input_sharding_spec.dim_partition_dict dim_partition_dict_for_output = {} for dim_index, permute_dim in enumerate(permute_dims): if permute_dim in dim_partition_dict_for_input: dim_partition_dict_for_output[dim_index] = dim_partition_dict_for_input[permute_dim] dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. name = f'{sharding_spec_mapping["input"].sharding_sequence} -> {sharding_spec_mapping["output"].sharding_sequence}_{index}' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list class TransposeGenerator(ReshapeGenerator): """ TransposeGenerator deals with the sharding strategies of permute op. """ def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] for index, strategy in enumerate(self.predecessor_node.strategies_vector): dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] dim_partition_dict_for_input = input_sharding_spec.dim_partition_dict dim_partition_dict_for_output = {} transpose_dims = self.op_data['transpose_dims'].data dim_0 = transpose_dims[0] dim_1 = transpose_dims[1] for dim, sharded_dims in dim_partition_dict_for_input.items(): if dim == dim_0: dim_partition_dict_for_output[dim_1] = dim_partition_dict_for_input[dim_0] elif dim == dim_1: dim_partition_dict_for_output[dim_0] = dim_partition_dict_for_input[dim_1] else: dim_partition_dict_for_output[dim] = sharded_dims dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. name = f'{sharding_spec_mapping["input"].sharding_sequence} -> {sharding_spec_mapping["output"].sharding_sequence}_{index}' strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list class SplitGenerator(ReshapeGenerator): """ SplitGenerator deals with the sharding strategies of split op. """ def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] for index, strategy in enumerate(self.predecessor_node.strategies_vector): recover_dims = None dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] dim_partition_dict_for_input = copy.deepcopy(input_sharding_spec.dim_partition_dict) split_size, split_dim = self.op_data['split_info'].data if split_dim in dim_partition_dict_for_input: recover_dims = dim_partition_dict_for_input.pop(split_dim) dim_partition_dict_for_output = [ copy.deepcopy(dim_partition_dict_for_input) for _ in range(len(self.op_data["output"].data)) ] assert len(dim_partition_dict_for_output) >= 2 dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. name = f'{sharding_spec_mapping["input"].sharding_sequence}_{index}' # add comm action if the input need to be recovered to replica in the split dimension. if recover_dims: # if there is only one sharding dimension, we should use the value instead of list as logical_process_axis. if len(recover_dims) == 1: recover_dims = recover_dims[0] input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.GATHER_FWD_SPLIT_BWD, logical_process_axis=recover_dims, comm_type=CommType.BEFORE, arg_index=0) # it will gather the input through gather_dim during forward phase. input_comm_action.comm_spec.gather_dim = split_dim # it will split the input activation grad through split_dim during backward phase. input_comm_action.comm_spec.shard_dim = split_dim elif len(recover_dims) >= 2: # original sharding spec source_spec = input_sharding_spec # target sharding spec target_spec = sharding_spec_mapping["input"] comm_spec = {'src_spec': source_spec, 'tgt_spec': target_spec} input_comm_action = CommAction(comm_spec=comm_spec, comm_type=CommType.BEFORE, arg_index=0) else: input_comm_action = None if input_comm_action is not None: communication_action_mapping["input"] = input_comm_action strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list class DefaultReshapeGenerator(ReshapeGenerator): """ DefaultReshapeGenerator which deals with the sharding strategies of Reshape Op which have to recover the tensor to Replica status. """ def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] # For default reshape strategy, to keep the computing correctness we keep the # sharding spec of input is fully replicated. In addition, we will keep the output # in replica status and let the successor node choose the way to resharding the # output node. Therefore, the different strategies of input node with same # output sharding spec will generate same strategy for reshape function. for index, strategy in enumerate(self.predecessor_node.strategies_vector): dim_partition_dict_mapping = {} communication_action_mapping = {} input_sharding_spec = strategy.output_sharding_specs[self.op_data["input"]] dim_partition_dict_for_input = input_sharding_spec.dim_partition_dict dim_partition_dict_for_output = {} if isinstance(self.op_data["output"].data, tuple): dim_partition_dict_for_output = [{} for _ in range(len(self.op_data["output"].data))] dim_partition_dict_mapping = { "input": dim_partition_dict_for_input, "output": dim_partition_dict_for_output, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # add index into name to pass the duplicated check # we keep same strategies with different name for node merging, and it will not increase the searching space, # because in solver, this node will be merged into other nodes, and solver will not create a new variable for this node. name = f'{sharding_spec_mapping["input"].sharding_sequence} -> FULLY REPLICATED_{index}' total_mesh_dim_list = [] for mesh_dim_list in dim_partition_dict_for_input.values(): total_mesh_dim_list.extend(mesh_dim_list) # if there is only one sharding dimension, we should use the value instead of list as logical_process_axis. if len(total_mesh_dim_list) == 1: total_mesh_dim_list = total_mesh_dim_list[0] input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.GATHER_FWD_SPLIT_BWD, logical_process_axis=total_mesh_dim_list, comm_type=CommType.BEFORE, arg_index=0) input_comm_action.comm_spec.gather_dim = total_mesh_dim_list elif len(total_mesh_dim_list) >= 2: source_spec = sharding_spec_mapping["input"] target_spec = ShardingSpec(device_mesh=self.device_mesh, entire_shape=source_spec.entire_shape, dim_partition_dict={}) comm_spec = {'src_spec': source_spec, 'tgt_spec': target_spec} input_comm_action = CommAction(comm_spec=comm_spec, comm_type=CommType.BEFORE, arg_index=0) else: input_comm_action = None if input_comm_action is not None: communication_action_mapping["input"] = input_comm_action strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) strategy_list.append(strategy) return strategy_list
import operator from ast import arg from functools import reduce from typing import List from colossalai.auto_parallel.tensor_shard.options import SolverPerference from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ignore_sharding_exception from colossalai.tensor.shape_consistency import CollectiveCommPattern from .strategy_generator import StrategyGenerator class MatMulStrategyGenerator(StrategyGenerator): """ MatMulStrategyGenerator is a generic class to cover all matrix multiplication cases. The operation data is defined as `output = input x other + bias`. """ def update_memory_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'other': self._compute_size_in_bytes(strategy, "other"), 'output': self._compute_size_in_bytes(strategy, "output") } if self.has_bias: bias_size = self._compute_size_in_bytes(strategy, "bias") size_mapping['bias'] = bias_size # compute fwd cost incurred # fwd_cost = input + other + bias + output fwd_activation_cost = sum([v for k, v in size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad + bias_grad bwd_activation_cost = sum([v for k, v in size_mapping.items() if k in ['input', 'other', 'bias']]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=0) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + 0) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost class DotProductStrategyGenerator(MatMulStrategyGenerator): def validate(self) -> bool: input_op_data = self.op_data['input'] other_op_data = self.op_data['other'] assert input_op_data.data.dim() == 1 and other_op_data.data.dim() == 1 def update_compute_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() fwd_compute_cost = sharded_input_shape[0] bwd_compute_cost = fwd_compute_cost * 2 compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) return compute_cost @ignore_sharding_exception def no_split(self): name = f'R = R dot R' dim_partition_dict = {"input": {}, "other": {}, "output": {}, 'bias': {}} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_one_dim(self, mesh_dim): name = f'R = S{mesh_dim} dot S{mesh_dim}' # get sharding spec dim_partition_dict = {"input": {0: [mesh_dim]}, "other": {0: [mesh_dim]}, "output": {}, "bias": {0: [mesh_dim]}} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) # get communication action output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['output'], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim, comm_type=CommType.AFTER) communication_action_mapping = {"output": output_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] # do not split dimensions for dot product # R = R dot R strategy_list.append(self.no_split()) # split two tensors in the same dimensions # S = S dot S strategy_list.append(self.split_one_dim(0)) strategy_list.append(self.split_one_dim(1)) return strategy_list class MatVecStrategyGenerator(MatMulStrategyGenerator): def validate(self) -> bool: input_op_data = self.op_data['input'] other_op_data = self.op_data['other'] assert input_op_data.data.dim() == 2 and other_op_data.data.dim() == 1 def update_compute_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() fwd_compute_cost = sharded_input_shape[0] bwd_compute_cost = fwd_compute_cost * 2 compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) return compute_cost @ignore_sharding_exception def no_split(self): name = "R = R x R" dim_partition_dict = {"input": {}, "other": {}, "output": {}} if self.has_bias: dim_partition_dict['bias'] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping={}) @ignore_sharding_exception def split_input_batch(self, mesh_dim): name = f'S{mesh_dim}R = S{mesh_dim}R x R' # get sharding spec dim_partition_dict = { "input": { 0: [mesh_dim] }, "other": {}, "output": { 0: [mesh_dim] }, } if self.has_bias: dim_partition_dict['bias'] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) # get communication action communication_action_mapping = {} if self.is_param('other'): other_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['other'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['other'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping['other'] = other_comm_action if self.has_bias: if self.is_param('bias'): bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim, comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim, comm_type=CommType.BEFORE, arg_index=2) communication_action_mapping['bias'] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] # no split strategy_list.append(self.no_split()) # split the batch dim for the first tensor only strategy_list.append(self.split_input_batch(0)) strategy_list.append(self.split_input_batch(1)) return strategy_list class LinearProjectionStrategyGenerator(MatMulStrategyGenerator): def __init__(self, operation_data_mapping, device_mesh, linear_projection_type='linear', solver_perference=SolverPerference.STANDARD): super().__init__(operation_data_mapping, device_mesh) self.linear_projection_type = linear_projection_type self.solver_perference = solver_perference def update_compute_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: # C = AB # C: [M, N], A: [M, P], B: [P, N] # fwd cost = MNP (only count mul) # bwd: 2 x fwd_cost sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() sharded_other_shape = strategy.sharding_specs[self.op_data['other']].get_sharded_shape_per_device() dim_m_val = reduce(operator.mul, sharded_input_shape[:-1]) dim_n_val = sharded_other_shape[-1] dim_p_val = sharded_other_shape[0] fwd_compute_cost = dim_m_val * dim_n_val * dim_p_val bwd_compute_cost = fwd_compute_cost * 2 compute_cost = TrainCycleItem(fwd=bwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) strategy.compute_cost = compute_cost def dp_strategies(self) -> List[ShardingStrategy]: strategies = [] # S01R = S01R x RR strategies.append(self.split_lhs_1st_dim_1d(0, 1)) return strategies def tp_strategies(self) -> List[ShardingStrategy]: strategies = [] # RR = RS01 x S01R strategies.append(self.split_lhs_2nd_dim_1d(0, 1)) # RS01 = RR x RS01 strategies.append(self.split_rhs_2nd_dim_1d(0, 1)) # RS = RS x SS strategies.append(self.split_rhs_space_both_contract(0, 1)) strategies.append(self.split_rhs_space_both_contract(1, 0)) # RR= RS x SR strategies.append(self.recompute_split_both_contract(0)) strategies.append(self.recompute_split_both_contract(1)) # RS = RR x RS strategies.append(self.split_rhs_space_only(0)) strategies.append(self.split_rhs_space_only(1)) return strategies def mix_strategies(self) -> List[ShardingStrategy]: strategies = [] # SS = SR x RS strategies.append(self.split_lhs_space_rhs_space(0, 1)) strategies.append(self.split_lhs_space_rhs_space(1, 0)) # SR = SS x SR strategies.append(self.split_lhs_space_both_contract(0, 1)) strategies.append(self.split_lhs_space_both_contract(1, 0)) # RR = RR x RR strategies.append(self.non_split()) return strategies def collate_strategies(self) -> List[ShardingStrategy]: strategies = [] if self.solver_perference == SolverPerference.STANDARD: strategies.extend(self.dp_strategies()) strategies.extend(self.tp_strategies()) strategies.extend(self.mix_strategies()) elif self.solver_perference == SolverPerference.DP: strategies.extend(self.dp_strategies()) elif self.solver_perference == SolverPerference.TP: strategies.extend(self.tp_strategies()) return strategies @ignore_sharding_exception def split_lhs_space_rhs_space(self, mesh_dim_0, mesh_dim_1): # handle case SS = SR x RS name = f'S{mesh_dim_0}S{mesh_dim_1} = S{mesh_dim_0}R x RS{mesh_dim_1}' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0] }, "other": { -1: [mesh_dim_1] }, "output": { 0: [mesh_dim_0], -1: [mesh_dim_1] }, } # linear bias only has one dimension, but addmm bias has same dimensions # as the output logically. if self.linear_projection_type == 'linear': dim_partition_dict_mapping['bias'] = {-1: [mesh_dim_1]} elif self.linear_projection_type == 'addmm': dim_partition_dict_mapping['bias'] = {0: [mesh_dim_0], -1: [mesh_dim_1]} else: raise ('Unsupported linear projection type') sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action communication_action_mapping = {} input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.BEFORE, arg_index=0) if self.is_param('other'): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping['input'] = input_comm_action communication_action_mapping['other'] = other_comm_action # we only add allreduce comm action for linear bias, because # allreduce comm action for addmm bias will be considered in post processing if self.has_bias and self.linear_projection_type == 'linear': if self.is_param('bias'): bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, key_for_kwarg='bias') communication_action_mapping['bias'] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_lhs_space_both_contract(self, mesh_dim_0, mesh_dim_1): # handle the case SR = SS x SR name = f'S{mesh_dim_0}R = S{mesh_dim_0}S{mesh_dim_1} x S{mesh_dim_1}R' # get sharding spec mapping dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0], -1: [mesh_dim_1] }, "other": { 0: [mesh_dim_1] }, "bias": {}, "output": { 0: [mesh_dim_0] }, } # linear bias only has one dimension, but addmm bias has same dimensions # as the output logically. if self.linear_projection_type == 'linear': dim_partition_dict_mapping['bias'] = {} elif self.linear_projection_type == 'addmm': dim_partition_dict_mapping['bias'] = {0: [mesh_dim_0]} else: raise ('Unsupported linear projection type') sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication action mapping communication_action_mapping = {} output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["output"], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.AFTER) if self.is_param('other'): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping['other'] = other_comm_action communication_action_mapping['output'] = output_comm_action # we only add allreduce comm action for linear bias, because # allreduce comm action for addmm bias will be considered in post processing if self.has_bias and self.linear_projection_type == 'linear': if self.is_param('bias'): bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, key_for_kwarg='bias') communication_action_mapping['bias'] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_rhs_space_both_contract(self, mesh_dim_0, mesh_dim_1): name = f'RS{mesh_dim_1} = RS{mesh_dim_0} x S{mesh_dim_0}S{mesh_dim_1}' # get sharding specs dim_partition_dict_mapping = { "input": { -1: [mesh_dim_0] }, "other": { 0: [mesh_dim_0], -1: [mesh_dim_1] }, "bias": { -1: [mesh_dim_1] }, "output": { -1: [mesh_dim_1] }, } # We don't have to do anything special for bias here, because # the bias is already the same sharding spec as the output. sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication actions communication_action_mapping = {} output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['output'], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.AFTER) input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['input'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping["input"] = input_comm_action communication_action_mapping['output'] = output_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def recompute_split_both_contract(self, mesh_dim): name = f'RR = RS{mesh_dim} x S{mesh_dim}R' # get sharding spec dim_partition_dict_mapping = { "input": { -1: [mesh_dim] }, "other": { 0: [mesh_dim] }, "bias": {}, "output": {}, } # We don't have to do anything special for bias here, because # the bias is already the same sharding spec as the output. sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication action communication_action_mapping = {} output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['output'], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim, comm_type=CommType.AFTER) communication_action_mapping['output'] = output_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_rhs_space_only(self, mesh_dim): name = f'RS{mesh_dim} = RR x RS{mesh_dim}' # get sharding spec dim_partition_dict_mapping = { "input": {}, "other": { -1: [mesh_dim] }, "bias": { -1: [mesh_dim] }, "output": { -1: [mesh_dim] }, } # We don't have to do anything special for bias here, because # the bias is already the same sharding spec as the output. sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication actions communication_action_mapping = {} input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['input'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping['input'] = input_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_lhs_1st_dim_1d(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}{mesh_dim_1}R = S{mesh_dim_0}{mesh_dim_1}R x RR' # get sharding spec dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0, mesh_dim_1] }, "other": {}, "bias": {}, "output": { 0: [mesh_dim_0, mesh_dim_1] }, } # linear bias only has one dimension, but addmm bias has same dimensions # as the output logically. if self.linear_projection_type == 'linear': dim_partition_dict_mapping['bias'] = {} elif self.linear_projection_type == 'addmm': dim_partition_dict_mapping['bias'] = {0: [mesh_dim_0, mesh_dim_1]} else: raise ('Unsupported linear projection type') sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication action communication_action_mapping = {} if self.is_param('other'): other_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['other'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['other'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping['other'] = other_comm_action # we only add allreduce comm action for linear bias, because # allreduce comm action for addmm bias will be considered in post processing if self.has_bias and self.linear_projection_type == 'linear': if self.is_param('bias'): bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.HOOK) else: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, key_for_kwarg='bias') communication_action_mapping['bias'] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_lhs_2nd_dim_1d(self, mesh_dim_0, mesh_dim_1): name = f'RR = RS{mesh_dim_0}{mesh_dim_1} x S{mesh_dim_0}{mesh_dim_1}R' # get sharding spec dim_partition_dict_mapping = { "input": { -1: [mesh_dim_0, mesh_dim_1] }, "other": { 0: [mesh_dim_0, mesh_dim_1] }, "bias": {}, "output": {}, } # We don't have to do anything special for bias here, because # the bias is already the same sharding spec as the output. sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication action communication_action_mapping = {} output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['output'], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.AFTER) communication_action_mapping['output'] = output_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_rhs_2nd_dim_1d(self, mesh_dim_0, mesh_dim_1): name = f'RS{mesh_dim_0}{mesh_dim_1} = RR x RS{mesh_dim_0}{mesh_dim_1}' # get sharding spec dim_partition_dict_mapping = { "input": {}, "other": { -1: [mesh_dim_0, mesh_dim_1] }, "bias": { -1: [mesh_dim_0, mesh_dim_1] }, "output": { -1: [mesh_dim_0, mesh_dim_1] }, } # We don't have to do anything special for bias here, because # the bias is already the same sharding spec as the output. sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication action communication_action_mapping = {} input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['input'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping['input'] = input_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def non_split(self): name = f'RR = RR x RR' # get sharding spec dim_partition_dict_mapping = { "input": {}, "other": {}, "bias": {}, "output": {}, } # We don't have to do anything special for bias here, because # the bias is already the same sharding spec as the output. sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # get communication action communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def validate(self) -> bool: assert "input" in self.op_data assert "other" in self.op_data # make sure the other has 2 dim input_data = self.op_data['input'] other_data = self.op_data['other'] assert input_data.data.dim() > 0 and other_data.data.dim() == 2 assert other_data.logical_shape[0] == input_data.logical_shape[-1] if self.has_bias: bias_data = self.op_data['bias'] assert bias_data.logical_shape[-1] == other_data.logical_shape[-1] class BatchedMatMulStrategyGenerator(MatMulStrategyGenerator): """ Generate sharding strategies for the batched matrix multiplication. A batched matrix multiplication can be viewed as [b, i, k] x [b, k, j] -> [b, i, j] The bias term is considered to have a 2D logical shape. Note: This class will be used to generate strategies for torch.bmm and torch.addbmm. However, the result of torch.addbmm is not correct, some extra runtime apply actions are required to keep numerical correctness. """ # TODO: torch.addbmm correctness issue need to be fixed. def __init__(self, *args, **kwargs): self.squeeze_batch_dim = False super().__init__(*args, **kwargs) def _pop_batch_dim_sharding_for_output(self, dim_partition_dict): # remove partition dict for dim 0 dim_partition_dict['output'].pop(0, None) # decrease the remaining dim index by 1 temp_dim_partition = {} keys = list(dim_partition_dict['output'].keys()) for key in keys: val = dim_partition_dict['output'].pop(key) temp_dim_partition[key - 1] = val dim_partition_dict['output'].update(temp_dim_partition) def validate(self) -> bool: input_op_data = self.op_data['input'] other_op_data = self.op_data['other'] assert len(input_op_data.logical_shape) == 3 or len(other_op_data.logical_shape) == 3 if 'bias' in self.op_data: bias_op_data = self.op_data['bias'] assert bias_op_data.data.dim() < 3 and len(bias_op_data.logical_shape) == 2 if self.op_data['output'].data.dim() == 2: # addbmm will shrink the first batch dim self.squeeze_batch_dim = True def update_compute_cost(self, strategy: ShardingStrategy) -> ShardingStrategy: fwd_compute_cost = self.op_data['input'].data.shape[-1] * reduce(operator.mul, self.op_data['output'].data.shape) bwd_compute_cost = fwd_compute_cost * 2 compute_cost = TrainCycleItem(fwd=fwd_compute_cost, bwd=bwd_compute_cost, total=fwd_compute_cost + bwd_compute_cost) strategy.compute_cost = compute_cost @ignore_sharding_exception def split_one_batch_dim(self, mesh_dim): name = f'Sb{mesh_dim} = Sb{mesh_dim} x Sb{mesh_dim}' # get sharding_spec dim_partition_dict = {"input": {0: [mesh_dim]}, "other": {0: [mesh_dim]}, "bias": {}, "output": {0: [mesh_dim]}} if self.squeeze_batch_dim: self._pop_batch_dim_sharding_for_output(dim_partition_dict) sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) # get communication actions communication_action_mapping = {} if self.has_bias: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping['bias'] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_two_batch_dim(self, mesh_dim_0, mesh_dim_1): name = f'Sb{mesh_dim_0}{mesh_dim_1} = Sb{mesh_dim_0}{mesh_dim_1} x Sb{mesh_dim_0}{mesh_dim_1}' dim_partition_dict = { "input": { 0: [mesh_dim_0, mesh_dim_1] }, "other": { 0: [mesh_dim_0, mesh_dim_1] }, "bias": {}, "output": { 0: [mesh_dim_0, mesh_dim_1] } } if self.squeeze_batch_dim: self._pop_batch_dim_sharding_for_output(dim_partition_dict) sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) # get communication actions communication_action_mapping = {} if self.has_bias: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping['bias'] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_batch_dim_lhs_space(self, mesh_dim_0, mesh_dim_1): name = f'Sb{mesh_dim_0}Si{mesh_dim_1} = Sb{mesh_dim_0}Si{mesh_dim_1} x Sb{mesh_dim_0}' dim_partition_dict = { "input": { 0: [mesh_dim_0], 1: [mesh_dim_1] }, "other": { 0: [mesh_dim_0] }, "bias": { 0: [mesh_dim_1] }, "output": { 0: [mesh_dim_0], 1: [mesh_dim_1] } } if self.squeeze_batch_dim: self._pop_batch_dim_sharding_for_output(dim_partition_dict) sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) # get communication actions communication_action_mapping = {} other_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['other'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping['other'] = other_comm_action if self.has_bias: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping['bias'] = bias_comm_action # for addbmm case, other is the third argument instead of second. communication_action_mapping['other'].arg_index += 1 return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_batch_dim_rhs_space(self, mesh_dim_0, mesh_dim_1): name = f'Sb{mesh_dim_0}Sj{mesh_dim_1} = Sb{mesh_dim_0}R x Sb{mesh_dim_0}Sj{mesh_dim_1}' dim_partition_dict = { "input": { 0: [mesh_dim_0] }, "other": { 0: [mesh_dim_0], 2: [mesh_dim_1] }, "bias": { 1: [mesh_dim_1] }, "output": { 0: [mesh_dim_0], 2: [mesh_dim_1] } } if self.squeeze_batch_dim: self._pop_batch_dim_sharding_for_output(dim_partition_dict) sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) # get communication actions communication_action_mapping = {} input_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['input'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping['input'] = input_comm_action if self.has_bias: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE) communication_action_mapping['bias'] = bias_comm_action # for addbmm case, other is the second argument instead of first. communication_action_mapping['input'].arg_index += 1 return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_batch_dim_both_contract(self, mesh_dim_0, mesh_dim_1): name = f'Sb{mesh_dim_0}R = Sb{mesh_dim_0}Sk{mesh_dim_1} x Sb{mesh_dim_0}Sk{mesh_dim_1}' dim_partition_dict = { "input": { 0: [mesh_dim_0], 2: [mesh_dim_1] }, "other": { 0: [mesh_dim_0], 1: [mesh_dim_1] }, "bias": {}, "output": { 0: [mesh_dim_0], } } if self.squeeze_batch_dim: self._pop_batch_dim_sharding_for_output(dim_partition_dict) sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict) # get communication actions communication_action_mapping = {} output_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['output'], communication_pattern=CollectiveCommPattern.ALLREDUCE_FWD_IDENTITY_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.AFTER) communication_action_mapping['output'] = output_comm_action if self.has_bias: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping['bias'], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping['bias'] = bias_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def collate_strategies(self) -> List[ShardingStrategy]: strategy_list = [] device_mesh_is_1d = True if len(self.device_mesh.mesh_shape) == 2 and 1 not in self.device_mesh.mesh_shape: device_mesh_is_1d = False if device_mesh_is_1d: # split only the batch dimension # Sb = Sb x Sb # can be None as it is only for 1D device mesh # only for 1D device mesh if len(self.device_mesh.mesh_shape) == 1: mesh_dim = 0 else: mesh_dim = self.device_mesh.mesh_shape.index(1) strategy_list.append(self.split_one_batch_dim(mesh_dim)) else: # for 2D device mesh # split batch dim of two inputs and the i dim of the first tensor # SbSi = SbSi x Sb strategy_list.append(self.split_batch_dim_lhs_space(0, 1)) strategy_list.append(self.split_batch_dim_lhs_space(1, 0)) # split batch dim of two inputs and the j of the second tensor # SbSj = Sb x SbSj strategy_list.append(self.split_batch_dim_rhs_space(0, 1)) strategy_list.append(self.split_batch_dim_rhs_space(1, 0)) # split batch dim of two inputs and the k dim of two inputs # Sb = SbSk x SbSk, need to all-reduce by k dim strategy_list.append(self.split_batch_dim_both_contract(0, 1)) strategy_list.append(self.split_batch_dim_both_contract(1, 0)) # split two batch dim strategy_list.append(self.split_two_batch_dim(0, 1)) return strategy_list
import copy import operator from functools import reduce from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ( enumerate_all_possible_1d_sharding, enumerate_all_possible_2d_sharding, ignore_sharding_exception, ) from colossalai.tensor.shape_consistency import CollectiveCommPattern from .strategy_generator import StrategyGenerator __all__ = ['LayerNormGenerator'] class LayerNormGenerator(StrategyGenerator): """ LayerNormGenerator is a generic class to generate strategies for LayerNorm operation. The operation data is defined as `output = input x other + bias`. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): ''' Compute the computation cost per device with this specific strategy. Note: compute_cost need to be devided by TFLOPS, now it just shows the computation size. ''' # TODO: compute_cost need to be devided by TFLOPS, now it just shows the computation size. # TODO: a constant coefficient need to be added. sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() sharded_weight_shape = strategy.sharding_specs[self.op_data['other']].get_sharded_shape_per_device() if self.has_bias: # bias add is an element wise operation, so the cost is equal to product of output shape. bias_compute_cost = reduce(operator.mul, sharded_weight_shape) # in LayerNorm context, batch dimensions mean all the dimensions do not join the normalization. input_batch_shape = sharded_input_shape[:-len(sharded_weight_shape)] input_batch_product = reduce(operator.mul, input_batch_shape, 1) norm_kernel_product = reduce(operator.mul, sharded_weight_shape, 1) forward_compute_cost = input_batch_product * norm_kernel_product backward_activation_compute_cost = input_batch_product * norm_kernel_product # To compute gradient of on norm kernel element requires input_batch_product times computation, so # the total cost is input_batch_product * norm_kernel_product backward_weight_compute_cost = input_batch_product * norm_kernel_product backward_compute_cost = backward_activation_compute_cost + backward_weight_compute_cost if self.has_bias: forward_compute_cost += bias_compute_cost backward_compute_cost += bias_compute_cost total_compute_cost = forward_compute_cost + backward_compute_cost compute_cost = TrainCycleItem(fwd=forward_compute_cost, bwd=backward_compute_cost, total=total_compute_cost) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): ''' Compute the memory cost per device with this specific strategy. ''' forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'other': self._compute_size_in_bytes(strategy, "other"), 'output': self._compute_size_in_bytes(strategy, "output") } if self.has_bias: bias_size = self._compute_size_in_bytes(strategy, "bias") forward_size_mapping['bias'] = bias_size backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + other + bias + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad + other_grad + bias_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost @ignore_sharding_exception def _generate_strategy_with_dim_partition(self, dim_partition): dim_partition_dict_mapping = { "input": dim_partition, "other": {}, "output": dim_partition, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) name = f'{sharding_spec_mapping["output"].sharding_sequence} = {sharding_spec_mapping["input"].sharding_sequence} x {sharding_spec_mapping["other"].sharding_sequence}' total_mesh_dim_list = [] for mesh_dim_list in dim_partition.values(): total_mesh_dim_list.extend(mesh_dim_list) # if there is only one sharding dimension, we should use the value instead of list as logical_process_axis. if len(total_mesh_dim_list) == 1: total_mesh_dim_list = total_mesh_dim_list[0] communication_action_mapping = {} other_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=total_mesh_dim_list, comm_type=CommType.HOOK) communication_action_mapping["other"] = other_comm_action if self.has_bias: bias_comm_action = self.get_communication_action( sharding_spec=sharding_spec_mapping["bias"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=total_mesh_dim_list, comm_type=CommType.HOOK) communication_action_mapping["bias"] = bias_comm_action strategy = self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) return strategy def split_input_batch_single_mesh_dim(self, mesh_dim_0, batch_dimension_length): strategy_list = [] dim_partition_list = enumerate_all_possible_1d_sharding(mesh_dim_0, batch_dimension_length) for dim_partition in dim_partition_list: strategy = self._generate_strategy_with_dim_partition(dim_partition) strategy_list.append(strategy) return strategy_list def split_input_batch_both_mesh_dim(self, mesh_dim_0, mesh_dim_1, batch_dimension_length): strategy_list = [] dim_partition_list = enumerate_all_possible_2d_sharding(mesh_dim_0, mesh_dim_1, batch_dimension_length) for dim_partition in dim_partition_list: strategy = self._generate_strategy_with_dim_partition(dim_partition) strategy_list.append(strategy) return strategy_list @ignore_sharding_exception def non_split(self): name = f'RR = RR x R' dim_partition_dict_mapping = { "input": {}, "other": {}, "output": {}, } if self.has_bias: dim_partition_dict_mapping["bias"] = {} sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def collate_strategies(self) -> List[ShardingStrategy]: ''' Generate every possible strategies for a LayerNorm node, and record all strategies into the strategies_vector. ''' strategy_list = [] input_data_dim = len(self.op_data["input"].logical_shape) weight_data_dim = len(self.op_data["other"].logical_shape) # in LayerNorm context, batch dimensions mean all the dimensions do not join the normalization. batch_dimension_length = input_data_dim - weight_data_dim # SR = SR x R with single mesh dim on batch dimensions strategy_list.extend(self.split_input_batch_single_mesh_dim(0, batch_dimension_length)) strategy_list.extend(self.split_input_batch_single_mesh_dim(1, batch_dimension_length)) # SR = SR x R with both mesh dims on batch dimensions strategy_list.extend(self.split_input_batch_both_mesh_dim(0, 1, batch_dimension_length)) # RR = RR x R strategy_list.append(self.non_split()) return strategy_list
import copy import operator import warnings from functools import reduce from typing import List from colossalai.auto_parallel.tensor_shard.sharding_strategy import ( CommAction, CommType, MemoryCost, ShardingStrategy, TrainCycleItem, ) from colossalai.auto_parallel.tensor_shard.utils import ignore_sharding_exception from colossalai.tensor.shape_consistency import CollectiveCommPattern from .strategy_generator import StrategyGenerator class EmbeddingStrategyGenerator(StrategyGenerator): """ EmbeddingStrategyGenerator is a generic class to generate strategies for nn.Embedding or F.embedding. The operation data is defined as `output = input x other`. """ def validate(self) -> bool: return super().validate() def update_compute_cost(self, strategy: ShardingStrategy): ''' Compute the computation cost per device with this specific strategy. Note: The computation cost for the embedding handler is estimated as dense computing now. It may not be accurate. ''' # TODO: estimate the embedding computation cost as sparse operation sharded_input_shape = strategy.sharding_specs[self.op_data['input']].get_sharded_shape_per_device() sharded_other_shape = strategy.sharding_specs[self.op_data['other']].get_sharded_shape_per_device() sharded_output_shape = strategy.sharding_specs[self.op_data['output']].get_sharded_shape_per_device() input_size_product = reduce(operator.mul, sharded_input_shape) other_size_product = reduce(operator.mul, sharded_other_shape) output_size_product = reduce(operator.mul, sharded_output_shape) forward_compute_cost = input_size_product * other_size_product backward_activation_cost = other_size_product * output_size_product / sharded_output_shape[-1] backward_weight_cost = input_size_product * other_size_product backward_compute_cost = backward_weight_cost + backward_activation_cost total_compute_cost = forward_compute_cost + backward_compute_cost compute_cost = TrainCycleItem(fwd=forward_compute_cost, bwd=backward_compute_cost, total=total_compute_cost) strategy.compute_cost = compute_cost def update_memory_cost(self, strategy: ShardingStrategy): forward_size_mapping = { 'input': self._compute_size_in_bytes(strategy, "input"), 'other': self._compute_size_in_bytes(strategy, "other"), 'output': self._compute_size_in_bytes(strategy, "output") } backward_size_mapping = copy.deepcopy(forward_size_mapping) backward_size_mapping.pop("output") # compute fwd cost incurred # fwd_cost = input + other + output fwd_activation_cost = sum([v for k, v in forward_size_mapping.items() if not self.is_param(k)]) fwd_parameter_cost = sum([v for k, v in forward_size_mapping.items() if self.is_param(k)]) fwd_mem_cost = MemoryCost(activation=fwd_activation_cost, parameter=fwd_parameter_cost) # compute bwd cost incurred # bwd_cost = input_grad + other_grad bwd_activation_cost = sum([v for k, v in backward_size_mapping.items() if not self.is_param(k)]) bwd_parameter_cost = sum([v for k, v in backward_size_mapping.items() if self.is_param(k)]) bwd_mem_cost = MemoryCost(activation=bwd_activation_cost, parameter=bwd_parameter_cost) # compute total cost total_mem_cost = MemoryCost(activation=fwd_activation_cost + bwd_activation_cost, parameter=fwd_parameter_cost + bwd_parameter_cost) memory_cost = TrainCycleItem(fwd=fwd_mem_cost, bwd=bwd_mem_cost, total=total_mem_cost) strategy.memory_cost = memory_cost @ignore_sharding_exception def non_split(self): name = f'RR = R x RR' dim_partition_dict_mapping = { "input": {}, "other": {}, "output": {}, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping={}) @ignore_sharding_exception def split_input(self, mesh_dim_0): name = f'S{mesh_dim_0}R = S{mesh_dim_0} x RR' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0] }, "other": {}, "output": { 0: [mesh_dim_0], }, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) communication_action_mapping = {} if self.is_param("other"): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping["other"] = other_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_input_and_embedding_dim(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}S{mesh_dim_1} = S{mesh_dim_0} x RS{mesh_dim_1}' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0], }, "other": { 1: [mesh_dim_1], }, "output": { 0: [mesh_dim_0], 1: [mesh_dim_1], }, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action input_comm_action = self.get_communication_action( sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_1, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping = {"input": input_comm_action} if self.is_param("other"): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping["other"] = other_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_1d_parallel_on_input(self, mesh_dim_0, mesh_dim_1): name = f'S{mesh_dim_0}{mesh_dim_1}R = S{mesh_dim_0}{mesh_dim_1} x RR' dim_partition_dict_mapping = { "input": { 0: [mesh_dim_0, mesh_dim_1] }, "other": {}, "output": { 0: [mesh_dim_0, mesh_dim_1], }, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action communication_action_mapping = {} if self.is_param("other"): other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.HOOK) else: other_comm_action = self.get_communication_action( sharding_spec_mapping["other"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=1) communication_action_mapping["other"] = other_comm_action return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_embedding_dim(self, mesh_dim_0): name = f'RS{mesh_dim_0} = R x RS{mesh_dim_0}' dim_partition_dict_mapping = { "input": {}, "other": { 1: [mesh_dim_0], }, "output": { 1: [mesh_dim_0], }, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action input_comm_action = self.get_communication_action( sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=mesh_dim_0, comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping = {"input": input_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) @ignore_sharding_exception def split_1d_parallel_on_embedding_dim(self, mesh_dim_0, mesh_dim_1): name = f'RS{mesh_dim_0}{mesh_dim_1} = R x RS{mesh_dim_0}{mesh_dim_1}' dim_partition_dict_mapping = { "input": {}, "other": { 1: [mesh_dim_0, mesh_dim_1], }, "output": { 1: [mesh_dim_0, mesh_dim_1], }, } sharding_spec_mapping = self.to_sharding_spec_mapping(dim_partition_dict_mapping) # set communication action input_comm_action = self.get_communication_action( sharding_spec_mapping["input"], communication_pattern=CollectiveCommPattern.IDENTITY_FWD_ALLREDUCE_BWD, logical_process_axis=[mesh_dim_0, mesh_dim_1], comm_type=CommType.BEFORE, arg_index=0) communication_action_mapping = {"input": input_comm_action} return self.get_sharding_strategy(name=name, sharding_spec_mapping=sharding_spec_mapping, communication_action_mapping=communication_action_mapping) def collate_strategies(self) -> List[ShardingStrategy]: strategies = [] # RR= R x RR strategies.append(self.non_split()) # SR = S x RR strategies.append(self.split_input(0)) strategies.append(self.split_input(1)) # SS = S x RS strategies.append(self.split_input_and_embedding_dim(0, 1)) strategies.append(self.split_input_and_embedding_dim(1, 0)) # S01R = S01 x RR strategies.append(self.split_1d_parallel_on_input(0, 1)) # RS = R x RS strategies.append(self.split_embedding_dim(0)) strategies.append(self.split_embedding_dim(1)) # RS01 = R x RS01 strategies.append(self.split_1d_parallel_on_embedding_dim(0, 1)) return strategies