# SPDX-License-Identifier: Apache-2.0 # SPDX-FileCopyrightText: Copyright contributors to the vLLM project from contextlib import contextmanager import torch import torch.cuda.nvtx as nvtx def print_tensor(tensor_obj, prefix, tensor_list=None): """Descends iterators that contains Tensors and prints the Tensor. Recursive function that descends iterator type arguments until it finds a Tensor object. """ if tensor_list is None: tensor_list = [] if isinstance(tensor_obj, (list, tuple)): for ten in tensor_obj: tensor_list = print_tensor(ten, prefix, tensor_list) elif isinstance(tensor_obj, torch.Tensor): tensor_dims = list(tensor_obj.size()) tensor_list.append(tensor_dims) return tensor_list def process_layer_params(module_obj): """Extract the static parameters from LLM and VLM relevant layer types""" param_info = {} # Extract parameters for layers commonly used in LLMs and VLMs if isinstance(module_obj, (torch.nn.Conv1d, torch.nn.Conv2d, torch.nn.Conv3d)): conv_params = {} conv_params["in_chan"] = module_obj.in_channels conv_params["out_chan"] = module_obj.out_channels conv_params["filter_dim"] = module_obj.kernel_size conv_params["stride"] = module_obj.stride conv_params["padding"] = module_obj.padding conv_params["dilation"] = module_obj.dilation conv_params["transposed"] = module_obj.transposed conv_params["output_padding"] = module_obj.output_padding conv_params["groups"] = module_obj.groups conv_params["padding_mode"] = module_obj.padding_mode param_info = conv_params elif isinstance( module_obj, ( torch.nn.ConvTranspose1d, torch.nn.ConvTranspose2d, torch.nn.ConvTranspose3d, ), ): convtranspose_params = {} convtranspose_params["in_chan"] = module_obj.in_channels convtranspose_params["out_chan"] = module_obj.out_channels convtranspose_params["filter_dim"] = module_obj.kernel_size convtranspose_params["stride"] = module_obj.stride convtranspose_params["padding"] = module_obj.padding convtranspose_params["dilation"] = module_obj.dilation convtranspose_params["transposed"] = module_obj.transposed convtranspose_params["output_padding"] = module_obj.output_padding convtranspose_params["groups"] = module_obj.groups convtranspose_params["padding_mode"] = module_obj.padding_mode param_info = convtranspose_params elif isinstance( module_obj, (torch.nn.MaxPool1d, torch.nn.MaxPool2d, torch.nn.MaxPool3d) ): def _handle_int_or_tuple(parameter): if isinstance(parameter, tuple): return list(parameter) elif isinstance(parameter, int): return [parameter, parameter] pooling_params = {} pooling_params["filter_dim"] = _handle_int_or_tuple(module_obj.kernel_size) pooling_params["stride"] = _handle_int_or_tuple(module_obj.stride) pooling_params["padding"] = _handle_int_or_tuple(module_obj.padding) pooling_params["dilation"] = _handle_int_or_tuple(module_obj.dilation) param_info = pooling_params elif isinstance( module_obj, (torch.nn.AvgPool1d, torch.nn.AvgPool2d, torch.nn.AvgPool3d) ): pooling_params = {} pooling_params["filter_dim"] = [ module_obj.kernel_size, module_obj.kernel_size, ] pooling_params["stride"] = [module_obj.stride, module_obj.stride] pooling_params["padding"] = [module_obj.padding, module_obj.padding] pooling_params["ceil_mode"] = module_obj.ceil_mode pooling_params["count_include_pad"] = module_obj.count_include_pad param_info = pooling_params elif isinstance( module_obj, ( torch.nn.AdaptiveAvgPool1d, torch.nn.AdaptiveAvgPool2d, torch.nn.AdaptiveAvgPool3d, ), ): pooling_params = {} pooling_params["output_size"] = [ module_obj.output_size, module_obj.output_size, ] param_info = pooling_params elif isinstance(module_obj, torch.nn.Linear): param_info["in_features"] = module_obj.in_features param_info["out_features"] = module_obj.out_features elif isinstance( module_obj, (torch.nn.BatchNorm1d, torch.nn.BatchNorm2d, torch.nn.BatchNorm3d), ): param_info["num_features"] = module_obj.num_features param_info["epsilon"] = module_obj.eps param_info["momentum"] = module_obj.momentum elif isinstance(module_obj, torch.nn.ReLU): param_info["in_place"] = module_obj.inplace elif isinstance(module_obj, torch.nn.Dropout): param_info["p"] = module_obj.p param_info["in_place"] = module_obj.inplace elif isinstance(module_obj, torch.nn.Embedding): param_info["num_embeddings"] = module_obj.num_embeddings param_info["embedding_dim"] = module_obj.embedding_dim elif isinstance( module_obj, ( torch.nn.Upsample, torch.nn.UpsamplingNearest2d, torch.nn.UpsamplingBilinear2d, ), ): param_info["scale_factor"] = module_obj.scale_factor return param_info def construct_marker_dict_and_push( module_name, module_obj, in_tensor, kwargs=None, out_tensor=None ): marker_dict = {} marker_dict["Module"] = module_name ## Get trainable parameters like weights and bias module_params = module_obj.named_parameters(recurse=False) for idx, (param_name, param_obj) in enumerate(module_params): if idx == 0: marker_dict["TrainableParams"] = {} marker_dict["TrainableParams"][param_name] = list(param_obj.size()) in_tensor_list = print_tensor(in_tensor, "Input") if in_tensor_list: marker_dict["Inputs"] = in_tensor_list out_tensor_list = print_tensor(out_tensor, "Output") if out_tensor_list: marker_dict["Outputs"] = out_tensor_list ## Get Kwargs like input_ids and positions for the top module if kwargs: for key, value in kwargs.items(): if isinstance(value, (torch.Tensor, list, tuple)): tensor_list = print_tensor(value, key) if tensor_list: marker_dict[key] = tensor_list param_info = process_layer_params(module_obj) if param_info: marker_dict["StaticParams"] = param_info nvtx.range_push("{}".format(marker_dict)) class ResultHolder: """Holder for storing results from within a context manager.""" result = None @contextmanager def layerwise_nvtx_marker_context(module_name, module_obj, in_tensor=None, kwargs=None): """Context manager for NVTX markers that automatically pushes on enter and pops on exit. Example: with nvtx_marker_context("Module:MyModule", module, in_tensor=args, kwargs=kwargs) as ctx: ctx.result = module(*args, **kwargs) return ctx.result """ holder = ResultHolder() # Push input marker construct_marker_dict_and_push( module_name, module_obj, in_tensor=in_tensor, kwargs=kwargs, ) try: yield holder finally: # Pop input marker nvtx.range_pop() # Push and pop output marker output_name = module_name.replace("(input)", "(output)") construct_marker_dict_and_push( output_name, module_obj, in_tensor=None, kwargs=None, out_tensor=holder.result, ) nvtx.range_pop() class PytHooks: """This module contains all the code needed to enable forward hooks in a pytorch network. To register the hooks for a given network, the user needs to instantiate a PytHook object. Then call the register_hooks method. Example: my_hook = PytHook() my_hook.register_hooks(my_network_model) """ def __init__(self): """Initialize module variables.""" super().__init__() self.module_to_name_map = {} def _process_layer_params(self, module_obj): return process_layer_params(module_obj) def module_fwd_hook(self, module_obj, in_tensor, out_tensor): """Callback function that ends the NVTX marker. Records the module name and tensor information. Called after the module executes the forward method. """ nvtx.range_pop() module_name = self.module_to_name_map.get(module_obj, "unknown") construct_marker_dict_and_push( module_name, module_obj, in_tensor=None, kwargs=None, out_tensor=out_tensor ) nvtx.range_pop() return def module_fwd_pre_hook(self, module_obj, in_tensor, kwargs): """Creates an NVTX marker with the module name in it. This function is called before the module executes. """ module_name = self.module_to_name_map.get(module_obj, "unknown") construct_marker_dict_and_push( module_name, module_obj, in_tensor=in_tensor, kwargs=kwargs, out_tensor=None ) return def register_hooks(self, network_model, module_prefix="top"): """User level function that activates all the hooks. The user needs to call this method from the network source code. The code descends all the modules in the network and registers their respective hooks. """ # Module types to skip (simple operations that don't need detailed profiling) skip_types = ( torch.nn.Identity, torch.nn.Dropout, torch.nn.Dropout1d, torch.nn.Dropout2d, torch.nn.Dropout3d, ) for name, module in network_model.named_modules(prefix=module_prefix): # Skip certain module types to reduce profiling overhead if isinstance(module, skip_types): continue module.register_forward_pre_hook(self.module_fwd_pre_hook, with_kwargs=True) module.register_forward_hook(self.module_fwd_hook) if module not in self.module_to_name_map: self.module_to_name_map[module] = name else: raise ValueError("Module instance {} is not unique ".format(module)) return