From bb1f0d5a62046c90eab24f037e48164773c33dc5 Mon Sep 17 00:00:00 2001 From: rjg-lyh <83491835+rjg-lyh@users.noreply.github.com> Date: Mon, 22 Sep 2025 17:17:31 +0800 Subject: [PATCH] [main] remove the redundant log prints in register_custom_ops.py (#3094) ### What this PR does / why we need it? This PR removed the redundant log prints in register_custom_ops.py, in order to make output clear. ### Does this PR introduce _any_ user-facing change? No. ### How was this patch tested? CI passed with new added/existing test. - vLLM version: v0.10.2 - vLLM main: https://github.com/vllm-project/vllm/commit/9607d5eb449711b349d4c2bee0a9c94afcc7ed14 Signed-off-by: rjg-lyh <1318825571@qq.com> --- vllm_ascend/ops/register_custom_ops.py | 7 ------- 1 file changed, 7 deletions(-) diff --git a/vllm_ascend/ops/register_custom_ops.py b/vllm_ascend/ops/register_custom_ops.py index 916b86f..1267066 100644 --- a/vllm_ascend/ops/register_custom_ops.py +++ b/vllm_ascend/ops/register_custom_ops.py @@ -7,7 +7,6 @@ from vllm.distributed import (get_tensor_model_parallel_rank, tensor_model_parallel_all_reduce, tensor_model_parallel_reduce_scatter) from vllm.forward_context import get_forward_context -from vllm.logger import logger from vllm.utils import direct_register_custom_op import vllm_ascend.envs as envs_ascend @@ -18,7 +17,6 @@ def _maybe_chunk_residual_impl(x: torch.Tensor, try: forward_context = get_forward_context() except AssertionError: - logger.info("Forward context is None, skipping the operation.") return residual if x.size(0) != residual.size(0): @@ -41,7 +39,6 @@ def _maybe_all_gather_and_maybe_unpad_impl(x: torch.Tensor, try: forward_context = get_forward_context() except AssertionError: - logger.info("Forward context is None, skipping the operation.") return x flashcomm_v1_enabled = forward_context.flashcomm_v1_enabled @@ -57,7 +54,6 @@ def _maybe_pad_and_reduce_impl(x: torch.Tensor) -> torch.Tensor: try: forward_context = get_forward_context() except AssertionError: - logger.info("Forward context is None, skipping the operation.") return tensor_model_parallel_all_reduce(x) flashcomm_v1_enabled = forward_context.flashcomm_v1_enabled @@ -75,7 +71,6 @@ def _maybe_prefetch_mlp_gate_up_proj_impl(x_dependency: torch.Tensor, try: forward_context = get_forward_context() except AssertionError: - logger.info("Forward context is None, skipping the operation.") return if not forward_context.prefetch_mlp_enabled: @@ -106,7 +101,6 @@ def _maybe_prefetch_mlp_down_proj_impl(x_dependency: torch.Tensor) -> None: try: forward_context = get_forward_context() except AssertionError: - logger.info("Forward context is None, skipping the operation.") return if not forward_context.prefetch_mlp_enabled: @@ -136,7 +130,6 @@ def _maybe_wait_prefetch_done_impl(x: torch.Tensor) -> None: try: forward_context = get_forward_context() except AssertionError: - logger.info("Forward context is None, skipping the operation.") return if not forward_context.prefetch_mlp_enabled: