[v0.18.0][Triton][Qwen3.5] delete expr for kernels args (#7646)

### What this PR does / why we need it?
Some parameters of Triton operators are unnecessarily modified with the
"constexpr" modifier. When these parameters change, recompilation is
triggered, which significantly affects the model performance. Therefore,
these parameters need to be rectified.
backport: https://github.com/vllm-project/vllm-ascend/pull/7482


Signed-off-by: w30012745 <wangxiaoshuai2@h-partners.com>
Co-authored-by: w30012745 <wangxiaoshuai2@h-partners.com>
This commit is contained in:
Mr.WXS
2026-03-25 23:31:27 +08:00
committed by GitHub
parent dd55736ee4
commit dba34d4915
4 changed files with 13 additions and 13 deletions

View File

@@ -12,7 +12,7 @@ from vllm.triton_utils import tl, triton
@triton.heuristics({"HAS_BIAS": lambda args: args["B"] is not None})
@triton.heuristics({"HAS_Z": lambda args: args["Z"] is not None})
@triton.jit
@triton.jit(do_not_specialize=["stride_x_row", "stride_y_row", "stride_z_row", "M", "N", "eps"])
def _layer_norm_fwd_1pass_kernel_npu(
X, # pointer to the input
Y, # pointer to the output