The attn mask was declared in the mla.py,we don't need the splitfuse
mask when mla chunkprefill, and this mask will cause memory problem when
long context like 64k or 128k
- vLLM version: v0.10.0
- vLLM main:
14a5d903ab
---------
Signed-off-by: haojiangzheng <justineric096@gmail.com>