Files
xc-llm-ascend/vllm_ascend
无脸男 ace300a549 [Bugfix] Fix the abnormal NPU memory usage in full graph mode. (#3331)
### What this PR does / why we need it?

In the full graph mode, since paged attention operators updates are
required, the parameters of this operators needs to be retained.
However, the tensor such as query、key cache、value cache, does not need
to be persistently saved, and we can manually release this space by
`weak_ref_tensor` to save the memory.

### Does this PR introduce _any_ user-facing change?

### How was this patch tested?

- vLLM version: v0.11.0rc3
- vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0

Signed-off-by: WithHades <244036962@qq.com>
2025-10-11 10:20:10 +08:00
..
2025-09-23 10:27:14 +08:00
2025-10-09 10:28:38 +08:00
2025-10-09 19:22:46 +08:00
2025-09-30 03:25:58 +08:00
2025-10-11 08:36:20 +08:00