[Fix] Window attention compatible with RadixAttention and chunked prefill (#1112)

This commit is contained in:
Ying Sheng
2024-08-15 10:33:20 -07:00
committed by GitHub
parent 9195d1362a
commit 93d4e354d8
5 changed files with 37 additions and 56 deletions

View File

@@ -334,11 +334,7 @@ class ModelRunner:
dtype=torch.uint8,
device="cuda",
)
self.flashinfer_prefill_wrapper_ragged = (
BatchPrefillWithRaggedKVCacheWrapper(
self.flashinfer_workspace_buffer, "NHD"
)
)
self.flashinfer_prefill_wrapper_ragged = None
self.flashinfer_prefill_wrapper_paged = []
self.flashinfer_decode_wrapper = []
for i in range(2):