Fix misusing the "_is_cuda". (#7091)

This commit is contained in:
sogalin
2025-06-12 02:21:31 +08:00
committed by GitHub
parent 25a6a9aa22
commit 02543b545c

View File

@@ -267,7 +267,7 @@ class MHATokenToKVPool(KVCache):
self.layer_transfer_counter = None
self.device_module = torch.get_device_module(self.device)
self.alt_stream = self.device_module.Stream() if is_cuda else None
self.alt_stream = self.device_module.Stream() if _is_cuda else None
k_size, v_size = self.get_kv_size_bytes()
logger.info(