Fix misusing the "_is_cuda". (#7091)
This commit is contained in:
@@ -267,7 +267,7 @@ class MHATokenToKVPool(KVCache):
|
|||||||
|
|
||||||
self.layer_transfer_counter = None
|
self.layer_transfer_counter = None
|
||||||
self.device_module = torch.get_device_module(self.device)
|
self.device_module = torch.get_device_module(self.device)
|
||||||
self.alt_stream = self.device_module.Stream() if is_cuda else None
|
self.alt_stream = self.device_module.Stream() if _is_cuda else None
|
||||||
|
|
||||||
k_size, v_size = self.get_kv_size_bytes()
|
k_size, v_size = self.get_kv_size_bytes()
|
||||||
logger.info(
|
logger.info(
|
||||||
|
|||||||
Reference in New Issue
Block a user