From 02543b545c4d7aa6df3057402ad0ef3143f58810 Mon Sep 17 00:00:00 2001 From: sogalin <39478626+sogalin@users.noreply.github.com> Date: Thu, 12 Jun 2025 02:21:31 +0800 Subject: [PATCH] Fix misusing the "_is_cuda". (#7091) --- python/sglang/srt/mem_cache/memory_pool.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/python/sglang/srt/mem_cache/memory_pool.py b/python/sglang/srt/mem_cache/memory_pool.py index 79fb1b3b4..2f3b7fdb6 100644 --- a/python/sglang/srt/mem_cache/memory_pool.py +++ b/python/sglang/srt/mem_cache/memory_pool.py @@ -267,7 +267,7 @@ class MHATokenToKVPool(KVCache): self.layer_transfer_counter = None self.device_module = torch.get_device_module(self.device) - self.alt_stream = self.device_module.Stream() if is_cuda else None + self.alt_stream = self.device_module.Stream() if _is_cuda else None k_size, v_size = self.get_kv_size_bytes() logger.info(