From 5f2595be430239ba13c5adbe559e21333f5adf9e Mon Sep 17 00:00:00 2001 From: Yineng Zhang Date: Sun, 15 Dec 2024 02:47:26 +0800 Subject: [PATCH] hotfix: checking for HIP (#2485) --- python/sglang/srt/layers/fused_moe_triton/layer.py | 2 +- python/sglang/srt/utils.py | 8 +------- 2 files changed, 2 insertions(+), 8 deletions(-) diff --git a/python/sglang/srt/layers/fused_moe_triton/layer.py b/python/sglang/srt/layers/fused_moe_triton/layer.py index d9503fe20..0785583cd 100644 --- a/python/sglang/srt/layers/fused_moe_triton/layer.py +++ b/python/sglang/srt/layers/fused_moe_triton/layer.py @@ -19,7 +19,7 @@ from sglang.srt.layers.quantization.base_config import ( ) from sglang.srt.utils import set_weight_attrs -if torch.cuda.is_available() or torch.hip.is_available(): +if torch.cuda.is_available(): from sglang.srt.layers.fused_moe_triton.fused_moe import fused_experts else: fused_experts = None # type: ignore diff --git a/python/sglang/srt/utils.py b/python/sglang/srt/utils.py index dabc608fa..be339d5d2 100644 --- a/python/sglang/srt/utils.py +++ b/python/sglang/srt/utils.py @@ -92,7 +92,7 @@ def is_flashinfer_available(): """ if not get_bool_env_var("SGLANG_IS_FLASHINFER_AVAILABLE", default="true"): return False - return torch.cuda.is_available() and not is_hip() + return torch.cuda.is_available() and torch.version.cuda def is_ipv6(address): @@ -1071,9 +1071,6 @@ def get_device_name(device_id: int = 0) -> str: if hasattr(torch, "cuda") and torch.cuda.is_available(): return torch.cuda.get_device_name(device_id) - if hasattr(torch, "hip") and torch.hip.is_available(): - return torch.hip.get_device_name(device_id) - if hasattr(torch, "xpu") and torch.xpu.is_available(): return torch.xpu.get_device_name(device_id) @@ -1086,9 +1083,6 @@ def get_device_capability(device_id: int = 0) -> Tuple[int, int]: if hasattr(torch, "cuda") and torch.cuda.is_available(): major, minor = torch.cuda.get_device_capability(device_id) - if hasattr(torch, "hip") and torch.hip.is_available(): - major, minor = torch.cuda.get_device_capability(device_id) - if hasattr(torch, "xpu") and torch.xpu.is_available(): major, minor, *_ = torch.xpu.get_device_capability(device_id)["version"].split( "."