diff --git a/python/sglang/srt/utils.py b/python/sglang/srt/utils.py index 707801604..6e2158c48 100644 --- a/python/sglang/srt/utils.py +++ b/python/sglang/srt/utils.py @@ -78,6 +78,11 @@ time_infos = {} HIP_FP8_E4M3_FNUZ_MAX = 224.0 +def get_bool_env_var(name: str, default: str = "false") -> bool: + value = os.getenv(name, default) + return value.lower() in ("true", "1") + + # https://pytorch.org/docs/stable/notes/hip.html#checking-for-hip def is_hip() -> bool: return torch.version.hip is not None @@ -128,9 +133,9 @@ def is_cuda_available(): return is_cuda() -_ENABLE_TORCH_INFERENCE_MODE = os.getenv( +_ENABLE_TORCH_INFERENCE_MODE = get_bool_env_var( "SGLANG_ENABLE_TORCH_INFERENCE_MODE", "false" -).lower() in ("true", "1") +) class DynamicGradMode(_DecoratorContextManager): @@ -1327,11 +1332,6 @@ def set_gpu_proc_affinity( logger.info(f"Process {pid} gpu_id {gpu_id} is running on CPUs: {p.cpu_affinity()}") -def get_bool_env_var(name: str, default: str = "false") -> bool: - value = os.getenv(name, default) - return value.lower() in ("true", "1") - - @lru_cache(maxsize=2) def disable_request_logging() -> bool: return get_bool_env_var("SGLANG_DISABLE_REQUEST_LOGGING")