Upgrade to vllm 0.17.0 corex v4.1 overlay

This commit is contained in:
2026-04-29 19:38:22 +08:00
parent 8fac6062e4
commit 938d0854a5
430 changed files with 35969 additions and 14511 deletions

View File

@@ -86,10 +86,7 @@ def cuda_platform_plugin() -> str | None:
pynvml.nvmlShutdown()
except Exception as e:
logger.debug("Exception happens when checking CUDA platform: %s", str(e))
nvml_related_error = (
"nvml" in e.__class__.__name__.lower() or "nvml" in str(e).lower()
)
if not nvml_related_error:
if "nvml" not in e.__class__.__name__.lower():
# If the error is not related to NVML, re-raise it.
raise e
@@ -105,40 +102,7 @@ def cuda_platform_plugin() -> str | None:
logger.debug("Confirmed CUDA platform is available on Jetson.")
is_cuda = True
else:
try:
import torch
if vllm_version_matches_substr("cpu"):
logger.debug(
"CUDA platform torch fallback is skipped because "
"vLLM is built with CPU."
)
elif not torch.cuda._is_compiled():
logger.debug(
"CUDA platform torch fallback reports CUDA is not compiled."
)
elif not torch.cuda.is_available():
logger.debug(
"CUDA platform torch fallback reports "
"torch.cuda.is_available()=False."
)
elif torch.cuda.device_count() <= 0:
logger.debug(
"CUDA platform torch fallback reports no GPU is found."
)
else:
logger.debug(
"Confirmed CUDA platform is available via torch fallback "
"after NVML check failed."
)
is_cuda = True
except Exception as torch_e:
logger.debug(
"CUDA platform torch fallback failed because: %s", str(torch_e)
)
if not is_cuda:
logger.debug("CUDA platform is not available because: %s", str(e))
logger.debug("CUDA platform is not available because: %s", str(e))
return "vllm.platforms.cuda.CudaPlatform" if is_cuda else None