Upgrade to vllm 0.17.0 corex v4.1 overlay
This commit is contained in:
@@ -66,6 +66,23 @@ class WorkspaceManager:
|
||||
],
|
||||
)
|
||||
|
||||
def unlock(self) -> None:
|
||||
"""Unlock the workspace to allow growth.
|
||||
|
||||
This is used during elastic EP scaling when the workspace size
|
||||
needs to grow due to changes in the number of experts.
|
||||
"""
|
||||
self._locked = False
|
||||
if envs.VLLM_DEBUG_WORKSPACE:
|
||||
logger.info(
|
||||
"[WORKSPACE DEBUG] Workspace unlocked. Current sizes: %s",
|
||||
[
|
||||
self._workspace_size_bytes(ws) / _MB
|
||||
for ws in self._current_workspaces
|
||||
if ws is not None
|
||||
],
|
||||
)
|
||||
|
||||
def is_locked(self) -> bool:
|
||||
"""Check if workspace is locked."""
|
||||
return self._locked
|
||||
@@ -242,6 +259,17 @@ def lock_workspace() -> None:
|
||||
current_workspace_manager().lock()
|
||||
|
||||
|
||||
def unlock_workspace() -> None:
|
||||
"""Unlock the workspace to allow growth.
|
||||
|
||||
This is used during elastic EP scaling when the workspace size
|
||||
needs to grow due to changes in the number of experts.
|
||||
After scaling operations complete, lock_workspace() should be
|
||||
called again to prevent unexpected allocations.
|
||||
"""
|
||||
current_workspace_manager().unlock()
|
||||
|
||||
|
||||
def reset_workspace_manager() -> None:
|
||||
"""Reset the workspace manager to uninitialized state.
|
||||
|
||||
|
||||
Reference in New Issue
Block a user