add vxpu
This commit is contained in:
48
vllm_kunlun/patch/platform/patch_executor.py
Normal file
48
vllm_kunlun/patch/platform/patch_executor.py
Normal file
@@ -0,0 +1,48 @@
|
||||
import time
|
||||
|
||||
from vllm.executor.executor_base import logger, ExecutorBase
|
||||
|
||||
|
||||
original_init = ExecutorBase.__init__
|
||||
def init(self, *args, **kwargs):
|
||||
original_init(self, *args, **kwargs)
|
||||
self.is_offloaded = False
|
||||
|
||||
|
||||
def offload_vram(self):
|
||||
if self.is_offloaded:
|
||||
logger.warning("Executor is already offloaded.")
|
||||
return
|
||||
time_before_offload = time.perf_counter()
|
||||
self.collective_rpc("offload_vram")
|
||||
time_after_offload = time.perf_counter()
|
||||
|
||||
self.is_offloaded = True
|
||||
logger.info(f"Offloading VRAM costs {time_after_offload - time_before_offload:.6f} seconds.")
|
||||
|
||||
|
||||
def reload_vram(self) -> bool:
|
||||
if not self.is_offloaded:
|
||||
logger.warning("Executor is not offloaded.")
|
||||
return True
|
||||
|
||||
while True:
|
||||
time_before_reload = time.perf_counter()
|
||||
res = self.collective_rpc("try_reload_vram")
|
||||
time_after_reload = time.perf_counter()
|
||||
|
||||
succ = all(x[0] for x in res)
|
||||
if succ:
|
||||
self.is_offloaded = False
|
||||
logger.info(f"Reloading VRAM costs {time_after_reload - time_before_reload:.6f} seconds.")
|
||||
prev_is_self = all(x[1] for x in res)
|
||||
return prev_is_self
|
||||
else:
|
||||
# some workers not get lock
|
||||
self.collective_rpc("vnpu_unlock_gpu")
|
||||
time.sleep(0.001)
|
||||
|
||||
|
||||
ExecutorBase.__init__ = init
|
||||
ExecutorBase.offload_vram = offload_vram
|
||||
ExecutorBase.reload_vram = reload_vram
|
||||
Reference in New Issue
Block a user