[Fix] Fix the case where prompt_len = 0 (#1593)

This commit is contained in:
Lianmin Zheng
2024-10-06 20:30:02 -07:00
committed by GitHub
parent 551a3a9d38
commit b6aad70ab1
3 changed files with 6 additions and 4 deletions

View File

@@ -526,7 +526,7 @@ class TokenizerManager:
async with self.model_update_lock:
# wait for the previous generation requests to finish
while len(self.rid_to_state) > 0:
await asyncio.sleep(0)
await asyncio.sleep(0.001)
self.send_to_scheduler.send_pyobj(obj)
self.model_update_result = asyncio.Future()
result = await self.model_update_result