Support server based rollout in Verlengine (#4848)
Co-authored-by: Jin Pan <jpan236@wisc.edu> Co-authored-by: Chayenne <zhaochen20@outlook.com> Co-authored-by: Jinn <47354855+jhinpan@users.noreply.github.com>
This commit is contained in:
@@ -38,6 +38,7 @@ import torch
|
||||
import uvloop
|
||||
|
||||
from sglang.srt.code_completion_parser import load_completion_template_for_openai_api
|
||||
from sglang.srt.entrypoints.EngineBase import EngineBase
|
||||
from sglang.srt.managers.data_parallel_controller import (
|
||||
run_data_parallel_controller_process,
|
||||
)
|
||||
@@ -78,7 +79,7 @@ logger = logging.getLogger(__name__)
|
||||
asyncio.set_event_loop_policy(uvloop.EventLoopPolicy())
|
||||
|
||||
|
||||
class Engine:
|
||||
class Engine(EngineBase):
|
||||
"""
|
||||
The entry point to the inference engine.
|
||||
|
||||
|
||||
Reference in New Issue
Block a user