Feature: support code completion (#3612)
This commit is contained in:
@@ -36,6 +36,7 @@ setattr(threading, "_register_atexit", lambda *args, **kwargs: None)
|
||||
import torch
|
||||
import uvloop
|
||||
|
||||
from sglang.srt.code_completion_parser import load_completion_template_for_openai_api
|
||||
from sglang.srt.managers.data_parallel_controller import (
|
||||
run_data_parallel_controller_process,
|
||||
)
|
||||
@@ -538,6 +539,9 @@ def _launch_subprocesses(
|
||||
tokenizer_manager, server_args.chat_template, server_args.model_path
|
||||
)
|
||||
|
||||
if server_args.completion_template:
|
||||
load_completion_template_for_openai_api(server_args.completion_template)
|
||||
|
||||
# Wait for the model to finish loading
|
||||
scheduler_infos = []
|
||||
for i in range(len(scheduler_pipe_readers)):
|
||||
|
||||
Reference in New Issue
Block a user