Feature: support code completion (#3612)

This commit is contained in:
woodx
2025-03-17 09:26:19 +08:00
committed by GitHub
parent 9b8333d992
commit 48efec7b05
6 changed files with 266 additions and 1 deletions

View File

@@ -36,6 +36,7 @@ setattr(threading, "_register_atexit", lambda *args, **kwargs: None)
import torch
import uvloop
from sglang.srt.code_completion_parser import load_completion_template_for_openai_api
from sglang.srt.managers.data_parallel_controller import (
run_data_parallel_controller_process,
)
@@ -538,6 +539,9 @@ def _launch_subprocesses(
tokenizer_manager, server_args.chat_template, server_args.model_path
)
if server_args.completion_template:
load_completion_template_for_openai_api(server_args.completion_template)
# Wait for the model to finish loading
scheduler_infos = []
for i in range(len(scheduler_pipe_readers)):