Improve the code style: more comments and remove useless packages (#1139)
This commit is contained in:
@@ -22,8 +22,6 @@ import uuid
|
||||
from dataclasses import dataclass
|
||||
from typing import Dict, List, Optional, Union
|
||||
|
||||
import torch
|
||||
|
||||
from sglang.srt.managers.schedule_batch import BaseFinishReason
|
||||
from sglang.srt.sampling_params import SamplingParams
|
||||
|
||||
@@ -43,9 +41,9 @@ class GenerateReqInput:
|
||||
rid: Optional[Union[List[str], str]] = None
|
||||
# Whether to return logprobs.
|
||||
return_logprob: Optional[Union[List[bool], bool]] = None
|
||||
# The start location of the prompt for return_logprob.
|
||||
# If return logprobs, the start location in the prompt for returning logprobs.
|
||||
logprob_start_len: Optional[Union[List[int], int]] = None
|
||||
# The number of top logprobs to return.
|
||||
# If return logprobs, the number of top logprobs to return at each position.
|
||||
top_logprobs_num: Optional[Union[List[int], int]] = None
|
||||
# Whether to detokenize tokens in text in the returned logprobs.
|
||||
return_text_in_logprobs: bool = False
|
||||
@@ -155,16 +153,27 @@ class GenerateReqInput:
|
||||
|
||||
@dataclass
|
||||
class TokenizedGenerateReqInput:
|
||||
# The request id
|
||||
rid: str
|
||||
# The input text
|
||||
input_text: str
|
||||
# The input token ids
|
||||
input_ids: List[int]
|
||||
# The pixel values for input images
|
||||
pixel_values: List[float]
|
||||
# The hash of input images
|
||||
image_hash: int
|
||||
# The image size
|
||||
image_size: List[int]
|
||||
# The sampling parameters
|
||||
sampling_params: SamplingParams
|
||||
# Whether to return the logprobs
|
||||
return_logprob: bool
|
||||
# If return logprobs, the start location in the prompt for returning logprobs.
|
||||
logprob_start_len: int
|
||||
# If return logprobs, the number of top logprobs to return at each position.
|
||||
top_logprobs_num: int
|
||||
# Whether to stream output
|
||||
stream: bool
|
||||
|
||||
|
||||
@@ -215,15 +224,21 @@ class EmbeddingReqInput:
|
||||
|
||||
@dataclass
|
||||
class TokenizedEmbeddingReqInput:
|
||||
# The request id
|
||||
rid: str
|
||||
# The input text
|
||||
input_text: str
|
||||
# The input token ids
|
||||
input_ids: List[int]
|
||||
# Dummy sampling params for compatibility
|
||||
sampling_params: SamplingParams
|
||||
|
||||
|
||||
@dataclass
|
||||
class BatchTokenIDOut:
|
||||
# The request id
|
||||
rids: List[str]
|
||||
# The version id to sync decode status with in detokenizer_manager
|
||||
vids: List[int]
|
||||
decoded_texts: List[str]
|
||||
decode_ids: List[int]
|
||||
@@ -236,17 +251,25 @@ class BatchTokenIDOut:
|
||||
|
||||
@dataclass
|
||||
class BatchStrOut:
|
||||
# The request id
|
||||
rids: List[str]
|
||||
# The output decoded strings
|
||||
output_strs: List[str]
|
||||
# The meta info
|
||||
meta_info: List[Dict]
|
||||
# The finish reason
|
||||
finished_reason: List[BaseFinishReason]
|
||||
|
||||
|
||||
@dataclass
|
||||
class BatchEmbeddingOut:
|
||||
# The request id
|
||||
rids: List[str]
|
||||
# The output embedding
|
||||
embeddings: List[List[float]]
|
||||
# The meta info
|
||||
meta_info: List[Dict]
|
||||
# The finish reason
|
||||
finished_reason: List[BaseFinishReason]
|
||||
|
||||
|
||||
@@ -257,9 +280,5 @@ class FlushCacheReq:
|
||||
|
||||
@dataclass
|
||||
class AbortReq:
|
||||
# The request id
|
||||
rid: str
|
||||
|
||||
|
||||
@dataclass
|
||||
class DetokenizeReqInput:
|
||||
input_ids: List[int]
|
||||
|
||||
Reference in New Issue
Block a user