[CI] upgrade to vllm 0.9.0 (#959)

Upgrade to vllm 0.9.0.
0.8.5 will not be supported any more.

Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com>
This commit is contained in:
wangxiyuan
2025-05-28 21:18:41 +08:00
committed by GitHub
parent e2a0c19cea
commit f6e5decc10
16 changed files with 79 additions and 146 deletions

View File

@@ -26,9 +26,9 @@ import torch
from vllm import SamplingParams
from vllm.sequence import PromptLogprobs, SampleLogprobs
from ....model_utils import (TokensTextLogprobs,
TokensTextLogprobsPromptLogprobs,
check_logprobs_close, check_outputs_equal)
from tests.model_utils import (TokensTextLogprobs,
TokensTextLogprobsPromptLogprobs,
check_logprobs_close, check_outputs_equal)
PROMPTS = [
"Hello, my name is",