[Version] Drop 0.16.0 support (#7153)
### What this PR does / why we need it?
Drop 0.16.0 support in main
- Fix eagle proposer break introduced by
https://github.com/vllm-project/vllm/pull/34552. Mainly change to use
the draft attention group to initialize the attention metadata builder.
- Fix the `ModelRunner` has no attribute `cudagraph_capture_sizes`
error, which is a bug in vLLM v0.17.0, and fixed by a later pr
https://github.com/vllm-project/vllm/pull/30515
- vLLM version: v0.16.0
- vLLM main:
4034c3d32e
---------
Signed-off-by: MengqingCao <cmq0113@163.com>
This commit is contained in:
@@ -1,4 +1,5 @@
|
||||
from unittest.mock import MagicMock, patch
|
||||
import unittest
|
||||
|
||||
import numpy as np
|
||||
import torch
|
||||
@@ -137,7 +138,7 @@ class TestEagleProposerInitialization(TestBase):
|
||||
expected_max_num_tokens = proposer.max_num_tokens
|
||||
self.assertEqual(proposer.hidden_states.shape, (expected_max_num_tokens, 2048))
|
||||
|
||||
|
||||
@unittest.skip("Skip due to the changes in #7153, fix me later")
|
||||
class TestEagleProposerLoadModel(TestBase):
|
||||
def setUp(self):
|
||||
self.vllm_config = MagicMock(spec=VllmConfig)
|
||||
|
||||
Reference in New Issue
Block a user