Update CI threshold (#2186)
This commit is contained in:
@@ -65,7 +65,9 @@ def patch_model(
|
||||
_to_torch(model)
|
||||
monkey_patch_vllm_all_gather()
|
||||
backup_ca_comm = tp_group.ca_comm
|
||||
# Use custom-allreduce here
|
||||
# Use custom-allreduce here.
|
||||
# We found the custom allreduce is much faster than the built-in allreduce in torch,
|
||||
# even with ENABLE_INTRA_NODE_COMM=1.
|
||||
# tp_group.ca_comm = None
|
||||
yield torch.compile(
|
||||
torch.no_grad()(model.forward), mode="max-autotune-no-cudagraphs"
|
||||
|
||||
@@ -36,7 +36,9 @@ class TestBenchServing(unittest.TestCase):
|
||||
)
|
||||
|
||||
if is_in_ci():
|
||||
self.assertGreater(res["output_throughput"], 950)
|
||||
# There is a regression with torch 2.5
|
||||
# This number was 950 for torch 2.4
|
||||
self.assertGreater(res["output_throughput"], 800)
|
||||
|
||||
def test_offline_throughput_without_radix_cache(self):
|
||||
res = run_bench_serving(
|
||||
|
||||
Reference in New Issue
Block a user