From 29a610b4d9aaf430da38a37c71d8d5f781f8cf99 Mon Sep 17 00:00:00 2001 From: Lifu Huang Date: Mon, 11 Aug 2025 22:59:56 -0700 Subject: [PATCH] Fix broken CI TestRequestLengthValidation (#9095) --- .../openai_server/validation/test_request_length_validation.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test/srt/openai_server/validation/test_request_length_validation.py b/test/srt/openai_server/validation/test_request_length_validation.py index b3c202f64..7276906d4 100644 --- a/test/srt/openai_server/validation/test_request_length_validation.py +++ b/test/srt/openai_server/validation/test_request_length_validation.py @@ -79,7 +79,7 @@ class TestRequestLengthValidation(CustomTestCase): ) self.assertIn( - "Requested token count exceeds the model's maximum context", + "max_completion_tokens is too large", str(cm.exception), )