diff --git a/.github/workflows/pr-test-rust.yml b/.github/workflows/pr-test-rust.yml index cc44192cb..d704488d9 100644 --- a/.github/workflows/pr-test-rust.yml +++ b/.github/workflows/pr-test-rust.yml @@ -56,7 +56,7 @@ jobs: e2e-python: if: github.repository == 'sgl-project/sglang' || github.event_name == 'pull_request' - runs-on: 2-gpu-runner + runs-on: BM.A10.4 timeout-minutes: 30 steps: - name: Checkout code @@ -66,6 +66,10 @@ jobs: run: | bash scripts/ci/ci_install_rust.sh + - name: Install SGLang dependencies + run: | + sudo bash scripts/ci/ci_install_dependency.sh + - name: Build python binding run: | source "$HOME/.cargo/env" diff --git a/sgl-router/README.md b/sgl-router/README.md index 42d1bb314..2bd4ed19f 100644 --- a/sgl-router/README.md +++ b/sgl-router/README.md @@ -43,6 +43,7 @@ python -m build && pip install --force-reinstall dist/*.whl ``` #### Option B: Development Mode + ```bash pip install -e . ``` diff --git a/sgl-router/py_test/test_launch_server.py b/sgl-router/py_test/test_launch_server.py index a01f5cfcc..f805ff117 100644 --- a/sgl-router/py_test/test_launch_server.py +++ b/sgl-router/py_test/test_launch_server.py @@ -251,7 +251,7 @@ class TestLaunchServer(unittest.TestCase): metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg) @@ -291,7 +291,7 @@ class TestLaunchServer(unittest.TestCase): ) metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg) @@ -305,7 +305,7 @@ class TestLaunchServer(unittest.TestCase): # 5. run mmlu again metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg) @@ -358,7 +358,7 @@ class TestLaunchServer(unittest.TestCase): ) metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg) @@ -485,7 +485,7 @@ class TestLaunchServer(unittest.TestCase): metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"dp aware MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg) @@ -529,7 +529,7 @@ class TestLaunchServer(unittest.TestCase): ) metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg) @@ -543,7 +543,7 @@ class TestLaunchServer(unittest.TestCase): # 5. Run mmlu again metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg) @@ -621,7 +621,7 @@ class TestLaunchServer(unittest.TestCase): ) metrics = run_eval(args) score = metrics["score"] - THRESHOLD = 0.65 + THRESHOLD = 0.635 passed = score >= THRESHOLD msg = f"MMLU test {'passed' if passed else 'failed'} with score {score:.3f} (threshold: {THRESHOLD})" self.assertGreaterEqual(score, THRESHOLD, msg)