Unverified Commit da6f8081 authored by Lianmin Zheng's avatar Lianmin Zheng Committed by GitHub
Browse files

Fix CI tests (#3132)

parent 9286740e
...@@ -43,6 +43,8 @@ jobs: ...@@ -43,6 +43,8 @@ jobs:
- name: Run test - name: Run test
timeout-minutes: 10 timeout-minutes: 10
env:
OPENAI_API_KEY: ${{ secrets.OPENAI_API_KEY }}
run: | run: |
cd test/lang cd test/lang
python3 run_suite.py --suite per-commit python3 run_suite.py --suite per-commit
......
...@@ -49,7 +49,7 @@ class TestBenchServing(unittest.TestCase): ...@@ -49,7 +49,7 @@ class TestBenchServing(unittest.TestCase):
) )
# There is a regression with torch 2.5 # There is a regression with torch 2.5
# This number was 950 for torch 2.4 # This number was 950 for torch 2.4
self.assertGreater(res["output_throughput"], 850) self.assertGreater(res["output_throughput"], 1000)
def test_offline_throughput_without_radix_cache(self): def test_offline_throughput_without_radix_cache(self):
res = run_bench_serving( res = run_bench_serving(
...@@ -114,7 +114,7 @@ class TestBenchServing(unittest.TestCase): ...@@ -114,7 +114,7 @@ class TestBenchServing(unittest.TestCase):
f"### test_offline_throughput_default_fp8\n" f"### test_offline_throughput_default_fp8\n"
f'Output throughput: {res["output_throughput"]:.2f} token/s\n' f'Output throughput: {res["output_throughput"]:.2f} token/s\n'
) )
self.assertGreater(res["output_throughput"], 3850) self.assertGreater(res["output_throughput"], 3900)
def test_online_latency_default(self): def test_online_latency_default(self):
res = run_bench_serving( res = run_bench_serving(
...@@ -129,7 +129,7 @@ class TestBenchServing(unittest.TestCase): ...@@ -129,7 +129,7 @@ class TestBenchServing(unittest.TestCase):
f"### test_online_latency_default\n" f"### test_online_latency_default\n"
f'median_e2e_latency_ms : {res["median_e2e_latency_ms"]:.2f} ms\n' f'median_e2e_latency_ms : {res["median_e2e_latency_ms"]:.2f} ms\n'
) )
self.assertLess(res["median_e2e_latency_ms"], 12000) self.assertLess(res["median_e2e_latency_ms"], 11000)
self.assertLess(res["median_ttft_ms"], 86) self.assertLess(res["median_ttft_ms"], 86)
self.assertLess(res["median_itl_ms"], 10) self.assertLess(res["median_itl_ms"], 10)
...@@ -161,7 +161,7 @@ class TestBenchServing(unittest.TestCase): ...@@ -161,7 +161,7 @@ class TestBenchServing(unittest.TestCase):
f"### test_online_latency_eagle\n" f"### test_online_latency_eagle\n"
f'median_e2e_latency_ms : {res["median_e2e_latency_ms"]:.2f} ms\n' f'median_e2e_latency_ms : {res["median_e2e_latency_ms"]:.2f} ms\n'
) )
self.assertLess(res["median_e2e_latency_ms"], 10000) self.assertLess(res["median_e2e_latency_ms"], 450)
def test_moe_offline_throughput_default(self): def test_moe_offline_throughput_default(self):
res = run_bench_serving( res = run_bench_serving(
...@@ -176,7 +176,7 @@ class TestBenchServing(unittest.TestCase): ...@@ -176,7 +176,7 @@ class TestBenchServing(unittest.TestCase):
f"### test_moe_offline_throughput_default\n" f"### test_moe_offline_throughput_default\n"
f'Output throughput: {res["output_throughput"]:.2f} token/s\n' f'Output throughput: {res["output_throughput"]:.2f} token/s\n'
) )
self.assertGreater(res["output_throughput"], 2150) self.assertGreater(res["output_throughput"], 2200)
def test_moe_offline_throughput_without_radix_cache(self): def test_moe_offline_throughput_without_radix_cache(self):
res = run_bench_serving( res = run_bench_serving(
...@@ -191,7 +191,7 @@ class TestBenchServing(unittest.TestCase): ...@@ -191,7 +191,7 @@ class TestBenchServing(unittest.TestCase):
f"### test_moe_offline_throughput_without_radix_cache\n" f"### test_moe_offline_throughput_without_radix_cache\n"
f'Output throughput: {res["output_throughput"]:.2f} token/s\n' f'Output throughput: {res["output_throughput"]:.2f} token/s\n'
) )
self.assertGreater(res["output_throughput"], 2150) self.assertGreater(res["output_throughput"], 2200)
if __name__ == "__main__": if __name__ == "__main__":
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment