"tests/git@developer.sourcefind.cn:OpenDAS/mmcv.git" did not exist on "508a322fba81aaf9dd0c215ca8fd44a511a53da4"
Unverified Commit 32318178 authored by Lianmin Zheng's avatar Lianmin Zheng Committed by GitHub
Browse files

Revert "[feat] Add math eval to CI" (#2656)

parent a11f8d5f
...@@ -68,17 +68,6 @@ class TestEvalAccuracyLarge(unittest.TestCase): ...@@ -68,17 +68,6 @@ class TestEvalAccuracyLarge(unittest.TestCase):
metrics = run_eval(args) metrics = run_eval(args)
self.assertGreater(metrics["score"], 0.835) self.assertGreater(metrics["score"], 0.835)
def test_math(self):
args = SimpleNamespace(
base_url=self.base_url,
model=self.model,
eval_name="math",
num_examples=5000,
num_threads=1024
)
metrics = run_eval(args)
self.assertGreaterEqual(metrics["score"], 0.519 - 0.01) # -1% to account for sampling variance
if __name__ == "__main__": if __name__ == "__main__":
unittest.main() unittest.main()
...@@ -37,18 +37,6 @@ class TestEvalAccuracyMini(unittest.TestCase): ...@@ -37,18 +37,6 @@ class TestEvalAccuracyMini(unittest.TestCase):
metrics = run_eval(args) metrics = run_eval(args)
self.assertGreaterEqual(metrics["score"], 0.65) self.assertGreaterEqual(metrics["score"], 0.65)
def test_math(self):
args = SimpleNamespace(
base_url=self.base_url,
model=self.model,
eval_name="math",
num_examples=64,
num_threads=32,
temperature=0.1,
)
metrics = run_eval(args)
self.assertGreaterEqual(metrics["score"], 0.519 - 0.03) # -3% to account for sampling variance
if __name__ == "__main__": if __name__ == "__main__":
unittest.main() unittest.main()
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment