Unverified Commit f4f8a1b4 authored by b8zhong's avatar b8zhong Committed by GitHub
Browse files

ci: update `lmms-eval` to speed up multimodal CI (#11000)

parent c44e985d
...@@ -22,7 +22,7 @@ case "${GPU_ARCH}" in ...@@ -22,7 +22,7 @@ case "${GPU_ARCH}" in
docker exec ci_sglang rm -rf python/pyproject.toml && mv python/pyproject_other.toml python/pyproject.toml docker exec ci_sglang rm -rf python/pyproject.toml && mv python/pyproject_other.toml python/pyproject.toml
docker exec ci_sglang pip install -e "python[dev_hip]" --no-deps # TODO: only for mi35x docker exec ci_sglang pip install -e "python[dev_hip]" --no-deps # TODO: only for mi35x
# For lmms_evals evaluating MMMU # For lmms_evals evaluating MMMU
docker exec -w / ci_sglang git clone --branch v0.3.3 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git docker exec -w / ci_sglang git clone --branch v0.4.1 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
docker exec -w /lmms-eval ci_sglang pip install -e . --no-deps # TODO: only for mi35x docker exec -w /lmms-eval ci_sglang pip install -e . --no-deps # TODO: only for mi35x
;; ;;
mi30x|mi300|mi325) mi30x|mi300|mi325)
...@@ -30,7 +30,7 @@ case "${GPU_ARCH}" in ...@@ -30,7 +30,7 @@ case "${GPU_ARCH}" in
docker exec ci_sglang rm -rf python/pyproject.toml && mv python/pyproject_other.toml python/pyproject.toml docker exec ci_sglang rm -rf python/pyproject.toml && mv python/pyproject_other.toml python/pyproject.toml
docker exec ci_sglang pip install -e "python[dev_hip]" docker exec ci_sglang pip install -e "python[dev_hip]"
# For lmms_evals evaluating MMMU # For lmms_evals evaluating MMMU
docker exec -w / ci_sglang git clone --branch v0.3.3 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git docker exec -w / ci_sglang git clone --branch v0.4.1 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
docker exec -w /lmms-eval ci_sglang pip install -e . docker exec -w /lmms-eval ci_sglang pip install -e .
;; ;;
*) *)
......
...@@ -66,7 +66,7 @@ $PIP_CMD install mooncake-transfer-engine==0.3.6.post1 nvidia-cuda-nvrtc-cu12 py ...@@ -66,7 +66,7 @@ $PIP_CMD install mooncake-transfer-engine==0.3.6.post1 nvidia-cuda-nvrtc-cu12 py
if [ "$IS_BLACKWELL" != "1" ]; then if [ "$IS_BLACKWELL" != "1" ]; then
# For lmms_evals evaluating MMMU # For lmms_evals evaluating MMMU
git clone --branch v0.3.3 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git git clone --branch v0.4.1 --depth 1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
$PIP_CMD install -e lmms-eval/ $PIP_INSTALL_SUFFIX $PIP_CMD install -e lmms-eval/ $PIP_INSTALL_SUFFIX
# Install xformers # Install xformers
......
...@@ -81,7 +81,7 @@ class TestVLMModels(CustomTestCase): ...@@ -81,7 +81,7 @@ class TestVLMModels(CustomTestCase):
model = "openai_compatible" model = "openai_compatible"
tp = 1 tp = 1
tasks = "mmmu_val" tasks = "mmmu_val"
batch_size = 2 batch_size = 32
log_suffix = "openai_compatible" log_suffix = "openai_compatible"
os.makedirs(output_path, exist_ok=True) os.makedirs(output_path, exist_ok=True)
...@@ -179,7 +179,15 @@ class TestVLMModels(CustomTestCase): ...@@ -179,7 +179,15 @@ class TestVLMModels(CustomTestCase):
self.run_mmmu_eval(model.model, output_path) self.run_mmmu_eval(model.model, output_path)
# Get the result file # Get the result file
result_file_path = glob.glob(f"{output_path}/*.json")[0] # Search recursively for JSON result files (lmms-eval v0.4.1+ creates subdirectories)
result_files = glob.glob(f"{output_path}/**/*.json", recursive=True)
if not result_files:
result_files = glob.glob(f"{output_path}/*.json")
if not result_files:
raise FileNotFoundError(f"No JSON result files found in {output_path}")
result_file_path = result_files[0]
with open(result_file_path, "r") as f: with open(result_file_path, "r") as f:
result = json.load(f) result = json.load(f)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment