Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
f4f8a1b4
Unverified
Commit
f4f8a1b4
authored
Oct 18, 2025
by
b8zhong
Committed by
GitHub
Oct 19, 2025
Browse files
ci: update `lmms-eval` to speed up multimodal CI (#11000)
parent
c44e985d
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
13 additions
and
5 deletions
+13
-5
scripts/ci/amd_ci_install_dependency.sh
scripts/ci/amd_ci_install_dependency.sh
+2
-2
scripts/ci/ci_install_dependency.sh
scripts/ci/ci_install_dependency.sh
+1
-1
test/srt/models/test_vlm_models.py
test/srt/models/test_vlm_models.py
+10
-2
No files found.
scripts/ci/amd_ci_install_dependency.sh
View file @
f4f8a1b4
...
...
@@ -22,7 +22,7 @@ case "${GPU_ARCH}" in
docker
exec
ci_sglang
rm
-rf
python/pyproject.toml
&&
mv
python/pyproject_other.toml python/pyproject.toml
docker
exec
ci_sglang pip
install
-e
"python[dev_hip]"
--no-deps
# TODO: only for mi35x
# For lmms_evals evaluating MMMU
docker
exec
-w
/ ci_sglang git clone
--branch
v0.
3.3
--depth
1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
docker
exec
-w
/ ci_sglang git clone
--branch
v0.
4.1
--depth
1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
docker
exec
-w
/lmms-eval ci_sglang pip
install
-e
.
--no-deps
# TODO: only for mi35x
;;
mi30x|mi300|mi325
)
...
...
@@ -30,7 +30,7 @@ case "${GPU_ARCH}" in
docker
exec
ci_sglang
rm
-rf
python/pyproject.toml
&&
mv
python/pyproject_other.toml python/pyproject.toml
docker
exec
ci_sglang pip
install
-e
"python[dev_hip]"
# For lmms_evals evaluating MMMU
docker
exec
-w
/ ci_sglang git clone
--branch
v0.
3.3
--depth
1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
docker
exec
-w
/ ci_sglang git clone
--branch
v0.
4.1
--depth
1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
docker
exec
-w
/lmms-eval ci_sglang pip
install
-e
.
;;
*
)
...
...
scripts/ci/ci_install_dependency.sh
View file @
f4f8a1b4
...
...
@@ -66,7 +66,7 @@ $PIP_CMD install mooncake-transfer-engine==0.3.6.post1 nvidia-cuda-nvrtc-cu12 py
if
[
"
$IS_BLACKWELL
"
!=
"1"
]
;
then
# For lmms_evals evaluating MMMU
git clone
--branch
v0.
3.3
--depth
1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
git clone
--branch
v0.
4.1
--depth
1 https://github.com/EvolvingLMMs-Lab/lmms-eval.git
$PIP_CMD
install
-e
lmms-eval/
$PIP_INSTALL_SUFFIX
# Install xformers
...
...
test/srt/models/test_vlm_models.py
View file @
f4f8a1b4
...
...
@@ -81,7 +81,7 @@ class TestVLMModels(CustomTestCase):
model
=
"openai_compatible"
tp
=
1
tasks
=
"mmmu_val"
batch_size
=
2
batch_size
=
3
2
log_suffix
=
"openai_compatible"
os
.
makedirs
(
output_path
,
exist_ok
=
True
)
...
...
@@ -179,7 +179,15 @@ class TestVLMModels(CustomTestCase):
self
.
run_mmmu_eval
(
model
.
model
,
output_path
)
# Get the result file
result_file_path
=
glob
.
glob
(
f
"
{
output_path
}
/*.json"
)[
0
]
# Search recursively for JSON result files (lmms-eval v0.4.1+ creates subdirectories)
result_files
=
glob
.
glob
(
f
"
{
output_path
}
/**/*.json"
,
recursive
=
True
)
if
not
result_files
:
result_files
=
glob
.
glob
(
f
"
{
output_path
}
/*.json"
)
if
not
result_files
:
raise
FileNotFoundError
(
f
"No JSON result files found in
{
output_path
}
"
)
result_file_path
=
result_files
[
0
]
with
open
(
result_file_path
,
"r"
)
as
f
:
result
=
json
.
load
(
f
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment