Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
1813bf04
"sgl-kernel/include/utils.h" did not exist on "0fe7c13be18d1edd8682747ce558b430a1aa1c9e"
Commit
1813bf04
authored
Aug 01, 2023
by
lintangsutawika
Browse files
skips metrics prep if process_result is not None
parent
0eb94c8b
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
16 additions
and
5 deletions
+16
-5
lm_eval/api/registry.py
lm_eval/api/registry.py
+9
-3
lm_eval/api/task.py
lm_eval/api/task.py
+7
-2
No files found.
lm_eval/api/registry.py
View file @
1813bf04
import
os
import
evaluate
from
lm_eval.api.model
import
LM
from
lm_eval.logger
import
eval_logger
MODEL_REGISTRY
=
{}
...
...
@@ -131,7 +132,7 @@ searching in HF Evaluate library..."
metric_object
=
evaluate
.
load
(
name
)
return
metric_object
.
compute
except
Exception
:
raise
Warning
(
eval_logger
.
error
(
"{} not found in the evaluate library!"
.
format
(
name
),
"Please check https://huggingface.co/evaluate-metric"
,
)
...
...
@@ -154,7 +155,7 @@ def get_aggregation(name):
try
:
return
AGGREGATION_REGISTRY
[
name
]
except
KeyError
:
raise
W
arning
(
eval_logger
.
w
arning
(
"{} not a registered aggregation metric!"
.
format
(
name
),
)
...
...
@@ -163,7 +164,9 @@ def get_default_aggregation(metric_name):
try
:
return
DEFAULT_AGGREGATION_REGISTRY
[
metric_name
]
except
KeyError
:
raise
Warning
(
f
"No default aggregation metric for metric '
{
metric_name
}
'!"
)
eval_logger
.
warning
(
f
"No default aggregation metric for metric '
{
metric_name
}
'!"
)
def
is_higher_better
(
metric_name
):
...
...
@@ -171,3 +174,6 @@ def is_higher_better(metric_name):
return
HIGHER_IS_BETTER_REGISTRY
[
metric_name
]
except
KeyError
:
raise
Warning
(
f
"higher_is_better not specified for metric '
{
metric_name
}
'!"
)
eval_logger
.
warning
(
f
"higher_is_better not specified for metric '
{
metric_name
}
'!"
)
lm_eval/api/task.py
View file @
1813bf04
...
...
@@ -554,8 +554,13 @@ class ConfigurableTask(Task):
for
key
in
metric_config
if
key
not
in
[
"metric"
,
"aggregation"
,
"higher_is_better"
]
}
self
.
_metric_fn_list
[
metric_name
]
=
get_metric
(
metric_name
)
self
.
_metric_fn_kwargs
[
metric_name
]
=
kwargs
if
self
.
_config
.
process_results
is
None
:
self
.
_metric_fn_list
[
metric_name
]
=
get_metric
(
metric_name
)
self
.
_metric_fn_kwargs
[
metric_name
]
=
kwargs
else
:
self
.
_metric_fn_list
[
metric_name
]
=
None
self
.
_metric_fn_kwargs
[
metric_name
]
=
{}
if
"aggregation"
in
metric_config
:
agg_name
=
metric_config
[
"aggregation"
]
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment