Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
ca8b00cc
Commit
ca8b00cc
authored
Nov 02, 2023
by
lintangsutawika
Browse files
very rough way of utilzing verbosity
parent
ae74b808
Changes
4
Hide whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
33 additions
and
18 deletions
+33
-18
lm_eval/__init__.py
lm_eval/__init__.py
+2
-1
lm_eval/__main__.py
lm_eval/__main__.py
+23
-11
lm_eval/evaluator.py
lm_eval/evaluator.py
+1
-4
lm_eval/tasks/__init__.py
lm_eval/tasks/__init__.py
+7
-2
No files found.
lm_eval/__init__.py
View file @
ca8b00cc
from
.evaluator
import
evaluate
,
simple_evaluate
# from .evaluator import evaluate, simple_evaluate
# from .logger import eval_logger, SPACING
\ No newline at end of file
lm_eval/__main__.py
View file @
ca8b00cc
import
os
import
os
import
re
import
re
import
json
import
json
import
fnmatch
import
argparse
import
logging
import
logging
from
pathlib
import
Path
import
argparse
import
numpy
as
np
import
numpy
as
np
from
lm_eval
import
evaluator
,
utils
from
lm_eval.api.registry
import
ALL_TASKS
from
lm_eval.logger
import
eval_logger
,
SPACING
from
lm_eval.tasks
import
include_path
from
pathlib
import
Path
from
typing
import
Union
from
typing
import
Union
import
logging
logging
.
basicConfig
(
format
=
"%(asctime)s,%(msecs)03d %(levelname)-8s [%(filename)s:%(lineno)d] %(message)s"
,
datefmt
=
"%Y-%m-%d:%H:%M:%S"
,
level
=
logging
.
INFO
,
)
eval_logger
=
logging
.
getLogger
(
"lm-eval"
)
SPACING
=
" "
*
47
def
_handle_non_serializable
(
o
):
def
_handle_non_serializable
(
o
):
if
isinstance
(
o
,
np
.
int64
)
or
isinstance
(
o
,
np
.
int32
):
if
isinstance
(
o
,
np
.
int64
)
or
isinstance
(
o
,
np
.
int32
):
...
@@ -29,7 +35,7 @@ def parse_eval_args() -> argparse.Namespace:
...
@@ -29,7 +35,7 @@ def parse_eval_args() -> argparse.Namespace:
parser
.
add_argument
(
parser
.
add_argument
(
"--tasks"
,
"--tasks"
,
default
=
None
,
default
=
None
,
help
=
"Available Tasks:
\n
- {}"
.
format
(
"
\n
- "
.
join
(
sorted
(
ALL_TASKS
))),
#
help="Available Tasks:\n - {}".format("\n - ".join(sorted(ALL_TASKS))),
)
)
parser
.
add_argument
(
parser
.
add_argument
(
"--model_args"
,
"--model_args"
,
...
@@ -115,13 +121,19 @@ def parse_eval_args() -> argparse.Namespace:
...
@@ -115,13 +121,19 @@ def parse_eval_args() -> argparse.Namespace:
def
cli_evaluate
(
args
:
Union
[
argparse
.
Namespace
,
None
]
=
None
)
->
None
:
def
cli_evaluate
(
args
:
Union
[
argparse
.
Namespace
,
None
]
=
None
)
->
None
:
if
not
args
:
# if not args:
# we allow for args to be passed externally, else we parse them ourselves
# # we allow for args to be passed externally, else we parse them ourselves
args
=
parse_eval_args
()
# from lm_eval.logger import eval_logger, SPACING
args
=
parse_eval_args
()
eval_logger
.
setLevel
(
getattr
(
logging
,
f
"
{
args
.
verbosity
}
"
))
eval_logger
.
setLevel
(
getattr
(
logging
,
f
"
{
args
.
verbosity
}
"
))
os
.
environ
[
"TOKENIZERS_PARALLELISM"
]
=
"false"
os
.
environ
[
"TOKENIZERS_PARALLELISM"
]
=
"false"
from
lm_eval
import
evaluator
,
utils
from
lm_eval.api.registry
import
ALL_TASKS
from
lm_eval.tasks
import
include_path
if
args
.
limit
:
if
args
.
limit
:
eval_logger
.
warning
(
eval_logger
.
warning
(
" --limit SHOULD ONLY BE USED FOR TESTING."
" --limit SHOULD ONLY BE USED FOR TESTING."
...
...
lm_eval/evaluator.py
View file @
ca8b00cc
...
@@ -25,10 +25,6 @@ from lm_eval.utils import (
...
@@ -25,10 +25,6 @@ from lm_eval.utils import (
from
lm_eval.logger
import
eval_logger
from
lm_eval.logger
import
eval_logger
logger
=
logging
.
getLogger
(
__name__
)
logger
.
setLevel
(
logging
.
INFO
)
logger
.
addHandler
(
logging
.
StreamHandler
(
sys
.
stdout
))
@
positional_deprecated
@
positional_deprecated
def
simple_evaluate
(
def
simple_evaluate
(
...
@@ -46,6 +42,7 @@ def simple_evaluate(
...
@@ -46,6 +42,7 @@ def simple_evaluate(
decontamination_ngrams_path
=
None
,
decontamination_ngrams_path
=
None
,
write_out
:
bool
=
False
,
write_out
:
bool
=
False
,
log_samples
:
bool
=
True
,
log_samples
:
bool
=
True
,
verbosity
:
str
=
"INFO"
,
):
):
"""Instantiate and evaluate a model on a list of tasks.
"""Instantiate and evaluate a model on a list of tasks.
...
...
lm_eval/tasks/__init__.py
View file @
ca8b00cc
...
@@ -16,7 +16,9 @@ from lm_eval.api.registry import (
...
@@ -16,7 +16,9 @@ from lm_eval.api.registry import (
import
logging
import
logging
eval_logger
=
logging
.
getLogger
(
"lm-eval"
)
eval_logger
=
logging
.
getLogger
(
"lm-eval"
)
# from lm_eval.logger import eval_logger
# print("tasks.py eval_logger.level")
print
(
eval_logger
.
level
)
def
register_configurable_task
(
config
:
Dict
[
str
,
str
])
->
int
:
def
register_configurable_task
(
config
:
Dict
[
str
,
str
])
->
int
:
SubClass
=
type
(
SubClass
=
type
(
...
@@ -141,8 +143,11 @@ def include_task_folder(task_dir: str, register_task: bool = True) -> None:
...
@@ -141,8 +143,11 @@ def include_task_folder(task_dir: str, register_task: bool = True) -> None:
else
:
else
:
if
type
(
config
[
"task"
])
==
list
:
if
type
(
config
[
"task"
])
==
list
:
register_configurable_group
(
config
,
yaml_path
)
register_configurable_group
(
config
,
yaml_path
)
# Log this silently and show it only when
# the user defines the appropriate verbosity.
except
ModuleNotFoundError
as
e
:
except
ModuleNotFoundError
as
e
:
eval_logger
.
warnin
g
(
eval_logger
.
debu
g
(
f
"
{
yaml_path
}
:
{
e
}
. Config will not be added to registry."
f
"
{
yaml_path
}
:
{
e
}
. Config will not be added to registry."
)
)
except
Exception
as
error
:
except
Exception
as
error
:
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment