Commit 80fe11d8 authored by Benjamin Fattori's avatar Benjamin Fattori
Browse files

keep local changes

parent c23dd0fd
...@@ -6,13 +6,8 @@ import lm_eval.api.metrics ...@@ -6,13 +6,8 @@ import lm_eval.api.metrics
import lm_eval.models import lm_eval.models
import lm_eval.tasks import lm_eval.tasks
import lm_eval.api import lm_eval.api
<<<<<<< HEAD
from lm_eval.utils import positional_deprecated, run_task_tests, make_table, create_iterator from lm_eval.utils import positional_deprecated, run_task_tests, make_table, create_iterator
import torch import torch
=======
from lm_eval.utils import positional_deprecated, run_task_tests, make_table, get_git_commit_hash
>>>>>>> upstream/big-refactor
@positional_deprecated @positional_deprecated
def simple_evaluate( def simple_evaluate(
...@@ -84,7 +79,6 @@ def simple_evaluate( ...@@ -84,7 +79,6 @@ def simple_evaluate(
decontamination_ngrams_path=decontamination_ngrams_path, decontamination_ngrams_path=decontamination_ngrams_path,
) )
<<<<<<< HEAD
if lm.rank == 0: if lm.rank == 0:
# add info about the model and few shot config # add info about the model and few shot config
results["config"] = { results["config"] = {
...@@ -101,20 +95,6 @@ def simple_evaluate( ...@@ -101,20 +95,6 @@ def simple_evaluate(
return results return results
else: else:
return None return None
=======
# add info about the model and few shot config
results["config"] = {
"model": model,
"model_args": model_args,
"num_fewshot": num_fewshot,
"batch_size": batch_size,
"device": device,
"no_cache": no_cache,
"limit": limit,
"bootstrap_iters": bootstrap_iters,
}
results["git_hash"] = get_git_commit_hash()
>>>>>>> upstream/big-refactor
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment