f"Tasks not found: {missing}. Try `lm-eval --tasks {{list_groups,list_subtasks,list_tags,list}}` to list out all available names for task groupings; only (sub)tasks; tags; or all of the above, or pass '--verbosity DEBUG' to troubleshoot task registration issues."
)
config.tasks=task_names
# Respect user's value passed in via CLI, otherwise default to True and add to comma-separated model args
f"The value of `model` passed to simple_evaluate() was of type {type(model)}, but is required to be a subclass of lm_eval.api.model.LM . This may be because you are passing an initialized Hugging Face PreTrainedModel without having wrapped it in `lm_eval.models.huggingface.HFLM(pretrained=my_model)` first."
f"The value of `model` passed to simple_evaluate() was of type {type(config.model)}, but is required to be a subclass of lm_eval.api.model.LM . This may be because you are passing an initialized Hugging Face PreTrainedModel without having wrapped it in `lm_eval.models.huggingface.HFLM(pretrained=my_model)` first."
)
eval_logger.info("Using pre-initialized model")
lm=model
lm=config.model
ifuse_cacheisnotNone:
eval_logger.info(f"Using cache at {use_cache+'_rank'+str(lm.rank)+'.db'}")
ifconfig.use_cacheisnotNone:
eval_logger.info(f"Using cache at {config.use_cache+'_rank'+str(lm.rank)+'.db'}")
lm=lm_eval.api.model.CachingLM(
lm,
use_cache
config.use_cache
# each rank receives a different cache db.
# necessary to avoid multiple writes to cache at once