Commit d859d1ca authored by Nathan Habib's avatar Nathan Habib
Browse files

batch commit

parent 6e49b1f6
from typing import List from typing import List
import pytest import pytest
import torch
from lm_eval import tasks from lm_eval import tasks
from lm_eval.api.instance import Instance from lm_eval.api.instance import Instance
...@@ -10,7 +11,7 @@ task_manager = tasks.TaskManager() ...@@ -10,7 +11,7 @@ task_manager = tasks.TaskManager()
@pytest.mark.skip(reason="requires CUDA") @pytest.mark.skip(reason="requires CUDA")
class Test_VLLM: class TEST_VLLM:
vllm = pytest.importorskip("vllm") vllm = pytest.importorskip("vllm")
try: try:
from lm_eval.models.vllm_causallms import VLLM from lm_eval.models.vllm_causallms import VLLM
...@@ -18,7 +19,7 @@ class Test_VLLM: ...@@ -18,7 +19,7 @@ class Test_VLLM:
LM = VLLM(pretrained="EleutherAI/pythia-70m") LM = VLLM(pretrained="EleutherAI/pythia-70m")
except ModuleNotFoundError: except ModuleNotFoundError:
pass pass
# torch.use_deterministic_algorithms(True) torch.use_deterministic_algorithms(True)
task_list = task_manager.load_task_or_group(["arc_easy", "gsm8k", "wikitext"]) task_list = task_manager.load_task_or_group(["arc_easy", "gsm8k", "wikitext"])
multiple_choice_task = task_list["arc_easy"] # type: ignore multiple_choice_task = task_list["arc_easy"] # type: ignore
multiple_choice_task.build_all_requests(limit=10, rank=0, world_size=1) multiple_choice_task.build_all_requests(limit=10, rank=0, world_size=1)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment