Commit 87dff8b0 authored by daniel-furman's avatar daniel-furman
Browse files

first stab at wrap_chat_template, print statements in loglikelihood for testing

parent 3e27f9dd
......@@ -664,6 +664,9 @@ class HFLM(LM):
return self.tokenizer.decode(tokens, skip_special_tokens=True)
def tok_wrap_chat_template(self, requests: List[Instance], system: bool = False) -> List[Instance]:
"""
Utility for adding chat templates via the apply_chat_template() method
"""
new_reqs = []
for req in requests:
context, continuation = req.args[0].strip(), req.args[1].strip()
......@@ -774,12 +777,9 @@ class HFLM(LM):
return context_enc, continuation_enc
def loglikelihood(self, requests: List[Instance]) -> List[Tuple[float, bool]]:
#print(requests)
print(requests[0])
print(requests[0].args)
print(f"First element before prompt formatting...\n{requests[0].args}")
requests = self.tok_wrap_chat_template(requests)
print(requests[0])
print(requests[0].args)
print(f"First element after prompt formatting...\n{requests[0].args}")
new_reqs = []
for context, continuation in [req.args for req in requests]:
if context == "":
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment