Commit 021232be authored by daniel-furman's avatar daniel-furman
Browse files

llama test

parent 49f43f9f
...@@ -685,14 +685,28 @@ class HFLM(LM): ...@@ -685,14 +685,28 @@ class HFLM(LM):
else: else:
chat.append({"role": "assistant", "content": f"{new_elements[i]}"})""" chat.append({"role": "assistant", "content": f"{new_elements[i]}"})"""
chat = [ chat = [
#{"role": "system", "content": "You are a helpful, respectful and honest assistant."}, {"role": "system", "content": "You are a helpful, respectful and honest assistant."},
{"role": "user", "content": context}, {"role": "user", "content": context},
{"role": "assistant", "content": continuation},
] ]
context = self.tokenizer.apply_chat_template( context = self.tokenizer.apply_chat_template(
chat, chat,
tokenize=False, tokenize=False,
add_generation_prompt=True, add_generation_prompt=True,
) )
print(context)
chat = [
{"role": "system", "content": "You are a helpful, respectful and honest assistant."},
{"role": "user", "content": context},
]
context = self.tokenizer.apply_chat_template(
chat,
tokenize=False,
add_generation_prompt=True,
)
print(context)
print("\n")
req.args = (context, continuation) req.args = (context, continuation)
new_reqs.append(req) new_reqs.append(req)
...@@ -734,14 +748,6 @@ class HFLM(LM): ...@@ -734,14 +748,6 @@ class HFLM(LM):
stopping_criteria = stop_sequences_criteria( stopping_criteria = stop_sequences_criteria(
self.tokenizer, stop, 1, context.shape[0] self.tokenizer, stop, 1, context.shape[0]
) )
print(self.model.generate(
input_ids=context,
max_length=max_length,
stopping_criteria=stopping_criteria,
pad_token_id=self.tokenizer.pad_token_id,
use_cache=True,
**generation_kwargs,
))
return self.model.generate( return self.model.generate(
input_ids=context, input_ids=context,
max_length=max_length, max_length=max_length,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment