Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
b6c75ed1
Commit
b6c75ed1
authored
Jan 10, 2024
by
daniel-furman
Browse files
llama test
parent
021232be
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
0 additions
and
28 deletions
+0
-28
lm_eval/models/huggingface.py
lm_eval/models/huggingface.py
+0
-28
No files found.
lm_eval/models/huggingface.py
View file @
b6c75ed1
...
@@ -670,32 +670,6 @@ class HFLM(LM):
...
@@ -670,32 +670,6 @@ class HFLM(LM):
new_reqs
=
[]
new_reqs
=
[]
for
req
in
requests
:
for
req
in
requests
:
context
,
continuation
=
req
.
args
[
0
].
strip
(),
req
.
args
[
1
].
strip
()
context
,
continuation
=
req
.
args
[
0
].
strip
(),
req
.
args
[
1
].
strip
()
# arc experiment with few-shot formatting
"""import re
elements = re.split('Answer:|Question:', context.replace('
\n
', ' '))
new_elements = []
for element in elements[1:-1]:
new_elements.append(element.strip())
new_elements
chat = []
for i in range(len(new_elements)):
if i % 2 == 0:
chat.append({"role": "user", "content": f"Question: {new_elements[i]} Answer:"})
else:
chat.append({"role": "assistant", "content": f"{new_elements[i]}"})"""
chat
=
[
{
"role"
:
"system"
,
"content"
:
"You are a helpful, respectful and honest assistant."
},
{
"role"
:
"user"
,
"content"
:
context
},
{
"role"
:
"assistant"
,
"content"
:
continuation
},
]
context
=
self
.
tokenizer
.
apply_chat_template
(
chat
,
tokenize
=
False
,
add_generation_prompt
=
True
,
)
print
(
context
)
chat
=
[
chat
=
[
{
"role"
:
"system"
,
"content"
:
"You are a helpful, respectful and honest assistant."
},
{
"role"
:
"system"
,
"content"
:
"You are a helpful, respectful and honest assistant."
},
{
"role"
:
"user"
,
"content"
:
context
},
{
"role"
:
"user"
,
"content"
:
context
},
...
@@ -705,8 +679,6 @@ class HFLM(LM):
...
@@ -705,8 +679,6 @@ class HFLM(LM):
tokenize
=
False
,
tokenize
=
False
,
add_generation_prompt
=
True
,
add_generation_prompt
=
True
,
)
)
print
(
context
)
print
(
"
\n
"
)
req
.
args
=
(
context
,
continuation
)
req
.
args
=
(
context
,
continuation
)
new_reqs
.
append
(
req
)
new_reqs
.
append
(
req
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment