Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
gaoqiong
lm-evaluation-harness
Commits
37a46351
Commit
37a46351
authored
Nov 28, 2023
by
lintangsutawika
Browse files
reformat
parent
525afc17
Changes
122
Hide whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
20 additions
and
20 deletions
+20
-20
lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu_flan_generative_template_yaml
...lan_n_shot/generative/_mmlu_flan_generative_template_yaml
+1
-1
lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu_flan_loglikelihood_template_yaml
...shot/loglikelihood/_mmlu_flan_loglikelihood_template_yaml
+1
-1
lm_eval/tasks/model_written_evals/advanced_ai_risk/_template_yaml
...tasks/model_written_evals/advanced_ai_risk/_template_yaml
+1
-1
lm_eval/tasks/model_written_evals/persona/_template_yaml
lm_eval/tasks/model_written_evals/persona/_template_yaml
+1
-1
lm_eval/tasks/model_written_evals/sycophancy/sycophancy_on_nlp_survey.yaml
...el_written_evals/sycophancy/sycophancy_on_nlp_survey.yaml
+1
-1
lm_eval/tasks/model_written_evals/sycophancy/sycophancy_on_philpapers2020.yaml
...ritten_evals/sycophancy/sycophancy_on_philpapers2020.yaml
+1
-1
lm_eval/tasks/model_written_evals/sycophancy/sycophancy_on_political_typology_quiz.yaml
...als/sycophancy/sycophancy_on_political_typology_quiz.yaml
+1
-1
lm_eval/tasks/model_written_evals/winogenerated/_template_yaml
...al/tasks/model_written_evals/winogenerated/_template_yaml
+1
-1
lm_eval/tasks/mutual/mutual.yaml
lm_eval/tasks/mutual/mutual.yaml
+1
-1
lm_eval/tasks/nq_open/nq_open.yaml
lm_eval/tasks/nq_open/nq_open.yaml
+1
-1
lm_eval/tasks/openbookqa/openbookqa.yaml
lm_eval/tasks/openbookqa/openbookqa.yaml
+1
-1
lm_eval/tasks/paws-x/pawsx_template_yaml
lm_eval/tasks/paws-x/pawsx_template_yaml
+1
-1
lm_eval/tasks/pile/pile_arxiv.yaml
lm_eval/tasks/pile/pile_arxiv.yaml
+1
-1
lm_eval/tasks/piqa/piqa.yaml
lm_eval/tasks/piqa/piqa.yaml
+1
-1
lm_eval/tasks/polemo2/polemo2_in.yaml
lm_eval/tasks/polemo2/polemo2_in.yaml
+1
-1
lm_eval/tasks/prost/corypaik_prost.yaml
lm_eval/tasks/prost/corypaik_prost.yaml
+1
-1
lm_eval/tasks/pubmedqa/pubmedqa.yaml
lm_eval/tasks/pubmedqa/pubmedqa.yaml
+1
-1
lm_eval/tasks/qa4mre/qa4mre_2011.yaml
lm_eval/tasks/qa4mre/qa4mre_2011.yaml
+1
-1
lm_eval/tasks/qasper/bool.yaml
lm_eval/tasks/qasper/bool.yaml
+1
-1
lm_eval/tasks/qasper/freeform.yaml
lm_eval/tasks/qasper/freeform.yaml
+1
-1
No files found.
lm_eval/tasks/mmlu/flan_n_shot/generative/_mmlu_flan_generative_template_yaml
View file @
37a46351
...
...
@@ -13,4 +13,4 @@ metric_list:
aggregation: mean
higher_is_better: true
metadata:
- version: 0.0
\ No newline at end of file
- version: 0.0
lm_eval/tasks/mmlu/flan_n_shot/loglikelihood/_mmlu_flan_loglikelihood_template_yaml
View file @
37a46351
...
...
@@ -13,4 +13,4 @@ metric_list:
aggregation: mean
higher_is_better: true
metadata:
- version: 0.0
\ No newline at end of file
- version: 0.0
lm_eval/tasks/model_written_evals/advanced_ai_risk/_template_yaml
View file @
37a46351
...
...
@@ -11,4 +11,4 @@ doc_to_decontamination_query: "{{sentence_good}} {{sentence_bad}}"
metric_list:
- metric: acc
metadata:
- version: 0.0
\ No newline at end of file
- version: 0.0
lm_eval/tasks/model_written_evals/persona/_template_yaml
View file @
37a46351
...
...
@@ -9,4 +9,4 @@ doc_to_choice: "{{[answer_matching_behavior, answer_not_matching_behavior]}}"
metric_list:
- metric: acc
metadata:
- version: 0.0
\ No newline at end of file
- version: 0.0
lm_eval/tasks/model_written_evals/sycophancy/sycophancy_on_nlp_survey.yaml
View file @
37a46351
...
...
@@ -11,4 +11,4 @@ doc_to_choice: "{{[answer_matching_behavior, answer_not_matching_behavior]}}"
metric_list
:
-
metric
:
acc
metadata
:
-
version
:
0.0
\ No newline at end of file
-
version
:
0.0
lm_eval/tasks/model_written_evals/sycophancy/sycophancy_on_philpapers2020.yaml
View file @
37a46351
...
...
@@ -11,4 +11,4 @@ doc_to_choice: "{{[answer_matching_behavior, answer_not_matching_behavior]}}"
metric_list
:
-
metric
:
acc
metadata
:
-
version
:
0.0
\ No newline at end of file
-
version
:
0.0
lm_eval/tasks/model_written_evals/sycophancy/sycophancy_on_political_typology_quiz.yaml
View file @
37a46351
...
...
@@ -11,4 +11,4 @@ doc_to_choice: "{{[answer_matching_behavior, answer_not_matching_behavior]}}"
metric_list
:
-
metric
:
acc
metadata
:
-
version
:
0.0
\ No newline at end of file
-
version
:
0.0
lm_eval/tasks/model_written_evals/winogenerated/_template_yaml
View file @
37a46351
...
...
@@ -9,4 +9,4 @@ doc_to_choice: "{{[answer_matching_behavior, answer_not_matching_behavior]}}"
metric_list:
- metric: acc
metadata:
- version: 0.0
\ No newline at end of file
- version: 0.0
lm_eval/tasks/mutual/mutual.yaml
View file @
37a46351
...
...
@@ -22,4 +22,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
2.0
\ No newline at end of file
-
version
:
2.0
lm_eval/tasks/nq_open/nq_open.yaml
View file @
37a46351
...
...
@@ -29,4 +29,4 @@ metric_list:
regexes_to_ignore
:
-
"
\b
an|a|the
\b
"
metadata
:
-
version
:
0.0
\ No newline at end of file
-
version
:
0.0
lm_eval/tasks/openbookqa/openbookqa.yaml
View file @
37a46351
...
...
@@ -18,4 +18,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
1.0
\ No newline at end of file
-
version
:
1.0
lm_eval/tasks/paws-x/pawsx_template_yaml
View file @
37a46351
...
...
@@ -17,4 +17,4 @@ metric_list:
aggregation: mean
higher_is_better: true
metadata:
- version: 0.0
\ No newline at end of file
- version: 0.0
lm_eval/tasks/pile/pile_arxiv.yaml
View file @
37a46351
...
...
@@ -20,4 +20,4 @@ metric_list:
aggregation
:
bits_per_byte
higher_is_better
:
false
metadata
:
-
version
:
2.0
\ No newline at end of file
-
version
:
2.0
lm_eval/tasks/piqa/piqa.yaml
View file @
37a46351
...
...
@@ -18,4 +18,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
1.0
\ No newline at end of file
-
version
:
1.0
lm_eval/tasks/polemo2/polemo2_in.yaml
View file @
37a46351
...
...
@@ -42,4 +42,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
0.0
\ No newline at end of file
-
version
:
0.0
lm_eval/tasks/prost/corypaik_prost.yaml
View file @
37a46351
...
...
@@ -16,4 +16,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
1.0
\ No newline at end of file
-
version
:
1.0
lm_eval/tasks/pubmedqa/pubmedqa.yaml
View file @
37a46351
...
...
@@ -13,4 +13,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
1.0
\ No newline at end of file
-
version
:
1.0
lm_eval/tasks/qa4mre/qa4mre_2011.yaml
View file @
37a46351
...
...
@@ -19,4 +19,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
1.0
\ No newline at end of file
-
version
:
1.0
lm_eval/tasks/qasper/bool.yaml
View file @
37a46351
...
...
@@ -11,4 +11,4 @@ doc_to_choice: ["no", "yes"]
metric_list
:
-
metric
:
f1
metadata
:
-
version
:
1.0
\ No newline at end of file
-
version
:
1.0
lm_eval/tasks/qasper/freeform.yaml
View file @
37a46351
...
...
@@ -15,4 +15,4 @@ metric_list:
aggregation
:
mean
higher_is_better
:
true
metadata
:
-
version
:
1.0
\ No newline at end of file
-
version
:
1.0
Prev
1
2
3
4
5
6
7
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment