Commit b58e5556 authored by Baber's avatar Baber
Browse files

Merge branch 'main' into tasklist

# Conflicts:
#	pyproject.toml
parents 6e1866f5 4f8195f1
......@@ -28,5 +28,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -28,5 +28,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -28,5 +28,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -28,5 +28,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -28,5 +28,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -31,5 +31,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -28,5 +28,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -28,5 +28,3 @@ filter_list:
- function: take_first
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -19,5 +19,3 @@ metric_list:
higher_is_better: true
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -25,5 +25,3 @@ filter_list:
- function: "take_first"
metadata:
version: 0.0
dataset_kwargs:
trust_remote_code: true
......@@ -11,7 +11,9 @@ try:
except (ModuleNotFoundError, ImportError):
raise ModuleNotFoundError(
"Please install evaluation metrics via pip install evaluate and pip install bert-score",
"Please install evaluation metrics via pip install evaluate bert-score "
"rouge_score>=0.1.2 nltk absl-py "
"git+https://github.com/google-research/bleurt.git"
)
except Exception as e:
raise RuntimeError(
......
......@@ -23,5 +23,3 @@ metric_list:
higher_is_better: false
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -11,7 +11,9 @@ try:
except (ModuleNotFoundError, ImportError):
raise ModuleNotFoundError(
"Please install evaluation metrics via pip install evaluate and pip install bert-score",
"Please install evaluation metrics via pip install evaluate bert-score "
"rouge_score>=0.1.2 nltk absl-py "
"git+https://github.com/google-research/bleurt.git"
)
except Exception as e:
raise RuntimeError(
......
......@@ -11,7 +11,9 @@ try:
except (ModuleNotFoundError, ImportError):
raise ModuleNotFoundError(
"Please install evaluation metrics via pip install evaluate and pip install bert-score",
"Please install evaluation metrics via pip install evaluate bert-score "
"rouge_score>=0.1.2 nltk absl-py "
"git+https://github.com/google-research/bleurt.git"
)
except Exception as e:
raise RuntimeError(
......
......@@ -11,7 +11,9 @@ try:
except (ModuleNotFoundError, ImportError):
raise ModuleNotFoundError(
"Please install evaluation metrics via pip install evaluate and pip install bert-score",
"Please install evaluation metrics via pip install evaluate bert-score "
"rouge_score>=0.1.2 nltk absl-py "
"git+https://github.com/google-research/bleurt.git"
)
except Exception as e:
raise RuntimeError(
......
......@@ -15,7 +15,9 @@ try:
except (ModuleNotFoundError, ImportError):
raise ModuleNotFoundError(
"Please install evaluation metrics via pip install evaluate and pip install bert-score",
"Please install evaluation metrics via pip install evaluate bert-score "
"rouge_score>=0.1.2 nltk absl-py radgraph"
"git+https://github.com/google-research/bleurt.git"
)
except Exception as e:
raise RuntimeError(
......
......@@ -7,7 +7,7 @@ dataset_name: algebra
output_type: generate_until
training_split: train
test_split: test
doc_to_text: !function utils.doc_to_text
doc_to_text: !function utils.doc_to_text
process_results: !function utils.process_results
doc_to_target: "{{answer if few_shot is undefined else solution}}"
generation_kwargs:
......@@ -25,8 +25,6 @@ metric_list:
num_fewshot: 4
metadata:
version: 2.0
dataset_kwargs:
trust_remote_code: true
fewshot_config:
sampler: first_n
samples: !function utils.list_fewshot_samples
dataset_path: facebook/mlqa
dataset_kwargs:
trust_remote_code: true
test_split: test
validation_split: validation
output_type: generate_until
......
......@@ -9,5 +9,3 @@ doc_to_choice: "{{choices}}"
doc_to_target: "{{answer}}"
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
......@@ -13,5 +13,3 @@ metric_list:
higher_is_better: true
metadata:
version: 1.0
dataset_kwargs:
trust_remote_code: true
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment