"library/vscode:/vscode.git/clone" did not exist on "7b1e2c379e0ddbeceec1f91f6d4d5f46cde29f7a"
Commit 6f4f9e1c authored by lintangsutawika's avatar lintangsutawika
Browse files

resolved merge conflict

parents 0d5748b7 aed90773
...@@ -16,5 +16,6 @@ generation_kwargs: ...@@ -16,5 +16,6 @@ generation_kwargs:
- "\n\n" - "\n\n"
do_sample: false do_sample: false
temperature: 0.0 temperature: 0.0
num_fewshot: 0
metadata: metadata:
- version: 0 - version: 0
...@@ -6,7 +6,7 @@ output_type: multiple_choice ...@@ -6,7 +6,7 @@ output_type: multiple_choice
training_split: train training_split: train
validation_split: validation validation_split: validation
test_split: test test_split: test
doc_to_text: "\nSentence 1: {{sentence1}}\nSentence 2: {{sentence2}}\nAnswer:" doc_to_text: "\nSentence 1: {{question1}}\nSentence 2: {{question2}}\nAnswer:"
doc_to_target: label doc_to_target: label
doc_to_choice: ["no", "yes"] doc_to_choice: ["no", "yes"]
metric_list: metric_list:
......
group: glue group: glue
task: sst task: sst2
dataset_path: glue dataset_path: glue
dataset_name: sst dataset_name: sst2
output_type: multiple_choice output_type: multiple_choice
training_split: train training_split: train
validation_split: validation validation_split: validation
......
# k_mmlu
### Paper
Title: `K-MMLU(work in progress) `
* ongoing project at publishing help/ non-english benchmark
Abstract: `The K-MMLU (Korean-MMLU) is a comprehensive suite designed to evaluate the advanced knowledge and reasoning abilities of large language models (LLMs) within the Korean language and cultural context. This suite encompasses 45 topics, primarily focusing on expert-level subjects. It includes general subjects like Physics and Ecology, and law and political science, alongside specialized fields such as Non-Destructive Training and Maritime Engineering. The datasets are derived from Korean licensing exams, with about 90% of the questions including human accuracy based on the performance of human test-takers in these exams. K-MMLU is segmented into training, testing, and development subsets, with the test subset ranging from a minimum of 100 to a maximum of 1000 questions, totaling 35,000 questions. Additionally, a set of 10 questions is provided as a development set for few-shot exemplar development. At total, K-MMLU consists of 254,334 instances.`
Homepage: https://huggingface.co/datasets/HAERAE-HUB/K-MMLU-Preview
### Citation
```
We'll be updating this section soon.
```
### Groups and Tasks
#### Groups
* `kmmlu`: 'All 45 subjects of the KMMLU dataset, evaluated following the methodology in MMLU's original implementation'
#### Tasks
The following tasks evaluate subjects in the KMMLU dataset
- `kmmlu_{subject_english}`
### Checklist
For adding novel benchmarks/datasets to the library:
* [x] Is the task an existing benchmark in the literature?
* [x] Have you referenced the original paper that introduced the task?
* [x] If yes, does the original paper provide a reference implementation? If so, have you checked against the reference implementation and documented how to run such a test?
If other tasks on this dataset are already supported:
* [ ] Is the "Main" variant of this task clearly denoted?
* [ ] Have you provided a short sentence in a README on what each new variant adds / evaluates?
* [ ] Have you noted which, if any, published evaluation setups are matched by this variant?
group: kmmlu
dataset_path: HAERAE-HUB/K-MMLU-Preview
output_type: multiple_choice
training_split: train
validation_split: dev
test_split: test
fewshot_split: dev
output_type: multiple_choice
process_docs: !function utils.process_docs
doc_to_text: "{{question}}"
doc_to_choice: "{{choices}}"
doc_to_target: "{{gold}}"
metric_list:
- metric: acc
aggregation: mean
higher_is_better: true
- metric: acc_norm
aggregation: mean
higher_is_better: true
metadata:
- version: 0.0
"dataset_name": "Accounting"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_accounting"
"dataset_name": "Agricultural Sciences"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_agricultural_sciences"
"dataset_name": "Aviation Engineering and Maintenance"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_aviation_engineering_and_maintenance"
"dataset_name": "Biology"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_biology"
"dataset_name": "Chemical Engineering"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_chemical_engineering"
"dataset_name": "Chemistry"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_chemistry"
"dataset_name": "Civil Engineering"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_civil_engineering"
"dataset_name": "Computer Science"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_computer_science"
"dataset_name": "Construction"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_construction"
"dataset_name": "Criminal Law"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_criminal_law"
"dataset_name": "Ecology"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_ecology"
"dataset_name": "Economics"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_economics"
"dataset_name": "Education"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_education"
"dataset_name": "Electrical Engineering"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_electrical_engineering"
"dataset_name": "Electronics Engineering"
"include": "_default_kmmlu_yaml"
"task": "kmmlu_electronics_engineering"
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment