| [lingoly](lingoly/README.md) | Challenging logical reasoning benchmark in low-resource languages with controls for memorization | English, Multilingual |
| [logiqa](logiqa/README.md) | Logical reasoning tasks requiring advanced inference and deduction. | English, Chinese |
| [logiqa2](logiqa2/README.md) | Large-scale logical reasoning dataset adapted from the Chinese Civil Service Examination. | English, Chinese |
| [mastermind](mastermind/README.md) | Reasoning benchmark based on the board game of Mastermind. | English |
| [mathqa](mathqa/README.md) | Question answering tasks involving mathematical reasoning and problem-solving. | English |
| [mbpp](mbpp/README.md) | A benchmark designed to measure the ability to synthesize short Python programs from natural language descriptions. | Python |
| [meddialog](meddialog/README.md) | Medical open-ended QA and Question Entailment stemming from the MedDialog dataset. | English |
title={MastermindEval: A Simple But Scalable Reasoning Benchmark},
author={Jonas Golde and Patrick Haller and Fabio Barth and Alan Akbik},
booktitle={Workshop on Reasoning and Planning for Large Language Models},
year={2025},
url={https://openreview.net/forum?id=H4donosutm}
}
```
### Groups, Tags, and Tasks
#### Groups
None.
#### Tags
*`mastermind`: Evaluates all settings.
*`mastermind_easy`: Evaluates all easy settings (random wrong answer options).
*`mastermind_hard`: Evaluates all hard settings (wrong answer options differ in one symbol from the secret code).
#### Tasks
*`mastermind_24_easy`
*`mastermind_24_hard`
*`mastermind_35_easy`
*`mastermind_35_hard`
*`mastermind_46_easy`
*`mastermind_46_hard`
### Checklist
For adding novel benchmarks/datasets to the library:
* [x] Is the task an existing benchmark in the literature?
* [x] Have you referenced the original paper that introduced the task?
* [ ] If yes, does the original paper provide a reference implementation? If so, have you checked against the reference implementation and documented how to run such a test?
If other tasks on this dataset are already supported:
* [ ] Is the "Main" variant of this task clearly denoted?
* [ ] Have you provided a short sentence in a README on what each new variant adds / evaluates?
* [ ] Have you noted which, if any, published evaluation setups are matched by this variant?