Unverified Commit 97ca18e0 authored by Leo Gao's avatar Leo Gao Committed by GitHub
Browse files

Update README.md

parent 50ac7dff
...@@ -21,7 +21,7 @@ pip install lm-eval ...@@ -21,7 +21,7 @@ pip install lm-eval
## Basic Usage ## Basic Usage
To evaluate a model, (e.g. GPT-2) on NLU tasks (e.g. LAMBADA, HellaSwag), you can run the following command. To evaluate a model, (e.g. GPT-2) on NLU tasks (e.g. LAMBADA, HellaSwag), you can run the following command. **When reporting results from eval harness, please include the task versions (shown in `results["versions"]`) for reproducibility.** This allows bug fixes to tasks while also ensuring that previously reported scores are reproducible.
```bash ```bash
python main.py \ python main.py \
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment