"llm/llama.go" did not exist on "2bc06565c7ca7f784fefc0d2f037269b8af2a33d"
Unverified Commit 14043a0f authored by Stella Biderman's avatar Stella Biderman Committed by GitHub
Browse files

Update README.md

parent 0542d35d
...@@ -104,6 +104,12 @@ python write_out.py \ ...@@ -104,6 +104,12 @@ python write_out.py \
This will write out one text file for each task. This will write out one text file for each task.
## Advanced Usage
We support wildcards in task names, for example you can run all of the machine-translated lambada tasks via `--task lambada_openai_mt_*`.
We currently only support one prompt per task, which we strive to make the "standard" as defined by the benchmark's authors. If you would like to study how varying prompts causes changes in the evaluation score, check out the [BigScience fork](https://github.com/bigscience-workshop/lm-evaluation-harness) of this repo. We are currently working on upstreaming this capability to `main`.
## Implementing new tasks ## Implementing new tasks
To implement a new task in the eval harness, see [this guide](./docs/task_guide.md). To implement a new task in the eval harness, see [this guide](./docs/task_guide.md).
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment