feat(dynamo-run): Batch mode (#142)
```
dynamo-run in=batch:prompts.jsonl out=mistralrs ~/llm_models/Llama-3.2-3B-Instruct/
```
The file has genai format, one entry per line:
```
{"text": "the prompt"}
{"text": ..etc
```
The prompt is evaluated and the output written to `output.jsonl` in the
same folder as the input.
At the end of the run various statistics are printed:
> Ran 5 files in 8s 679ms. Tokens in: 40 (5/s). Tokens out: 346 (43/s)
This is also helpful for pushing load into the system and stressing the
various components. Not intended for performance measurement, it's a
batch inference tool.
Showing
Please register or sign in to comment