• Graham King's avatar
    feat(dynamo-run): Batch mode (#142) · 2cca070c
    Graham King authored
    ```
    dynamo-run in=batch:prompts.jsonl out=mistralrs ~/llm_models/Llama-3.2-3B-Instruct/
    ```
    
    The file has genai format, one entry per line:
    ```
    {"text": "the prompt"}
    {"text": ..etc
    ```
    
    The prompt is evaluated and the output written to `output.jsonl` in the
    same folder as the input.
    
    At the end of the run various statistics are printed:
    > Ran 5 files in 8s 679ms. Tokens in: 40 (5/s). Tokens out: 346 (43/s)
    
    This is also helpful for pushing load into the system and stressing the
    various components. Not intended for performance measurement, it's a
    batch inference tool.
    2cca070c
README.md 15.1 KB