> TensorRT-LLM Support is currently available on a [branch](https://github.com/ai-dynamo/dynamo/tree/dynamo/trtllm_llmapi_v1/examples/trtllm#building-the-environment)
### Development Environment
For a consistent development environment, you can use the provided devcontainer configuration. This requires:
This allows quick iteration on the engine setup. Note how the `-n``1` is included. Flags `--leader-addr` and `--model-config` will also be added if provided to `dynamo-run`.
#### TensorRT-LLM `pystr` engine
To run a TRT-LLM model with dynamo-run we have included a python based [async engine] (/examples/tensorrt_llm/engines/agg_engine.py).
To configure the TensorRT-LLM async engine please see [llm_api_config.yaml](/examples/tensorrt_llm/configs/llm_api_config.yaml). The file defines the options that need to be passed to the LLM engine. Follow the steps below to serve trtllm on dynamo run.
##### Step 1: Build the environment
See instructions [here](/examples/tensorrt_llm/README.md#build-docker) to build the dynamo container with TensorRT-LLM.
##### Step 2: Run the environment
See instructions [here](/examples/tensorrt_llm/README.md#run-container) to run the built environment.
##### Step 3: Execute `dynamo run` command
Execute the following to load the TensorRT-LLM model specified in the configuration.
```
dynamo run out=pystr:/workspace/examples/tensorrt_llm/engines/agg_engine.py -- --engine_args /workspace/examples/tensorrt_llm/configs/llm_api_config.yaml
```
#### Dynamo does the pre-processing
If the Python engine wants to receive and return tokens - the prompt templating and tokenization is already done - run it like this: