"Generate completions. This is similar to the `/v1/completions` in OpenAI API. Detailed parameters can be found in the [sampling parameters](https://docs.sglang.ai/backend/sampling_params.html)."
"Generate completions. This is similar to the `/v1/completions` in OpenAI API. Detailed parameters can be found in the [sampling parameters](./sampling_params.md)."
This doc describes the sampling parameters of the SGLang Runtime.
This doc describes the sampling parameters of the SGLang Runtime. It is the low-level endpoint of the runtime.
It is the low-level endpoint of the runtime.
If you want a high-level endpoint that can automatically handle chat templates, consider using the [OpenAI Compatible API](https://docs.sglang.ai/backend/openai_api_completions.html).
If you want a high-level endpoint that can automatically handle chat templates, consider using the [OpenAI Compatible API](./openai_api_completions.ipynb).
## `/generate` Endpoint
## `/generate` Endpoint
The `/generate` endpoint accepts the following parameters in JSON format. For in detail usage see the [native api doc](https://docs.sglang.ai/backend/native_api.html).
The `/generate` endpoint accepts the following parameters in JSON format. For in detail usage see the [native api doc](./native_api.ipynb).
*`prompt`: The input prompt. Can be a single prompt or a batch of prompts.`Optional[Union[List[str], str]] = None`
*`prompt: Optional[Union[List[str], str]] = None` The input prompt. Can be a single prompt or a batch of prompts.
*`input_ids`: Alternative to `text`. Specify the input as token IDs instead of text.`Optional[Union[List[List[int]], List[int]]] = None`
*`input_ids: Optional[Union[List[List[int]], List[int]]] = None` Alternative to `text`. Specify the input as token IDs instead of text.
*`sampling_params`: The sampling parameters as described in the sections below.`Optional[Union[List[Dict], Dict]] = None`
*`sampling_params: Optional[Union[List[Dict], Dict]] = None` The sampling parameters as described in the sections below.
*`return_logprob`: Whether to return log probabilities for tokens.`Optional[Union[List[bool], bool]] = None`
*`return_logprob: Optional[Union[List[bool], bool]] = None` Whether to return log probabilities for tokens.
*`logprob_start_len`: If returning log probabilities, specifies the start position in the prompt. Default is "-1" which returns logprobs only for output tokens.`Optional[Union[List[int], int]] = None`
*`logprob_start_len: Optional[Union[List[int], int]] = None` If returning log probabilities, specifies the start position in the prompt. Default is "-1" which returns logprobs only for output tokens.
*`top_logprobs_num`: If returning log probabilities, specifies the number of top logprobs to return at each position.`Optional[Union[List[int], int]] = None`
*`top_logprobs_num: Optional[Union[List[int], int]] = None` If returning log probabilities, specifies the number of top logprobs to return at each position.
*`stream`: Whether to stream the output.`bool = False`
*`stream: bool = False` Whether to stream the output.
*`lora_path`: Path to LoRA weights. `Optional[Union[List[Optional[str]], Optional[str]]] = None`
*`lora_path: Optional[Union[List[Optional[str]], Optional[str]]] = None` Path to LoRA weights.
*`custom_logit_processor`: Custom logit processor for advanced sampling control. For usage see below.`Optional[Union[List[Optional[str]], str]] = None`
*`custom_logit_processor: Optional[Union[List[Optional[str]], str]] = None` Custom logit processor for advanced sampling control. For usage see below.
*`return_hidden_states`: Whether to return hidden states of the model. Note that each time it changes, the cuda graph will be recaptured, which might lead to a performance hit. See the [examples](https://github.com/sgl-project/sglang/blob/main/examples/runtime/engine/hidden_states.py) for more information.`bool = False`
*`return_hidden_states: bool = False` Whether to return hidden states of the model. Note that each time it changes, the cuda graph will be recaptured, which might lead to a performance hit. See the [examples](https://github.com/sgl-project/sglang/blob/main/examples/runtime/engine/hidden_states.py) for more information.
## Sampling params
## Sampling params
### Core Parameters
### Core Parameters
*`max_new_tokens`: The maximum output length measured in tokens.`int = 128`
*`max_new_tokens: int = 128` The maximum output length measured in tokens.
*`stop`: One or multiple [stop words](https://platform.openai.com/docs/api-reference/chat/create#chat-create-stop). Generation will stop if one of these words is sampled.`Optional[Union[str, List[str]]] = None`
*`stop: Optional[Union[str, List[str]]] = None` One or multiple [stop words](https://platform.openai.com/docs/api-reference/chat/create#chat-create-stop). Generation will stop if one of these words is sampled.
*`stop_token_ids`: Provide stop words in form of token ids. Generation will stop if one of these token ids is sampled.`Optional[List[int]] = []`
*`stop_token_ids: Optional[List[int]] = []` Provide stop words in form of token ids. Generation will stop if one of these token ids is sampled.
*`temperature`: [Temperature](https://platform.openai.com/docs/api-reference/chat/create#chat-create-temperature) when sampling the next token. `temperature = 0` corresponds to greedy sampling, higher temperature leads to more diversity.`float = 1.0`
*`temperature: float = 1.0`[Temperature](https://platform.openai.com/docs/api-reference/chat/create#chat-create-temperature) when sampling the next token. `temperature = 0` corresponds to greedy sampling, higher temperature leads to more diversity.
*`top_p`: [Top-p](https://platform.openai.com/docs/api-reference/chat/create#chat-create-top_p) selects tokens from the smallest sorted set whose cumulative probability exceeds `top_p`. When `top_p = 1`, this reduces to unrestricted sampling from all tokens.`top_p: float = 1.0`
*`top_p: float = 1.0`[Top-p](https://platform.openai.com/docs/api-reference/chat/create#chat-create-top_p) selects tokens from the smallest sorted set whose cumulative probability exceeds `top_p`. When `top_p = 1`, this reduces to unrestricted sampling from all tokens.
*`top_k`: [Top-k](https://developer.nvidia.com/blog/how-to-get-better-outputs-from-your-large-language-model/#predictability_vs_creativity) randomly selects from the `k` highest-probability tokens.`int = -1`
*`top_k: int = -1`[Top-k](https://developer.nvidia.com/blog/how-to-get-better-outputs-from-your-large-language-model/#predictability_vs_creativity) randomly selects from the `k` highest-probability tokens.
*`min_p`: [Min-p](https://github.com/huggingface/transformers/issues/27670) samples from tokens with probability larger than `min_p * highest_token_probability`.`float = 0.0`
*`min_p: float = 0.0`[Min-p](https://github.com/huggingface/transformers/issues/27670) samples from tokens with probability larger than `min_p * highest_token_probability`.
### Penalizers
### Penalizers
To use penalizers you will need to `--disable-overlap`. Please note that this might degrade performance.
To use penalizers you will need to `--disable-overlap`. Please note that this might degrade performance.
*`frequency_penalty`: Penalizes tokens based on their frequency in generation so far. Must be between `-2` and `2` where negative numbers encourage repeatment of tokens and positive number encourages sampling of new tokens. The scaling of penalization grows linearly with each appearance of a token.`float = 0.0`
*`frequency_penalty: float = 0.0`: Penalizes tokens based on their frequency in generation so far. Must be between `-2` and `2` where negative numbers encourage repeatment of tokens and positive number encourages sampling of new tokens. The scaling of penalization grows linearly with each appearance of a token.
*`presence_penalty`: Penalizes tokens if they appeared in the generation so far. Must be between `-2` and `2` where negative numbers encourage repeatment of tokens and positive number encourages sampling of new tokens. The scaling of the penalization is constant if a token occured.`float = 0.0`
*`presence_penalty: float = 0.0`: Penalizes tokens if they appeared in the generation so far. Must be between `-2` and `2` where negative numbers encourage repeatment of tokens and positive number encourages sampling of new tokens. The scaling of the penalization is constant if a token occured.
*`repetition_penalty`: Penalizes tokens if they appeared in prompt or generation so far. Must be between `0` and `2` where numbers smaller than `1` encourage repeatment of tokens and numbers larger than `2` encourages sampling of new tokens. The penalization scales multiplicatively.`float = 0.0`
*`repetition_penalty: float = 0.0`: Penalizes tokens if they appeared in prompt or generation so far. Must be between `0` and `2` where numbers smaller than `1` encourage repeatment of tokens and numbers larger than `2` encourages sampling of new tokens. The penalization scales multiplicatively.
*`min_new_tokens`: Forces the model to generate at least `min_new_tokens` until a stop word or EOS token is sampled. Note that this might lead to unintended behavior for example if the distribution is highly skewed towards these tokens.`int = 0`
*`min_new_tokens: int = 0`: Forces the model to generate at least `min_new_tokens` until a stop word or EOS token is sampled. Note that this might lead to unintended behavior for example if the distribution is highly skewed towards these tokens.
### Constrained decoding
### Constrained decoding
Please refer to our dedicated guide on [constrained decoding](https://docs.sglang.ai/backend/structured_outputs.html#Native-API-and-SGLang-Runtime-(SRT)) for the following parameters.
Please refer to our dedicated guide on [constrained decoding](./structured_outputs.ipynb) for the following parameters.
*`json_schema`: `Optional[str] = None`
*`json_schema: Optional[str] = None`: JSON schema for structured outputs.
*`regex`: `Optional[str] = None`
*`regex: Optional[str] = None`: Regex for structured outputs.
*`ebnf`: `Optional[str] = None`
*`ebnf: Optional[str] = None`: EBNF for structured outputs.
### Other options
### Other options
*`n`: Specifies the number of output sequences to generate per request. (Generating multiple outputs in one request (n > 1) is discouraged; repeat the same prompts for several times offer better control and efficiency.)`int = 1`
*`n: int = 1`: Specifies the number of output sequences to generate per request. (Generating multiple outputs in one request (n > 1) is discouraged; repeat the same prompts for several times offer better control and efficiency.)
*`spaces_between_special_tokens`: Whether or not to add spaces between special tokens during detokenization.`bool = True`
*`spaces_between_special_tokens: bool = True`: Whether or not to add spaces between special tokens during detokenization.
*`no_stop_trim`: Don't trim stop words or EOS token from the generated text.`bool = False`
*`no_stop_trim: bool = False`: Don't trim stop words or EOS token from the generated text.
*`ignore_eos`: Don't stop generation when EOS token is sampled.`bool = False`
*`ignore_eos: bool = False`: Don't stop generation when EOS token is sampled.
*`skip_special_tokens`: Remove special tokens during decoding.`bool = True`
*`skip_special_tokens: bool = True`: Remove special tokens during decoding.
*`custom_params`: Used when employing `CustomLogitProcessor`. For usage see below.`Optional[List[Optional[Dict[str, Any]]]] = None`
*`custom_params: Optional[List[Optional[Dict[str, Any]]]] = None`: Used when employing `CustomLogitProcessor`. For usage see below.
### Custom Logit Processor
## Examples
Launch a server with `--enable-custom-logit-processor` flag on.
"text":"<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n"
"<|im_start|>user\n<image>\nDescribe this image in a very short sentence.<|im_end|>\n"
"<|im_start|>assistant\n",
"image_data":"example_image.png",
"sampling_params":{
"temperature":0,
"max_new_tokens":32,
},
},
)
print(response.json())
```
The `image_data` can be a file name, a URL, or a base64 encoded string. See also `python/sglang/srt/utils.py:load_image`.
Streaming is supported in a similar manner as [above](#streaming).
Detailed example in [openai api vision](./openai_api_vision.ipynb).
### Structured Outputs (JSON, Regex, EBNF)
You can specify a JSON schema, regular expression or [EBNF](https://en.wikipedia.org/wiki/Extended_Backus%E2%80%93Naur_form) to constrain the model output. The model output will be guaranteed to follow the given constraints. Only one constraint parameter (`json_schema`, `regex`, or `ebnf`) can be specified for a request.
SGLang supports two grammar backends:
-[Outlines](https://github.com/dottxt-ai/outlines)(default): Supports JSON schema and regular expression constraints.
-[XGrammar](https://github.com/mlc-ai/xgrammar): Supports JSON schema, regular expression, and EBNF constraints.
- XGrammar currently uses the [GGML BNF format](https://github.com/ggerganov/llama.cpp/blob/master/grammars/README.md)
Initialize the XGrammar backend using `--grammar-backend xgrammar` flag
- To enable multi-GPU data parallelism, add `--dp 2`. Data parallelism is better for throughput if there is enough memory. It can also be used together with tensor parallelism. The following command uses 4 GPUs in total. We recommend [SGLang Router](https://docs.sglang.ai/router/router.html) for data parallelism.
- To enable multi-GPU data parallelism, add `--dp 2`. Data parallelism is better for throughput if there is enough memory. It can also be used together with tensor parallelism. The following command uses 4 GPUs in total. We recommend [SGLang Router](../router/router.md) for data parallelism.
@@ -49,8 +49,8 @@ Please consult the documentation below to learn more about the parameters you ma
...
@@ -49,8 +49,8 @@ Please consult the documentation below to learn more about the parameters you ma
*`kv_cache_dtype`: Dtype of the kv cache, defaults to the `dtype`.
*`kv_cache_dtype`: Dtype of the kv cache, defaults to the `dtype`.
*`context_length`: The number of tokens our model can process *including the input*. Note that extending the default might lead to strange behavior.
*`context_length`: The number of tokens our model can process *including the input*. Note that extending the default might lead to strange behavior.
*`device`: The device we put the model, defaults to `cuda`.
*`device`: The device we put the model, defaults to `cuda`.
*`chat_template`: The chat template to use. Deviating from the default might lead to unexpected responses. For multi-modal chat templates, refer to [here](https://docs.sglang.ai/backend/openai_api_vision.html#Chat-Template).
*`chat_template`: The chat template to use. Deviating from the default might lead to unexpected responses. For multi-modal chat templates, refer to [here](https://docs.sglang.ai/backend/openai_api_vision.ipynb#Chat-Template).
*`is_embedding`: Set to true to perform [embedding](https://docs.sglang.ai/backend/openai_api_embeddings.html) / [encode](https://docs.sglang.ai/backend/native_api.html#Encode-(embedding-model)) and [reward](https://docs.sglang.ai/backend/native_api.html#Classify-(reward-model)) tasks.
*`is_embedding`: Set to true to perform [embedding](./openai_api_embeddings.ipynb) / [encode](https://docs.sglang.ai/backend/native_api#Encode-(embedding-model)) and [reward](https://docs.sglang.ai/backend/native_api#Classify-(reward-model)) tasks.
*`revision`: Adjust if a specific version of the model should be used.
*`revision`: Adjust if a specific version of the model should be used.
*`skip_tokenizer_init`: Set to true to provide the tokens to the engine and get the output tokens directly, typically used in RLHF. Please see this [example for reference](https://github.com/sgl-project/sglang/blob/main/examples/runtime/engine/token_in_token_out_llm.py).
*`skip_tokenizer_init`: Set to true to provide the tokens to the engine and get the output tokens directly, typically used in RLHF. Please see this [example for reference](https://github.com/sgl-project/sglang/blob/main/examples/runtime/engine/token_in_token_out_llm.py).
*`json_model_override_args`: Override model config with the provided JSON.
*`json_model_override_args`: Override model config with the provided JSON.
...
@@ -79,7 +79,7 @@ Please consult the documentation below to learn more about the parameters you ma
...
@@ -79,7 +79,7 @@ Please consult the documentation below to learn more about the parameters you ma
### Data parallelism
### Data parallelism
*`dp_size`: Will be deprecated. The number of data-parallel copies of the model. [SGLang router](https://docs.sglang.ai/router/router.html) is recommended instead of the current naive data parallel.
*`dp_size`: Will be deprecated. The number of data-parallel copies of the model. [SGLang router](../router/router.md) is recommended instead of the current naive data parallel.
*`load_balance_method`: Will be deprecated. Load balancing strategy for data parallel requests.
*`load_balance_method`: Will be deprecated. Load balancing strategy for data parallel requests.
@@ -39,7 +39,7 @@ Again, please go through the entire documentation to confirm your system is usin
...
@@ -39,7 +39,7 @@ Again, please go through the entire documentation to confirm your system is usin
## Installing SGLang
## Installing SGLang
For general installation instructions, see the official [SGLang Installation Docs](https://docs.sglang.ai/start/install.html). Below are the AMD-specific steps summarized for convenience.
For general installation instructions, see the official [SGLang Installation Docs](../start/install.md). Below are the AMD-specific steps summarized for convenience.
The quantization and limited context length (`--dtype half --context-length 8192`) are due to the limited computational resources in [Nvidia jetson kit](https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-orin/). A detailed explanation can be found in [Server Arguments](https://docs.sglang.ai/backend/server_arguments.html).
The quantization and limited context length (`--dtype half --context-length 8192`) are due to the limited computational resources in [Nvidia jetson kit](https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-orin/). A detailed explanation can be found in [Server Arguments](../backend/server_arguments.md).
After launching the engine, refer to [Chat completions](https://docs.sglang.ai/backend/openai_api_completions.html#Usage) to test the usability.
After launching the engine, refer to [Chat completions](https://docs.sglang.ai/backend/openai_api_completions.html#Usage) to test the usability.
* * * * *
* * * * *
...
@@ -66,7 +66,7 @@ This enables TorchAO's int4 weight-only quantization with a 128-group size. The
...
@@ -66,7 +66,7 @@ This enables TorchAO's int4 weight-only quantization with a 128-group size. The
* * * * *
* * * * *
Structured output with XGrammar
Structured output with XGrammar
-------------------------------
-------------------------------
Please refer to [SGLang doc structured output](https://docs.sglang.ai/backend/structured_outputs.html).
Please refer to [SGLang doc structured output](../backend/structured_outputs.ipynb).
* * * * *
* * * * *
Thanks to the support from [shahizat](https://github.com/shahizat).
Thanks to the support from [shahizat](https://github.com/shahizat).