Unverified Commit 704f8e8e authored by Chayenne's avatar Chayenne Committed by GitHub
Browse files

Add Reward API Docs etc (#1910)


Co-authored-by: default avatarChayenne <zhaochenyang@g.ucla.edu>
parent 1853c352
...@@ -48,9 +48,10 @@ jobs: ...@@ -48,9 +48,10 @@ jobs:
make html make html
cd _build/html cd _build/html
git clone https://$GITHUB_TOKEN@github.com/sgl-project/sgl-project.github.io.git ../sgl-project.github.io git clone https://$GITHUB_TOKEN@github.com/sgl-project/sgl-project.github.io.git ../sgl-project.github.io --depth 1
rm -rf ../sgl-project.github.io/* rm -rf ../sgl-project.github.io/*
cp -r * ../sgl-project.github.io cp -r * ../sgl-project.github.io
cp ../../README.md ../sgl-project.github.io/README.md
cd ../sgl-project.github.io cd ../sgl-project.github.io
git config user.name "zhaochenyang20" git config user.name "zhaochenyang20"
git config user.email "zhaochenyang20@gmail.com" git config user.email "zhaochenyang20@gmail.com"
......
...@@ -36,11 +36,13 @@ The core features include: ...@@ -36,11 +36,13 @@ The core features include:
- **Fast Backend Runtime**: Provides efficient serving with RadixAttention for prefix caching, jump-forward constrained decoding, continuous batching, token attention (paged attention), tensor parallelism, FlashInfer kernels, chunked prefill, and quantization (INT4/FP8/AWQ/GPTQ). - **Fast Backend Runtime**: Provides efficient serving with RadixAttention for prefix caching, jump-forward constrained decoding, continuous batching, token attention (paged attention), tensor parallelism, FlashInfer kernels, chunked prefill, and quantization (INT4/FP8/AWQ/GPTQ).
- **Flexible Frontend Language**: Offers an intuitive interface for programming LLM applications, including chained generation calls, advanced prompting, control flow, multi-modal inputs, parallelism, and external interactions. - **Flexible Frontend Language**: Offers an intuitive interface for programming LLM applications, including chained generation calls, advanced prompting, control flow, multi-modal inputs, parallelism, and external interactions.
- **Extensive Model Support**: Supports a wide range of generative models (Llama, Gemma, Mistral, QWen, DeepSeek, LLaVA, etc.) and embedding models (e5-mistral), with easy extensibility for integrating new models. - **Extensive Model Support**: Supports a wide range of generative models (Llama, Gemma, Mistral, QWen, DeepSeek, LLaVA, etc.), embedding models (e5-mistral, gte) and reward models (Skywork), with easy extensibility for integrating new models.
- **Active Community**: SGLang is open-source and backed by an active community with industry adoption. - **Active Community**: SGLang is open-source and backed by an active community with industry adoption.
## Install ## Getting Started
See [https://sgl-project.github.io/start/install.html](https://sgl-project.github.io/start/install.html) Install SGLang: See [https://sgl-project.github.io/start/install.html](https://sgl-project.github.io/start/install.html)
Send requests: See [https://sgl-project.github.io/start/send_request.html](https://sgl-project.github.io/start/send_request.html)
## Backend: SGLang Runtime (SRT) ## Backend: SGLang Runtime (SRT)
See [https://sgl-project.github.io/backend/backend.html](https://sgl-project.github.io/backend/backend.html) See [https://sgl-project.github.io/backend/backend.html](https://sgl-project.github.io/backend/backend.html)
......
...@@ -5,9 +5,10 @@ ...@@ -5,9 +5,10 @@
"metadata": {}, "metadata": {},
"source": [ "source": [
"# Native APIs\n", "# Native APIs\n",
"\n",
"Apart from the OpenAI compatible APIs, the SGLang Runtime also provides its native server APIs. We introduce these following APIs:\n", "Apart from the OpenAI compatible APIs, the SGLang Runtime also provides its native server APIs. We introduce these following APIs:\n",
"\n", "\n",
"- `/generate`\n", "- `/generate` (text generation model)\n",
"- `/get_server_args`\n", "- `/get_server_args`\n",
"- `/get_model_info`\n", "- `/get_model_info`\n",
"- `/health`\n", "- `/health`\n",
...@@ -15,7 +16,8 @@ ...@@ -15,7 +16,8 @@
"- `/flush_cache`\n", "- `/flush_cache`\n",
"- `/get_memory_pool_size`\n", "- `/get_memory_pool_size`\n",
"- `/update_weights`\n", "- `/update_weights`\n",
"- `/encode`\n", "- `/encode`(embedding model)\n",
"- `/judge`(reward model)\n",
"\n", "\n",
"We mainly use `requests` to test these APIs in the following examples. You can also use `curl`." "We mainly use `requests` to test these APIs in the following examples. You can also use `curl`."
] ]
...@@ -40,6 +42,8 @@ ...@@ -40,6 +42,8 @@
" print_highlight,\n", " print_highlight,\n",
")\n", ")\n",
"\n", "\n",
"import requests\n",
"\n",
"server_process = execute_shell_command(\n", "server_process = execute_shell_command(\n",
" \"\"\"\n", " \"\"\"\n",
"python3 -m sglang.launch_server --model-path meta-llama/Llama-3.2-1B-Instruct --port=30010\n", "python3 -m sglang.launch_server --model-path meta-llama/Llama-3.2-1B-Instruct --port=30010\n",
...@@ -53,7 +57,7 @@ ...@@ -53,7 +57,7 @@
"cell_type": "markdown", "cell_type": "markdown",
"metadata": {}, "metadata": {},
"source": [ "source": [
"## Generate\n", "## Generate (text generation model)\n",
"Generate completions. This is similar to the `/v1/completions` in OpenAI API. Detailed parameters can be found in the [sampling parameters](../references/sampling_params.html)." "Generate completions. This is similar to the `/v1/completions` in OpenAI API. Detailed parameters can be found in the [sampling parameters](../references/sampling_params.html)."
] ]
}, },
...@@ -63,8 +67,6 @@ ...@@ -63,8 +67,6 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"import requests\n",
"\n",
"url = \"http://localhost:30010/generate\"\n", "url = \"http://localhost:30010/generate\"\n",
"data = {\"text\": \"What is the capital of France?\"}\n", "data = {\"text\": \"What is the capital of France?\"}\n",
"\n", "\n",
...@@ -252,7 +254,7 @@ ...@@ -252,7 +254,7 @@
"cell_type": "markdown", "cell_type": "markdown",
"metadata": {}, "metadata": {},
"source": [ "source": [
"## Encode\n", "## Encode (embedding model)\n",
"\n", "\n",
"Encode text into embeddings. Note that this API is only available for [embedding models](openai_api_embeddings.html#openai-apis-embedding) and will raise an error for generation models.\n", "Encode text into embeddings. Note that this API is only available for [embedding models](openai_api_embeddings.html#openai-apis-embedding) and will raise an error for generation models.\n",
"Therefore, we launch a new server to server an embedding model." "Therefore, we launch a new server to server an embedding model."
...@@ -292,13 +294,76 @@ ...@@ -292,13 +294,76 @@
"print_highlight(f\"Text embedding (first 10): {response_json['embedding'][:10]}\")" "print_highlight(f\"Text embedding (first 10): {response_json['embedding'][:10]}\")"
] ]
}, },
{
"cell_type": "markdown",
"metadata": {},
"source": [
"## Judge (reward model)\n",
"\n",
"SGLang Runtime also supports reward models. Here we use a reward model to judge the quality of pairwise generations."
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"terminate_process(embedding_process)\n",
"\n",
"# Note that SGLang now treats embedding models and reward models as the same type of models.\n",
"# This will be updated in the future.\n",
"\n",
"reward_process = execute_shell_command(\n",
" \"\"\"\n",
"python -m sglang.launch_server --model-path Skywork/Skywork-Reward-Llama-3.1-8B-v0.2 --port 30030 --host 0.0.0.0 --is-embedding\n",
"\"\"\"\n",
")\n",
"\n",
"wait_for_server(\"http://localhost:30030\")"
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"from transformers import AutoTokenizer\n",
"\n",
"PROMPT = (\n",
" \"What is the range of the numeric output of a sigmoid node in a neural network?\"\n",
")\n",
"\n",
"RESPONSE1 = \"The output of a sigmoid node is bounded between -1 and 1.\"\n",
"RESPONSE2 = \"The output of a sigmoid node is bounded between 0 and 1.\"\n",
"\n",
"CONVS = [\n",
" [{\"role\": \"user\", \"content\": PROMPT}, {\"role\": \"assistant\", \"content\": RESPONSE1}],\n",
" [{\"role\": \"user\", \"content\": PROMPT}, {\"role\": \"assistant\", \"content\": RESPONSE2}],\n",
"]\n",
"\n",
"tokenizer = AutoTokenizer.from_pretrained(\"Skywork/Skywork-Reward-Llama-3.1-8B-v0.2\")\n",
"prompts = tokenizer.apply_chat_template(CONVS, tokenize=False)\n",
"\n",
"url = \"http://localhost:30030/judge\"\n",
"data = {\n",
" \"model\": \"Skywork/Skywork-Reward-Llama-3.1-8B-v0.2\", \n",
" \"text\": prompts\n",
"}\n",
"\n",
"responses = requests.post(url, json=data).json()\n",
"for response in responses:\n",
" print_highlight(f\"reward: {response['embedding'][0]}\")"
]
},
{ {
"cell_type": "code", "cell_type": "code",
"execution_count": 43, "execution_count": 15,
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"terminate_process(embedding_process)" "terminate_process(reward_process)"
] ]
} }
], ],
......
...@@ -7,7 +7,7 @@ The core features include: ...@@ -7,7 +7,7 @@ The core features include:
- **Fast Backend Runtime**: Provides efficient serving with RadixAttention for prefix caching, jump-forward constrained decoding, continuous batching, token attention (paged attention), tensor parallelism, FlashInfer kernels, chunked prefill, and quantization (INT4/FP8/AWQ/GPTQ). - **Fast Backend Runtime**: Provides efficient serving with RadixAttention for prefix caching, jump-forward constrained decoding, continuous batching, token attention (paged attention), tensor parallelism, FlashInfer kernels, chunked prefill, and quantization (INT4/FP8/AWQ/GPTQ).
- **Flexible Frontend Language**: Offers an intuitive interface for programming LLM applications, including chained generation calls, advanced prompting, control flow, multi-modal inputs, parallelism, and external interactions. - **Flexible Frontend Language**: Offers an intuitive interface for programming LLM applications, including chained generation calls, advanced prompting, control flow, multi-modal inputs, parallelism, and external interactions.
- **Extensive Model Support**: Supports a wide range of generative models (Llama 3, Gemma 2, Mistral, QWen, DeepSeek, LLaVA, etc.) and embedding models (e5-mistral), with easy extensibility for integrating new models. - **Extensive Model Support**: Supports a wide range of generative models (Llama, Gemma, Mistral, QWen, DeepSeek, LLaVA, etc.), embedding models (e5-mistral, gte) and reward models (Skywork), with easy extensibility for integrating new models.
- **Active Community**: SGLang is open-source and backed by an active community with industry adoption. - **Active Community**: SGLang is open-source and backed by an active community with industry adoption.
......
...@@ -5,14 +5,19 @@ ...@@ -5,14 +5,19 @@
"metadata": {}, "metadata": {},
"source": [ "source": [
"# Quick Start: Sending Requests\n", "# Quick Start: Sending Requests\n",
"This notebook provides a quick-start guide for using SGLang after installation." "This notebook provides a quick-start guide to use SGLang in chat completions after installation.\n",
"\n",
"- For Vision Language Models, see [OpenAI APIs - Vision](../backend/openai_api_vision.ipynb).\n",
"- For Embedding Models, see [OpenAI APIs - Embedding](../backend/openai_api_embeddings.ipynb) and [Encode (embedding model)](../backend/native_api.html#Encode-(embedding-model)).\n",
"- For Reward Models, see [Judge (reward model)](../backend/native_api.html#Judge-(reward-model))."
] ]
}, },
{ {
"cell_type": "markdown", "cell_type": "markdown",
"metadata": {}, "metadata": {},
"source": [ "source": [
"## Launch a server\n", "## Launch A Server\n",
"\n",
"This code block is equivalent to executing \n", "This code block is equivalent to executing \n",
"\n", "\n",
"```bash\n", "```bash\n",
...@@ -254,7 +259,7 @@ ...@@ -254,7 +259,7 @@
}, },
{ {
"cell_type": "code", "cell_type": "code",
"execution_count": 6, "execution_count": 8,
"metadata": { "metadata": {
"execution": { "execution": {
"iopub.execute_input": "2024-11-01T02:46:52.898411Z", "iopub.execute_input": "2024-11-01T02:46:52.898411Z",
......
...@@ -22,6 +22,7 @@ from sglang.test.runners import HFRunner, SRTRunner ...@@ -22,6 +22,7 @@ from sglang.test.runners import HFRunner, SRTRunner
MODELS = [ MODELS = [
("LxzGordon/URM-LLaMa-3.1-8B", 1, 3e-2), ("LxzGordon/URM-LLaMa-3.1-8B", 1, 3e-2),
("Skywork/Skywork-Reward-Llama-3.1-8B-v0.2", 1, 3e-2),
] ]
TORCH_DTYPES = [torch.float16] TORCH_DTYPES = [torch.float16]
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment