router.py 4.59 KB
Newer Older
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
# SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
# SPDX-License-Identifier: Apache-2.0
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.


import asyncio
from argparse import Namespace
from enum import Enum
20
from typing import AsyncIterator
21
22

import uvloop
23
from common.protocol import Tokens
Neelay Shah's avatar
Neelay Shah committed
24
25
26
27
from vllm.logger import logger as vllm_logger

from triton_distributed.llm import KvRouter
from triton_distributed.runtime import (
28
29
30
31
32
    DistributedRuntime,
    triton_endpoint,
    triton_worker,
)

33
34
WorkerId = str

35
36
37
38
39
40
41
42
43
44
45
46
47
48

class RoutingStrategy(Enum):
    PREFIX = "prefix"
    ROUND_ROBIN = "round_robin"
    RANDOM = "random"


class Router:
    """
    Request handler for the generate endpoint
    """

    def __init__(
        self,
49
        router: KvRouter,
50
51
52
53
54
55
56
57
        routing_strategy: RoutingStrategy = RoutingStrategy.PREFIX,
    ):
        vllm_logger.info(
            f"Initializing KV Router with strategy: {routing_strategy.value}"
        )
        self.router = router
        self.routing_strategy = routing_strategy

58
59
    @triton_endpoint(Tokens, WorkerId)
    async def generate(self, request) -> AsyncIterator[WorkerId]:
60
61
62
63
64
65
66
67
68
69
70
71
72
73
        lora_id = 0
        worker_id = ""
        if self.routing_strategy == RoutingStrategy.PREFIX:
            try:
                worker_id = await self.router.schedule(request.tokens, lora_id)
            except Exception as e:
                vllm_logger.info(f"{e}")
                if "No worker found" in str(e):
                    worker_id = ""
                else:
                    vllm_logger.exception(f"Error during worker selection: {e}")

            vllm_logger.info(f"Scheduling to worker_id: {worker_id}")

74
75
            yield worker_id

76
        else:
77
78
79
80
            # TODO: Do we implement round_robin and random here?
            # or just skip this router and directly enable in preprocess?
            raise NotImplementedError(
                f"Routing strategy {self.routing_strategy} not implemented"
81
82
83
84
85
            )


@triton_worker()
async def worker(runtime: DistributedRuntime, args: Namespace):
86
87
88
89
    """
    Set up the worker clients.
    Serve the triton-init.router.generate endpoint.
    """
90
91
92
    workers_client = (
        await runtime.namespace("triton-init")
        .component("vllm")
93
        .endpoint("generate")
94
95
        .client()
    )
96
97
98
99
100
101
102
103
    wait_task = workers_client.wait_for_endpoints()
    await asyncio.sleep(1)

    while not wait_task.done():
        vllm_logger.info("Waiting for workers to be ready...")
        await asyncio.sleep(5)

    wait_task.result()
104
105
106
107
108
109
110
111
112
113
114
115
116

    while len(workers_client.endpoint_ids()) < args.min_workers:
        vllm_logger.info(
            f"Waiting for more workers... Current: {len(workers_client.endpoint_ids())}, Required: {args.min_workers}"
        )
        await asyncio.sleep(5)

    vllm_logger.info(
        f"Required number of workers ({args.min_workers}) are ready:\n"
        + "\n".join(f"id: {id}" for id in workers_client.endpoint_ids())
    )

    # TODO Router is a fixed namespace separate from the others
117
    kv_listener = runtime.namespace("router").component(args.model_name)
118
119
120
121
122
    await kv_listener.create_service()

    router_component = runtime.namespace("triton-init").component("router")
    await router_component.create_service()

123
    router = KvRouter(runtime, kv_listener)
124
125

    endpoint = router_component.endpoint("generate")
126
    await endpoint.serve_endpoint(Router(router, args.routing_strategy).generate)
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147


if __name__ == "__main__":
    uvloop.install()

    import argparse

    parser = argparse.ArgumentParser()
    parser.add_argument(
        "--routing-strategy",
        type=RoutingStrategy,
        default=RoutingStrategy.PREFIX,
        choices=list(RoutingStrategy),
        help="Routing strategy to use",
    )
    parser.add_argument(
        "--min-workers",
        type=int,
        default=1,
        help="Minimum number of workers required before proceeding",
    )
148
149
150
151
152
153
    parser.add_argument(
        "--model-name",
        type=str,
        default="deepseek-ai/DeepSeek-R1-Distill-Llama-8B",
        help="Model that is being served",
    )
154
155
156
    args = parser.parse_args()

    asyncio.run(worker(args))