// SPDX-FileCopyrightText: Copyright (c) 2024-2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. // SPDX-License-Identifier: Apache-2.0 // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. use serde::{Deserialize, Serialize}; use std::borrow::BorrowMut; use std::cmp::min; use uuid::Uuid; use crate::kv_router::indexer::OverlapScores; pub use crate::kv_router::protocols::{ForwardPassMetrics, KV_BLOCK_SIZE}; use crate::kv_router::scoring::ProcessedEndpoints; #[allow(dead_code)] #[derive(Debug, thiserror::Error)] pub enum KvSchedulerError { #[error("no endpoints aviailable to route work")] NoEndpoints, #[error("all workers busy")] AllWorkersBusy, #[error("endpoint subscriber shutdown")] SubscriberShutdown, } #[derive(Debug, Clone, Serialize, Deserialize)] pub struct Endpoint { pub name: String, pub subject: String, pub data: ForwardPassMetrics, } impl Endpoint { pub fn worker_id(&self) -> Uuid { Uuid::parse_str( self.subject .split(".") .last() .expect("invalid subject") .to_string() .as_str(), ) .expect("invalid uuid") } } #[derive(Debug, Clone, Serialize, Deserialize)] pub struct Service { pub name: String, pub id: String, pub version: String, pub started: String, pub endpoints: Vec, } pub struct SchedulingRequest { isl_tokens: usize, overlap: OverlapScores, resp_tx: tokio::sync::oneshot::Sender, } impl SchedulingRequest { pub fn respond(self, worker_id: String) { if self.resp_tx.send(worker_id).is_err() { tracing::trace!("failed to send response to requestor"); } } } pub struct KvScheduler { request_tx: tokio::sync::mpsc::Sender, } impl KvScheduler { pub async fn start( endpoints_rx: tokio::sync::mpsc::Receiver, ) -> Result { let mut endpoints_rx = endpoints_rx; tracing::trace!("awaiting the start of the background endpoint subscriber"); let mut endpoints = match endpoints_rx.recv().await { Some(endpoints) => endpoints, None => { return Err(KvSchedulerError::SubscriberShutdown); } }; // Channel to accept new scheduling requests let (request_tx, request_rx) = tokio::sync::mpsc::channel::(16); tracing::debug!("scheduler starting"); // Background task to handle scheduling requests tokio::spawn(async move { let mut request: SchedulingRequest; let mut request_rx = request_rx; tracing::debug!("scheduler background task started"); 'outer: loop { request = tokio::select! { biased; new_request = request_rx.recv() => { match new_request { Some(new_request) => { tracing::trace!("received request to be scheduled"); new_request }, None => { tracing::trace!("scheduler shutdown"); break 'outer; } } } new_endpoints = endpoints_rx.recv() => { match new_endpoints { Some(new_endpoints) => { tracing::trace!("updated endpoints"); endpoints = new_endpoints; continue 'outer; } None => { tracing::trace!("endpoint subscriber shutdown"); break 'outer; } } } }; tracing::debug!("selected"); loop { match select_worker(endpoints.borrow_mut(), &request) { Ok(worker_id) => { request.respond(worker_id); continue 'outer; } Err(KvSchedulerError::AllWorkersBusy) => { tracing::trace!("all workers busy; waiting for more capacity"); endpoints = match endpoints_rx.recv().await { Some(endpoints) => endpoints, None => { tracing::trace!("endpoint subscriber shutdown"); break 'outer; } }; } Err(e) => { tracing::error!("error scheduling request: {:?}", e); break 'outer; } } } } tracing::trace!("background endpoint subscriber shutting down"); }); Ok(KvScheduler { request_tx }) } #[allow(dead_code)] pub async fn schedule( &self, overlap: OverlapScores, isl_tokens: usize, ) -> Result { let (resp_tx, resp_rx) = tokio::sync::oneshot::channel(); let request = SchedulingRequest { isl_tokens, overlap, resp_tx, }; tracing::debug!("before sending request"); self.request_tx .send(request) .await .map_err(|_| KvSchedulerError::SubscriberShutdown)?; tracing::debug!("after sending request"); let res = resp_rx .await .map_err(|_| KvSchedulerError::SubscriberShutdown)?; tracing::debug!("after receiving response"); Ok(res) } } pub fn select_worker( workers: &mut ProcessedEndpoints, request: &SchedulingRequest, ) -> Result { // balance mode prioritizes balancing load across workers let balance_threshold: f64 = 0.1; let balance_mode = workers.load_std > balance_threshold * workers.load_avg; // Determine alpha based on mode let alpha = if balance_mode { 0.7 } else { 0.3 }; let gamma = 0.1; // example tuning param // Compute each worker's score let mut best_index = None; let mut best_cost = f64::INFINITY; if workers.endpoints.is_empty() { return Err(KvSchedulerError::NoEndpoints); } for (i, w) in workers.endpoints.iter().enumerate() { // Exclude workers that are at capacity if w.data.request_active_slots >= w.data.request_total_slots || w.data.kv_active_blocks >= w.data.kv_total_blocks { continue; } let kv_load_ratio = w.data.kv_active_blocks as f64 / w.data.kv_total_blocks as f64; let load_deviation = kv_load_ratio - workers.load_avg; let worker_id = workers.worker_ids[i]; let overlap_score = request.overlap.scores.get(&worker_id).map_or(0, |x| *x); let overlap_score = overlap_score as usize * KV_BLOCK_SIZE; let new_tokens = request.isl_tokens.saturating_sub(overlap_score); let normalized_new_tokens = new_tokens as f64 / request.isl_tokens as f64; let request_load_ratio = w.data.request_active_slots as f64 / w.data.request_total_slots as f64; // cost = alpha * load_deviation + (1 - alpha)*normalized_new_tokens + gamma * request_load_ratio let cost = alpha * load_deviation + (1.0 - alpha) * normalized_new_tokens + gamma * request_load_ratio; tracing::debug!("worker: {}; load_deviation: {}; normalized new blocks: {}; request_load_ratio: {} cost: {}", worker_id, load_deviation, normalized_new_tokens, request_load_ratio, cost ); if cost < best_cost { best_cost = cost; best_index = Some(i); } } if let Some(best_index) = best_index { let total_blocks = min(request.isl_tokens / KV_BLOCK_SIZE, 1); workers.endpoints[best_index].data.request_active_slots += 1; workers.endpoints[best_index].data.kv_active_blocks += total_blocks as u64; } match best_index { Some(i) => { tracing::info!( "selected worker: {}; cost: {}", workers.endpoints[i].subject, best_cost ); Ok(workers.endpoints[i].subject.clone()) } None => { tracing::debug!("all workers busy"); Err(KvSchedulerError::AllWorkersBusy) } } }