Commit 08fcd7e9 authored by Neelay Shah's avatar Neelay Shah Committed by GitHub
Browse files

refactor: move libs to lib dir


Signed-off-by: default avatarNeelay Shah <neelays@nvidia.com>
Co-authored-by: default avatarRyan McCormick <rmccormick@nvidia.com>
parent 0bfd9a76
...@@ -32,7 +32,7 @@ use super::{ ...@@ -32,7 +32,7 @@ use super::{
MAX_TOP_P, MIN_FREQUENCY_PENALTY, MIN_PRESENCE_PENALTY, MIN_TEMPERATURE, MIN_TOP_P, MAX_TOP_P, MIN_FREQUENCY_PENALTY, MIN_PRESENCE_PENALTY, MIN_TEMPERATURE, MIN_TOP_P,
}; };
use triton_distributed::protocols::annotated::AnnotationsProvider; use triton_distributed_runtime::protocols::annotated::AnnotationsProvider;
/// Legacy OpenAI CompletionRequest /// Legacy OpenAI CompletionRequest
/// ///
...@@ -176,7 +176,7 @@ pub struct CompletionRequest { ...@@ -176,7 +176,7 @@ pub struct CompletionRequest {
/// in the vocabulary of the model. /// in the vocabulary of the model.
/// ///
/// ```rust /// ```rust
/// use triton_llm::protocols::openai::completions::CompletionRequest; /// use triton_distributed_llm::protocols::openai::completions::CompletionRequest;
/// ///
/// let request = CompletionRequest::builder() /// let request = CompletionRequest::builder()
/// .prompt("What is the meaning of life?") /// .prompt("What is the meaning of life?")
...@@ -243,7 +243,7 @@ impl CompletionRequestBuilder { ...@@ -243,7 +243,7 @@ impl CompletionRequestBuilder {
/// Builds and validates the CompletionRequest /// Builds and validates the CompletionRequest
/// ///
/// ```rust /// ```rust
/// use triton_llm::protocols::openai::completions::CompletionRequest; /// use triton_distributed_llm::protocols::openai::completions::CompletionRequest;
/// ///
/// let request = CompletionRequest::builder() /// let request = CompletionRequest::builder()
/// .model("mixtral-8x7b-instruct-v0.1") /// .model("mixtral-8x7b-instruct-v0.1")
......
...@@ -19,7 +19,7 @@ pub use protocols::{Annotated, TokenIdType}; ...@@ -19,7 +19,7 @@ pub use protocols::{Annotated, TokenIdType};
pub mod openai { pub mod openai {
use super::*; use super::*;
use triton_distributed::pipeline::{ServerStreamingEngine, UnaryEngine}; use triton_distributed_runtime::pipeline::{ServerStreamingEngine, UnaryEngine};
pub mod completions { pub mod completions {
use super::*; use super::*;
......
...@@ -14,7 +14,7 @@ ...@@ -14,7 +14,7 @@
// limitations under the License. // limitations under the License.
use futures::StreamExt; use futures::StreamExt;
use triton_llm::protocols::{ use triton_distributed_llm::protocols::{
codec::{create_message_stream, Message, SseCodecError}, codec::{create_message_stream, Message, SseCodecError},
openai::{chat_completions::ChatCompletionResponse, completions::CompletionResponse}, openai::{chat_completions::ChatCompletionResponse, completions::CompletionResponse},
ContentProvider, DataStream, ContentProvider, DataStream,
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment