/// HTTP Server logic use crate::config::Config; use crate::infer::tool_grammar::ToolGrammar; use crate::infer::{Backend, Infer, InferError, InferResponse, InferStreamResponse}; #[cfg(feature = "kserve")] use crate::kserve::{ kerve_server_metadata, kserve_health_live, kserve_health_ready, kserve_model_infer, kserve_model_metadata, kserve_model_metadata_ready, }; use crate::validation::ValidationError; use crate::ChatTokenizeResponse; use crate::{ usage_stats, BestOfSequence, Details, ErrorResponse, FinishReason, FunctionName, GenerateParameters, GenerateRequest, GenerateResponse, GrammarType, HubModelInfo, HubProcessorConfig, HubTokenizerConfig, Info, Message, MessageChunk, MessageContent, OutputMessage, PrefillToken, SimpleToken, StreamDetails, StreamResponse, TextMessage, Token, TokenizeResponse, ToolCallDelta, ToolCallMessage, Url, Usage, Validation, }; use crate::{ ChatCompletion, ChatCompletionChoice, ChatCompletionChunk, ChatCompletionComplete, ChatCompletionDelta, ChatCompletionLogprob, ChatCompletionLogprobs, ChatCompletionTopLogprob, ChatRequest, Chunk, CompatGenerateRequest, Completion, CompletionComplete, CompletionFinal, CompletionRequest, CompletionType, DeltaToolCall, Function, Prompt, Tool, VertexRequest, VertexResponse, }; use crate::{FunctionDefinition, HubPreprocessorConfig, ToolCall, ToolChoice, ToolType, Tools}; use async_stream::__private::AsyncStream; use axum::extract::Extension; use axum::http::{HeaderMap, HeaderValue, Method, StatusCode}; use axum::response::sse::{Event, KeepAlive, Sse}; use axum::response::{IntoResponse, Response}; use axum::routing::{get, post}; use axum::{http, Json, Router}; use axum_tracing_opentelemetry::middleware::OtelAxumLayer; use futures::stream::StreamExt; use futures::stream::{FuturesOrdered, FuturesUnordered}; use futures::Stream; use futures::TryStreamExt; use hf_hub::api::tokio::{Api, ApiBuilder, ApiRepo}; use hf_hub::{Cache, Repo, RepoType}; use http::header::AUTHORIZATION; use metrics_exporter_prometheus::{Matcher, PrometheusBuilder, PrometheusHandle}; use serde_json::Value; use std::convert::Infallible; use std::fs::File; use std::io::BufReader; use std::net::{IpAddr, Ipv4Addr, SocketAddr}; use std::path::{Path, PathBuf}; use thiserror::Error; use tokenizers::processors::template::TemplateProcessing; use tokenizers::Tokenizer; use tokio::select; use tokio::signal; use tokio::sync::oneshot; use tokio::time::Instant; use tower_http::cors::{AllowOrigin, CorsLayer}; use tracing::{info_span, instrument, Instrument}; use utoipa::OpenApi; use utoipa_swagger_ui::SwaggerUi; /// Generate tokens if `stream == false` or a stream of token if `stream == true` #[utoipa::path( post, tag = "Text Generation Inference", path = "/", request_body = CompatGenerateRequest, responses( (status = 200, description = "Generated Text", content( ("application/json" = GenerateResponse), ("text/event-stream" = StreamResponse), )), (status = 424, description = "Generation Error", body = ErrorResponse, example = json ! ({"error": "Request failed during generation"})), (status = 429, description = "Model is overloaded", body = ErrorResponse, example = json ! ({"error": "Model is overloaded"})), (status = 422, description = "Input validation error", body = ErrorResponse, example = json ! ({"error": "Input validation error"})), (status = 500, description = "Incomplete generation", body = ErrorResponse, example = json ! ({"error": "Incomplete generation"})), ) )] #[instrument(skip(infer, req))] async fn compat_generate( Extension(default_return_full_text): Extension, infer: Extension, compute_type: Extension, Json(mut req): Json, ) -> Result)> { // default return_full_text given the pipeline_tag if req.parameters.return_full_text.is_none() { req.parameters.return_full_text = Some(default_return_full_text) } // switch on stream if req.stream { Ok(generate_stream(infer, compute_type, Json(req.into())) .await .into_response()) } else { let (headers, Json(generation)) = generate(infer, compute_type, Json(req.into())).await?; // wrap generation inside a Vec to match api-inference Ok((headers, Json(vec![generation])).into_response()) } } /// Text Generation Inference endpoint info #[utoipa::path( get, tag = "Text Generation Inference", path = "/info", responses((status = 200, description = "Served model info", body = Info)) )] #[instrument] async fn get_model_info(info: Extension) -> Json { Json(info.0) } #[utoipa::path( post, tag = "Text Generation Inference", path = "/chat_tokenize", request_body = ChatRequest, responses((status = 200, description = "Templated and tokenized ChatRequest", body = ChatTokenizeResponse)) )] async fn get_chat_tokenize( Extension(infer): Extension, Json(req): Json, ) -> Result<(HeaderMap, Json), (StatusCode, Json)> { metrics::counter!("tgi_request_count").increment(1); let ChatRequest { model, max_tokens, messages, seed, stop, stream, tools, tool_choice, tool_prompt, temperature, response_format, guideline, .. } = req; let tool_prompt = tool_prompt.unwrap_or_default(); let (inputs, _grammar, _tool_grammar) = prepare_chat_input( &infer, response_format, tools, tool_choice, &tool_prompt, guideline, messages, )?; let generate_request = GenerateRequest { inputs, parameters: GenerateParameters { best_of: None, temperature, repetition_penalty: None, frequency_penalty: None, top_k: None, top_p: None, typical_p: None, do_sample: true, max_new_tokens: max_tokens, return_full_text: None, stop: stop.unwrap_or_default(), truncate: None, watermark: false, details: false, decoder_input_details: !stream, seed, top_n_tokens: None, grammar: _grammar, adapter_id: model.as_ref().filter(|m| *m != "tgi").map(String::from), }, }; let input = generate_request.inputs.clone(); let encoding = infer.tokenize(generate_request).await?; if let Some(encoding) = encoding { let tokens: Vec = encoding .get_ids() .iter() .zip(encoding.get_offsets()) .map(|(&id, &(start, stop))| { let text = input .chars() .skip(start) .take(stop - start) .collect::(); SimpleToken { id, text, start, stop, } }) .collect(); let resp = ChatTokenizeResponse { tokenize_response: TokenizeResponse(tokens), templated_text: input, }; Ok((HeaderMap::new(), Json(resp))) } else { Err(( StatusCode::NOT_FOUND, Json(ErrorResponse { error: "No fast tokenizer or tokenizer.json for this model".to_string(), error_type: "no fast tokenizer".to_string(), }), )) } } #[utoipa::path( get, tag = "Text Generation Inference", path = "/health", responses( (status = 200, description = "Everything is working fine"), (status = 503, description = "Text generation inference is down", body = ErrorResponse, example = json ! ({"error": "unhealthy", "error_type": "healthcheck"})), ) )] #[instrument(skip(infer))] /// Health check method async fn health(infer: Extension) -> Result<(), (StatusCode, Json)> { match infer.health().await { true => Ok(()), false => Err(( StatusCode::SERVICE_UNAVAILABLE, Json(ErrorResponse { error: "unhealthy".to_string(), error_type: "healthcheck".to_string(), }), )), } } /// Generate tokens #[utoipa::path( post, tag = "Text Generation Inference", path = "/generate", request_body = GenerateRequest, responses( (status = 200, description = "Generated Text", body = GenerateResponse), (status = 424, description = "Generation Error", body = ErrorResponse, example = json ! ({"error": "Request failed during generation"})), (status = 429, description = "Model is overloaded", body = ErrorResponse, example = json ! ({"error": "Model is overloaded"})), (status = 422, description = "Input validation error", body = ErrorResponse, example = json ! ({"error": "Input validation error"})), (status = 500, description = "Incomplete generation", body = ErrorResponse, example = json ! ({"error": "Incomplete generation"})), ) )] #[instrument( skip_all, fields( parameters = ? req.parameters, total_time, validation_time, queue_time, inference_time, time_per_token, seed, ) )] async fn generate( infer: Extension, Extension(ComputeType(compute_type)): Extension, Json(req): Json, ) -> Result<(HeaderMap, Json), (StatusCode, Json)> { let span = tracing::Span::current(); generate_internal(infer, ComputeType(compute_type), Json(req), span).await } pub(crate) async fn generate_internal( infer: Extension, ComputeType(compute_type): ComputeType, Json(req): Json, span: tracing::Span, ) -> Result<(HeaderMap, Json), (StatusCode, Json)> { let start_time = Instant::now(); metrics::counter!("tgi_request_count").increment(1); // Do not long ultra long inputs, like image payloads. tracing::debug!("Input: {}", &req.inputs[..1000.min(req.inputs.len())]); let compute_characters = req.inputs.chars().count(); let mut add_prompt = None; if req.parameters.return_full_text.unwrap_or(false) { add_prompt = Some(req.inputs.clone()); } let details: bool = req.parameters.details || req.parameters.decoder_input_details; // Inference let (response, best_of_responses) = match req.parameters.best_of { Some(best_of) if best_of > 1 => { let (response, best_of_responses) = infer.generate_best_of(req, best_of).await?; (response, Some(best_of_responses)) } _ => (infer.generate(req).await?, None), }; // Token details let input_length = response._input_length; let details = match details { true => { // convert best_of_responses let best_of_sequences = best_of_responses.map(|responses: Vec| { responses .into_iter() .map(|response: InferResponse| { // Add prompt if return_full_text let mut output_text = response.generated_text.text; if let Some(prompt) = &add_prompt { output_text = prompt.clone() + &output_text; } BestOfSequence { generated_text: output_text, finish_reason: response.generated_text.finish_reason, generated_tokens: response.generated_text.generated_tokens, prefill: response.prefill, tokens: response.tokens, top_tokens: response.top_tokens, seed: response.generated_text.seed, } }) .collect() }); Some(Details { finish_reason: response.generated_text.finish_reason, generated_tokens: response.generated_text.generated_tokens, prefill: response.prefill, tokens: response.tokens, seed: response.generated_text.seed, best_of_sequences, top_tokens: response.top_tokens, }) } false => None, }; // Timings let total_time = start_time.elapsed(); let validation_time = response.queued - start_time; let queue_time = response.start - response.queued; let inference_time = Instant::now() - response.start; let time_per_token = inference_time / response.generated_text.generated_tokens; // Tracing metadata span.record("total_time", format!("{total_time:?}")); span.record("validation_time", format!("{validation_time:?}")); span.record("queue_time", format!("{queue_time:?}")); span.record("inference_time", format!("{inference_time:?}")); span.record("time_per_token", format!("{time_per_token:?}")); span.record("seed", format!("{:?}", response.generated_text.seed)); // Headers let mut headers = HeaderMap::new(); headers.insert("x-compute-type", compute_type.parse().unwrap()); headers.insert( "x-compute-time", total_time.as_secs_f64().to_string().parse().unwrap(), ); headers.insert( "x-compute-characters", compute_characters.to_string().parse().unwrap(), ); headers.insert( "x-total-time", total_time.as_millis().to_string().parse().unwrap(), ); headers.insert( "x-validation-time", validation_time.as_millis().to_string().parse().unwrap(), ); headers.insert( "x-queue-time", queue_time.as_millis().to_string().parse().unwrap(), ); headers.insert( "x-inference-time", inference_time.as_millis().to_string().parse().unwrap(), ); headers.insert( "x-time-per-token", time_per_token.as_millis().to_string().parse().unwrap(), ); headers.insert("x-prompt-tokens", input_length.into()); headers.insert( "x-generated-tokens", response.generated_text.generated_tokens.into(), ); // Metrics metrics::counter!("tgi_request_success").increment(1); metrics::histogram!("tgi_request_duration").record(total_time.as_secs_f64()); metrics::histogram!("tgi_request_validation_duration").record(validation_time.as_secs_f64()); metrics::histogram!("tgi_request_queue_duration").record(queue_time.as_secs_f64()); metrics::histogram!("tgi_request_inference_duration").record(inference_time.as_secs_f64()); metrics::histogram!("tgi_request_mean_time_per_token_duration") .record(time_per_token.as_secs_f64()); metrics::histogram!("tgi_request_generated_tokens") .record(response.generated_text.generated_tokens as f64); // Send response let mut output_text = response.generated_text.text; if let Some(prompt) = add_prompt { output_text = prompt + &output_text; } tracing::debug!("Output: {}", output_text); tracing::info!("Success"); let response = GenerateResponse { generated_text: output_text, details, }; Ok((headers, Json(response))) } /// Generate a stream of token using Server-Sent Events #[utoipa::path( post, tag = "Text Generation Inference", path = "/generate_stream", request_body = GenerateRequest, responses( (status = 200, description = "Generated Text", body = StreamResponse, content_type = "text/event-stream"), (status = 424, description = "Generation Error", body = ErrorResponse, example = json ! ({"error": "Request failed during generation"}), content_type = "text/event-stream"), (status = 429, description = "Model is overloaded", body = ErrorResponse, example = json ! ({"error": "Model is overloaded"}), content_type = "text/event-stream"), (status = 422, description = "Input validation error", body = ErrorResponse, example = json ! ({"error": "Input validation error"}), content_type = "text/event-stream"), (status = 500, description = "Incomplete generation", body = ErrorResponse, example = json ! ({"error": "Incomplete generation"}), content_type = "text/event-stream"), ) )] #[instrument( skip_all, fields( parameters = ? req.parameters, total_time, validation_time, queue_time, inference_time, time_per_token, seed, ) )] async fn generate_stream( Extension(infer): Extension, Extension(compute_type): Extension, Json(req): Json, ) -> ( HeaderMap, Sse>>, ) { let span = tracing::Span::current(); let on_message_callback = |stream_token: StreamResponse| { let event = Event::default(); event.json_data(stream_token).unwrap() }; let (headers, response_stream) = generate_stream_internal(infer, compute_type, Json(req), on_message_callback, span).await; let sse = Sse::new(response_stream).keep_alive(KeepAlive::default()); (headers, sse) } async fn generate_stream_internal( infer: Infer, ComputeType(compute_type): ComputeType, Json(req): Json, on_message_callback: impl Fn(StreamResponse) -> Event, span: tracing::Span, ) -> (HeaderMap, impl Stream>) { let start_time = Instant::now(); metrics::counter!("tgi_request_count").increment(1); tracing::debug!("Input: {}", req.inputs); let compute_characters = req.inputs.chars().count(); let mut headers = HeaderMap::new(); headers.insert("x-compute-type", compute_type.parse().unwrap()); headers.insert( "x-compute-characters", compute_characters.to_string().parse().unwrap(), ); headers.insert("X-Accel-Buffering", "no".parse().unwrap()); let stream = async_stream::stream! { // Inference let mut end_reached = false; let mut error = false; let mut add_prompt = None; if req.parameters.return_full_text.unwrap_or(false) { add_prompt = Some(req.inputs.clone()); } let details = req.parameters.details; let best_of = req.parameters.best_of.unwrap_or(1); if best_of != 1 { let err = InferError::from(ValidationError::BestOfStream); metrics::counter!("tgi_request_failure", "err" => "validation").increment(1); tracing::error!("{err}"); yield Ok(Event::from(err)); } else if req.parameters.decoder_input_details { let err = InferError::from(ValidationError::PrefillDetailsStream); metrics::counter!("tgi_request_failure", "err" => "validation").increment(1); tracing::error!("{err}"); yield Ok(Event::from(err)); } else { match infer.generate_stream(req).instrument(info_span!(parent: &span, "async_stream")).await { // Keep permit as long as generate_stream lives Ok((_permit, input_length, response_stream)) => { let mut index = 0; let mut response_stream = Box::pin(response_stream); // Server-Sent Event stream while let Some(response) = response_stream.next().await { index += 1; match response { Ok(response) => { match response { // Prefill is ignored InferStreamResponse::Prefill(_) => {} // Yield event for every new token InferStreamResponse::Intermediate{ token, top_tokens, } => { tracing::debug!(parent: &span, "Token: {:?}", token); // StreamResponse let stream_token = StreamResponse { index, token, top_tokens, generated_text: None, details: None, }; let event = on_message_callback(stream_token); yield Ok(event); } // Yield event for last token and compute timings InferStreamResponse::End { token, generated_text, start, queued, top_tokens, } => { // Token details let details = match details { true => Some(StreamDetails { finish_reason: generated_text.finish_reason, generated_tokens: generated_text.generated_tokens, seed: generated_text.seed, input_length, }), false => None, }; // Timings let total_time = start_time.elapsed(); let validation_time = queued - start_time; let queue_time = start - queued; let inference_time = Instant::now() - start; let time_per_token = inference_time / generated_text.generated_tokens; // Tracing metadata span.record("total_time", format!("{total_time:?}")); span.record("validation_time", format!("{validation_time:?}")); span.record("queue_time", format!("{queue_time:?}")); span.record("inference_time", format!("{inference_time:?}")); span.record("time_per_token", format!("{time_per_token:?}")); span.record("seed", format!("{:?}", generated_text.seed)); // Metrics metrics::counter!("tgi_request_success").increment(1); metrics::histogram!("tgi_request_duration").record(total_time.as_secs_f64()); metrics::histogram!("tgi_request_validation_duration").record(validation_time.as_secs_f64()); metrics::histogram!("tgi_request_queue_duration").record(queue_time.as_secs_f64()); metrics::histogram!("tgi_request_inference_duration").record(inference_time.as_secs_f64()); metrics::histogram!("tgi_request_mean_time_per_token_duration").record(time_per_token.as_secs_f64()); metrics::histogram!("tgi_request_generated_tokens").record(generated_text.generated_tokens as f64); // StreamResponse end_reached = true; let mut output_text = generated_text.text; if let Some(prompt) = add_prompt { output_text = prompt + &output_text; } tracing::debug!(parent: &span, "Output: {}", output_text); tracing::info!(parent: &span, "Success"); let stream_token = StreamResponse { index, token, top_tokens, generated_text: Some(output_text), details }; let event = on_message_callback(stream_token); yield Ok(event); break; } } } // yield error Err(err) => { error = true; yield Ok(Event::from(err)); break; } } } }, // yield error Err(err) => { error = true; yield Ok(Event::from(err)); } } // Check if generation reached the end // Skip if we already sent an error if !end_reached && !error { let err = InferError::IncompleteGeneration; metrics::counter!("tgi_request_failure", "err" => "incomplete").increment(1); tracing::error!("{err}"); yield Ok(Event::from(err)); } } }; (headers, stream) } /// Generate tokens #[utoipa::path( post, tag = "Text Generation Inference", path = "/v1/completions", request_body = CompletionRequest, responses( (status = 200, description = "Generated Chat Completion", content( ("application/json" = CompletionFinal), ("text/event-stream" = Chunk), )), (status = 424, description = "Generation Error", body = ErrorResponse, example = json ! ({"error": "Request failed during generation"})), (status = 429, description = "Model is overloaded", body = ErrorResponse, example = json ! ({"error": "Model is overloaded"})), (status = 422, description = "Input validation error", body = ErrorResponse, example = json ! ({"error": "Input validation error"})), (status = 500, description = "Incomplete generation", body = ErrorResponse, example = json ! ({"error": "Incomplete generation"})), ) )] #[instrument( skip_all, fields( // parameters = ? req.parameters, total_time, validation_time, queue_time, inference_time, time_per_token, seed, ) )] async fn completions( Extension(infer): Extension, Extension(compute_type): Extension, Extension(info): Extension, Json(req): Json, ) -> Result)> { let span = tracing::Span::current(); metrics::counter!("tgi_request_count").increment(1); let CompletionRequest { model, max_tokens, seed, stop, stream, temperature, .. } = req; let max_new_tokens = max_tokens.or(Some(100)); let stop = stop.unwrap_or_default(); // enable greedy only when temperature is 0 let (do_sample, temperature) = match temperature { Some(temperature) if temperature == 0.0 => (false, None), other => (true, other), }; // if suffix is present throw an error if req.suffix.is_some() { metrics::counter!("tgi_request_failure", "err" => "validation").increment(1); return Err(( StatusCode::UNPROCESSABLE_ENTITY, Json(ErrorResponse { error: "Suffix is not supported and can be achieved by preprocessing the prompt." .to_string(), error_type: "suffix not supported".to_string(), }), )); } if req.prompt.0.len() > info.max_client_batch_size { metrics::counter!("tgi_request_failure", "err" => "validation").increment(1); return Err(( StatusCode::UNPROCESSABLE_ENTITY, Json(ErrorResponse { error: format!( "Number of prompts exceeds the maximum allowed batch size of {}", info.max_client_batch_size ), error_type: "batch size exceeded".to_string(), }), )); } let generate_requests: Vec = req .prompt .0 .iter() .map(|prompt| GenerateRequest { inputs: prompt.to_string(), parameters: GenerateParameters { best_of: None, temperature, repetition_penalty: req.repetition_penalty, frequency_penalty: req.frequency_penalty, top_k: None, top_p: req.top_p, typical_p: None, do_sample, max_new_tokens, return_full_text: None, stop: stop.clone(), truncate: None, watermark: false, details: true, decoder_input_details: !stream, seed, top_n_tokens: None, grammar: None, adapter_id: model.as_ref().filter(|m| *m != "tgi").map(String::from), }, }) .collect(); let mut x_compute_type = None; let mut x_compute_characters = 0u32; let mut x_accel_buffering = None; if stream { let mut response_streams = FuturesOrdered::new(); for (index, generate_request) in generate_requests.into_iter().enumerate() { let model_id = info.model_id.clone(); let system_fingerprint = format!("{}-{}", info.version, info.docker_label.unwrap_or("native")); let infer_clone = infer.clone(); let compute_type_clone = compute_type.clone(); let span_clone = span.clone(); // Create a future for each generate_stream_internal call. let generate_future = async move { let on_message_callback = move |stream_token: StreamResponse| { let event = Event::default(); let current_time = std::time::SystemTime::now() .duration_since(std::time::UNIX_EPOCH) .unwrap_or_else(|_| std::time::Duration::from_secs(0)) .as_secs(); let message = match stream_token.details { Some(details) => { let completion_tokens = details.generated_tokens; let prompt_tokens = details.input_length; let total_tokens = prompt_tokens + completion_tokens; Completion::Final(CompletionFinal { id: String::new(), created: current_time, model: model_id.clone(), system_fingerprint: system_fingerprint.clone(), choices: vec![CompletionComplete { finish_reason: details.finish_reason.to_string(), index: index as u32, logprobs: None, text: stream_token.token.text, }], usage: Usage { prompt_tokens, completion_tokens, total_tokens, }, }) } None => Completion::Chunk(Chunk { id: String::new(), created: current_time, choices: vec![CompletionComplete { finish_reason: String::new(), index: index as u32, logprobs: None, text: stream_token.token.text, }], model: model_id.clone(), system_fingerprint: system_fingerprint.clone(), }), }; event .json_data(message) .unwrap_or_else(|_e| Event::default()) }; let (header_tx, header_rx) = oneshot::channel(); let (sse_tx, sse_rx) = tokio::sync::mpsc::unbounded_channel(); tokio::spawn(async move { let (header_map, sse) = generate_stream_internal( infer_clone.clone(), compute_type_clone.clone(), Json(generate_request), on_message_callback, span_clone.clone(), ) .await; // send and dont wait for response let _ = header_tx.send(header_map); // pin an emit messages to the sse_tx let mut sse = Box::pin(sse); while let Some(event) = sse.next().await { if sse_tx.send(event).is_err() { tracing::error!("Failed to send event. Receiver dropped."); break; } } }); (header_rx, sse_rx) }; response_streams.push_back(generate_future); } let mut all_rxs = vec![]; while let Some((header_rx, sse_rx)) = response_streams.next().await { all_rxs.push(sse_rx); // get the headers from the first response of each stream let headers = header_rx.await.map_err(|e| { tracing::error!("Failed to get headers: {:?}", e); ( StatusCode::INTERNAL_SERVER_ERROR, Json(ErrorResponse { error: "Failed to get headers".to_string(), error_type: "headers".to_string(), }), ) })?; if x_compute_type.is_none() { x_compute_type = headers .get("x-compute-type") .and_then(|v| v.to_str().ok()) .map(|v| v.to_string()); x_accel_buffering = headers .get("x-accel-buffering") .and_then(|v| v.to_str().ok()) .map(|v| v.to_string()); } x_compute_characters += headers .get("x-compute-characters") .and_then(|v| v.to_str().ok()) .and_then(|v| v.parse().ok()) .unwrap_or(0); } let mut headers = HeaderMap::new(); if let Some(x_compute_type) = x_compute_type { headers.insert("x-compute-type", x_compute_type.parse().unwrap()); } headers.insert("x-compute-characters", x_compute_characters.into()); if let Some(x_accel_buffering) = x_accel_buffering { headers.insert("x-accel-buffering", x_accel_buffering.parse().unwrap()); } // now sink the sse streams into a single stream and remove the ones that are done let stream: AsyncStream, _> = async_stream::stream! { loop { let mut i = 0; while i < all_rxs.len() { let rx = &mut all_rxs[i]; select! { Some(event) = rx.recv() => { yield event; } else => { all_rxs.remove(i); continue; // skip the increment to handle the next element at the same index } } i += 1; // only increment when no element was removed } if all_rxs.is_empty() { break; } } }; let stream = stream.chain(futures::stream::once(async { Ok(Event::default().data("[DONE]")) })); let sse = Sse::new(stream).keep_alive(KeepAlive::default()); Ok((headers, sse).into_response()) } else { let current_time = std::time::SystemTime::now() .duration_since(std::time::UNIX_EPOCH) .unwrap_or_else(|_| std::time::Duration::from_secs(0)) .as_secs(); let responses = FuturesUnordered::new(); for (index, generate_request) in generate_requests.into_iter().enumerate() { let infer_clone = infer.clone(); let compute_type_clone = compute_type.clone(); let span_clone = span.clone(); let response_future = async move { let result = generate_internal( Extension(infer_clone), compute_type_clone, Json(generate_request), span_clone, ) .await; result.map(|(headers, generation)| (index, headers, generation)) }; responses.push(response_future); } let generate_responses = responses.try_collect::>().await?; let mut prompt_tokens = 0u32; let mut completion_tokens = 0u32; let mut total_tokens = 0u32; let mut x_compute_time = 0u32; let mut x_total_time = 0u32; let mut x_validation_time = 0u32; let mut x_queue_time = 0u32; let mut x_inference_time = 0u32; let mut x_time_per_token = 0u32; let mut x_prompt_tokens = 0u32; let mut x_generated_tokens = 0u32; let choices = generate_responses .into_iter() .map(|(index, headers, Json(generation))| { let details = generation.details.ok_or(( // this should never happen but handle if details are missing unexpectedly StatusCode::INTERNAL_SERVER_ERROR, Json(ErrorResponse { error: "No details in generation".to_string(), error_type: "no details".to_string(), }), ))?; if x_compute_type.is_none() { x_compute_type = headers .get("x-compute-type") .and_then(|v| v.to_str().ok()) .map(|v| v.to_string()); } // accumulate headers and usage from each response x_compute_time += headers .get("x-compute-time") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_compute_characters += headers .get("x-compute-characters") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_total_time += headers .get("x-total-time") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_validation_time += headers .get("x-validation-time") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_queue_time += headers .get("x-queue-time") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_inference_time += headers .get("x-inference-time") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_time_per_token += headers .get("x-time-per-token") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_prompt_tokens += headers .get("x-prompt-tokens") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); x_generated_tokens += headers .get("x-generated-tokens") .and_then(|v| v.to_str().ok()?.parse().ok()) .unwrap_or(0); prompt_tokens += details.prefill.len() as u32; completion_tokens += details.generated_tokens; total_tokens += details.prefill.len() as u32 + details.generated_tokens; Ok(CompletionComplete { finish_reason: details.finish_reason.format(true), index: index as u32, logprobs: None, text: generation.generated_text, }) }) .collect::, _>>() .map_err(|(status, Json(err))| (status, Json(err)))?; let response = Completion::Final(CompletionFinal { id: "".to_string(), created: current_time, model: info.model_id.clone(), system_fingerprint: format!( "{}-{}", info.version, info.docker_label.unwrap_or("native") ), choices, usage: Usage { prompt_tokens, completion_tokens, total_tokens, }, }); // headers similar to `generate` but aggregated let mut headers = HeaderMap::new(); if let Some(x_compute_type) = x_compute_type { headers.insert("x-compute-type", x_compute_type.parse().unwrap()); } headers.insert("x-compute-characters", x_compute_characters.into()); headers.insert("x-total-time", x_total_time.into()); headers.insert("x-validation-time", x_validation_time.into()); headers.insert("x-queue-time", x_queue_time.into()); headers.insert("x-inference-time", x_inference_time.into()); headers.insert("x-time-per-token", x_time_per_token.into()); headers.insert("x-prompt-tokens", x_prompt_tokens.into()); headers.insert("x-generated-tokens", x_generated_tokens.into()); if let Some(x_accel_buffering) = x_accel_buffering { headers.insert("x-accel-buffering", x_accel_buffering.parse().unwrap()); } Ok((headers, Json(response)).into_response()) } } /// Generate tokens #[utoipa::path( post, tag = "Text Generation Inference", path = "/v1/chat/completions", request_body = ChatRequest, responses( (status = 200, description = "Generated Chat Completion", content( ("application/json" = ChatCompletion), ("text/event-stream" = ChatCompletionChunk), )), (status = 424, description = "Generation Error", body = ErrorResponse, example = json ! ({"error": "Request failed during generation"})), (status = 429, description = "Model is overloaded", body = ErrorResponse, example = json ! ({"error": "Model is overloaded"})), (status = 422, description = "Input validation error", body = ErrorResponse, example = json ! ({"error": "Input validation error"})), (status = 500, description = "Incomplete generation", body = ErrorResponse, example = json ! ({"error": "Incomplete generation"})), ) )] #[instrument( skip_all, fields( // parameters = ? req.parameters, total_time, validation_time, queue_time, inference_time, time_per_token, seed, ) )] async fn chat_completions( Extension(infer): Extension, Extension(compute_type): Extension, Extension(info): Extension, Json(req): Json, ) -> Result)> { let span = tracing::Span::current(); metrics::counter!("tgi_request_count").increment(1); let ChatRequest { model, logprobs, max_tokens, messages, presence_penalty, seed, stop, stream, tools, tool_choice, tool_prompt, temperature, response_format, guideline, .. } = req; let repetition_penalty = presence_penalty.map(|x| x + 2.0); let max_new_tokens = max_tokens.or(Some(100)); let logprobs = logprobs.unwrap_or(false); let tool_prompt = tool_prompt.unwrap_or_default(); let stop = stop.unwrap_or_default(); // enable greedy only when temperature is 0 let (do_sample, temperature) = match temperature { Some(temperature) if temperature == 0.0 => (false, None), other => (true, other), }; let (inputs, grammar, tool_grammar) = prepare_chat_input( &infer, response_format, tools, tool_choice, &tool_prompt, guideline, messages, )?; // build the request passing some parameters let generate_request = GenerateRequest { inputs: inputs.to_string(), parameters: GenerateParameters { best_of: None, temperature, repetition_penalty, frequency_penalty: req.frequency_penalty, top_k: None, top_p: req.top_p, typical_p: None, do_sample, max_new_tokens, return_full_text: None, stop, truncate: None, watermark: false, details: true, decoder_input_details: !stream, seed, top_n_tokens: req.top_logprobs, grammar, adapter_id: model.filter(|m| *m != "tgi").map(String::from), }, }; // static values that will be returned in all cases let model_id = info.model_id.clone(); let system_fingerprint = format!("{}-{}", info.version, info.docker_label.unwrap_or("native")); // switch on stream if stream { // pass this callback to the stream generation and build the required event structure let on_message_callback = move |stream_token: StreamResponse| { let event = Event::default(); let current_time = std::time::SystemTime::now() .duration_since(std::time::UNIX_EPOCH) .unwrap_or_else(|_| std::time::Duration::from_secs(0)) .as_secs(); let logprobs = logprobs.then(|| { ChatCompletionLogprobs::from((stream_token.token.clone(), stream_token.top_tokens)) }); // replace the content with the tool calls if grammar is present let (content, tool_calls) = if tool_grammar.is_some() { (None, Some(vec![stream_token.token.text])) } else { let content = if !stream_token.token.special { Some(stream_token.token.text) } else { None }; (content, None) }; event .json_data(CompletionType::ChatCompletionChunk( ChatCompletionChunk::new( model_id.clone(), system_fingerprint.clone(), content, tool_calls, current_time, logprobs, stream_token.details.map(|d| d.finish_reason.format(true)), ), )) .unwrap_or_else(|e| { println!("Failed to serialize ChatCompletionChunk: {:?}", e); Event::default() }) }; let (headers, response_stream) = generate_stream_internal( infer, compute_type, Json(generate_request), on_message_callback, span, ) .await; let response_stream = response_stream.chain(futures::stream::once(async { Ok(Event::default().data("[DONE]")) })); let sse = Sse::new(response_stream).keep_alive(KeepAlive::default()); Ok((headers, sse).into_response()) } else { let (headers, Json(generation)) = generate_internal(Extension(infer), compute_type, Json(generate_request), span).await?; let current_time = std::time::SystemTime::now() .duration_since(std::time::UNIX_EPOCH) .unwrap_or_else(|_| std::time::Duration::from_secs(0)) .as_secs(); let (tool_calls, output) = if tool_grammar.is_some() { let gen_text_value: Value = serde_json::from_str(&generation.generated_text).map_err(|e| { InferError::ToolError(format!( "Failed to parse generated text: {} {:?}", e, generation.generated_text )) })?; let function = gen_text_value.get("function").ok_or(InferError::ToolError( "No function found in generated text".to_string(), ))?; let name = function .get("_name") .and_then(Value::as_str) .ok_or(InferError::ToolError( "No _name found in generated text".to_string(), ))? .to_string(); let mut arguments = function.clone(); if let Value::Object(ref mut props) = arguments { props.remove("_name"); } let tool_calls = vec![ToolCall { id: "0".to_string(), r#type: "function".to_string(), function: FunctionDefinition { description: None, name, arguments, }, }]; (Some(tool_calls), None) } else { (None, Some(generation.generated_text)) }; // build the complete response object with the full text let response = CompletionType::ChatCompletion(ChatCompletion::new( model_id, system_fingerprint, output, current_time, generation.details.unwrap(), logprobs, tool_calls, )); // wrap generation inside a Vec to match api-inference Ok((headers, Json(response)).into_response()) } } /// Generate tokens from Vertex request #[utoipa::path( post, tag = "Text Generation Inference", path = "/vertex", request_body = VertexRequest, responses( (status = 200, description = "Generated Text", body = VertexResponse), (status = 424, description = "Generation Error", body = ErrorResponse, example = json ! ({"error": "Request failed during generation"})), (status = 429, description = "Model is overloaded", body = ErrorResponse, example = json ! ({"error": "Model is overloaded"})), (status = 422, description = "Input validation error", body = ErrorResponse, example = json ! ({"error": "Input validation error"})), (status = 500, description = "Incomplete generation", body = ErrorResponse, example = json ! ({"error": "Incomplete generation"})), ) )] #[instrument( skip_all, fields( total_time, validation_time, queue_time, inference_time, time_per_token, seed, ) )] async fn vertex_compatibility( Extension(infer): Extension, Extension(compute_type): Extension, Json(req): Json, ) -> Result)> { let span = tracing::Span::current(); metrics::counter!("tgi_request_count").increment(1); // check that theres at least one instance if req.instances.is_empty() { return Err(( StatusCode::UNPROCESSABLE_ENTITY, Json(ErrorResponse { error: "Input validation error".to_string(), error_type: "Input validation error".to_string(), }), )); } // Process all instances let predictions = req .instances .iter() .map(|instance| { let generate_request = GenerateRequest { inputs: instance.inputs.clone(), parameters: GenerateParameters { do_sample: true, max_new_tokens: instance.parameters.as_ref().and_then(|p| p.max_new_tokens), seed: instance.parameters.as_ref().and_then(|p| p.seed), details: true, decoder_input_details: true, ..Default::default() }, }; async { generate_internal( Extension(infer.clone()), compute_type.clone(), Json(generate_request), span.clone(), ) .await .map(|(_, Json(generation))| generation.generated_text) .map_err(|_| { ( StatusCode::INTERNAL_SERVER_ERROR, Json(ErrorResponse { error: "Incomplete generation".into(), error_type: "Incomplete generation".into(), }), ) }) } }) .collect::>() .try_collect::>() .await?; let response = VertexResponse { predictions }; Ok((HeaderMap::new(), Json(response)).into_response()) } /// Tokenize inputs #[utoipa::path( post, tag = "Text Generation Inference", path = "/tokenize", request_body = GenerateRequest, responses( (status = 200, description = "Tokenized ids", body = TokenizeResponse), (status = 404, description = "No tokenizer found", body = ErrorResponse, example = json ! ({"error": "No fast tokenizer available"})), ) )] #[instrument(skip_all)] async fn tokenize( Extension(infer): Extension, Json(req): Json, ) -> Result, (StatusCode, Json)> { let input = req.inputs.clone(); let encoding = infer.tokenize(req).await?; if let Some(encoding) = encoding { let tokens: Vec = encoding .get_ids() .iter() .zip(encoding.get_offsets()) .map(|(&id, &(start, stop))| { let text = input .chars() .skip(start) .take(stop - start) .collect::(); SimpleToken { id, text, start, stop, } }) .collect(); Ok(Json(TokenizeResponse(tokens))) } else { Err(( StatusCode::NOT_FOUND, Json(ErrorResponse { error: "No fast tokenizer or tokenizer.json for this model".to_string(), error_type: "no fast tokenizer".to_string(), }), )) } } /// Prometheus metrics scrape endpoint #[utoipa::path( get, tag = "Text Generation Inference", path = "/metrics", responses((status = 200, description = "Prometheus Metrics", body = String)) )] async fn metrics(prom_handle: Extension) -> String { prom_handle.render() } #[derive(Clone, Debug)] pub(crate) struct ComputeType(String); // OpenAPI documentation #[derive(OpenApi)] #[openapi( paths( health, get_model_info, compat_generate, generate, generate_stream, chat_completions, completions, tokenize, metrics, ), components( schemas( Info, CompatGenerateRequest, GenerateRequest, GrammarType, ChatRequest, Message, MessageContent, MessageChunk, Url, FunctionName, OutputMessage, TextMessage, ToolCallMessage, ToolCallDelta, ChatCompletionComplete, ChatCompletionChoice, ChatCompletionDelta, ChatCompletionChunk, ChatCompletionLogprob, ChatCompletionLogprobs, ChatCompletionTopLogprob, ChatCompletion, CompletionRequest, CompletionComplete, Chunk, Completion, CompletionFinal, Prompt, GenerateParameters, PrefillToken, Token, GenerateResponse, TokenizeResponse, SimpleToken, BestOfSequence, Details, FinishReason, StreamResponse, StreamDetails, ErrorResponse, GrammarType, Usage, DeltaToolCall, ToolType, Tool, ToolCall, Function, FunctionDefinition, ToolChoice, ) ), tags( (name = "Text Generation Inference", description = "Hugging Face Text Generation Inference API") ), info( title = "Text Generation Inference", license( name = "Apache 2.0", url = "https://www.apache.org/licenses/LICENSE-2.0" ) ) )] pub struct ApiDoc; pub fn schema() -> ApiDoc { ApiDoc } /// Serving method #[allow(clippy::too_many_arguments)] pub async fn run( backend: impl Backend + Send + Sync + 'static, max_concurrent_requests: usize, max_best_of: usize, max_stop_sequences: usize, max_top_n_tokens: u32, max_input_tokens: usize, max_total_tokens: usize, validation_workers: usize, api_key: Option, tokenizer_name: String, tokenizer_config_path: Option, revision: Option, hostname: String, port: u16, cors_allow_origin: Option>, ngrok: bool, _ngrok_authtoken: Option, _ngrok_edge: Option, messages_api_enabled: bool, disable_grammar_support: bool, max_client_batch_size: usize, usage_stats_level: usage_stats::UsageStatsLevel, ) -> Result<(), WebServerError> { // CORS allowed origins // map to go inside the option and then map to parse from String to HeaderValue // Finally, convert to AllowOrigin let allow_origin: Option = cors_allow_origin.map(|cors_allow_origin| { AllowOrigin::list( cors_allow_origin .iter() .map(|origin| origin.parse::().unwrap()), ) }); // Parse Huggingface hub token let authorization_token = std::env::var("HF_TOKEN") .or_else(|_| std::env::var("HUGGING_FACE_HUB_TOKEN")) .ok(); // Tokenizer instance // This will only be used to validate payloads let local_path = Path::new(&tokenizer_name); // Shared API builder initialization let api_builder = || { let mut builder = ApiBuilder::new() .with_progress(false) .with_token(authorization_token); if let Ok(cache_dir) = std::env::var("HUGGINGFACE_HUB_CACHE") { builder = builder.with_cache_dir(cache_dir.into()); } builder }; // Decide if we need to use the API based on the revision and local path let use_api = revision.is_some() || !local_path.exists() || !local_path.is_dir(); // Initialize API if needed #[derive(Clone)] enum Type { Api(Api), Cache(Cache), None, } let api = if use_api { if std::env::var("HF_HUB_OFFLINE") == Ok("1".to_string()) { let cache = std::env::var("HUGGINGFACE_HUB_CACHE") .map_err(|_| ()) .map(|cache_dir| Cache::new(cache_dir.into())) .unwrap_or_else(|_| Cache::default()); tracing::warn!("Offline mode active using cache defaults"); Type::Cache(cache) } else { tracing::info!("Using the Hugging Face API"); match api_builder().build() { Ok(api) => Type::Api(api), Err(_) => { tracing::warn!("Unable to build the Hugging Face API"); Type::None } } } } else { Type::None }; // Load tokenizer and model info let ( tokenizer_filename, config_filename, tokenizer_config_filename, preprocessor_config_filename, processor_config_filename, model_info, ) = match api { Type::None => ( Some(local_path.join("tokenizer.json")), Some(local_path.join("config.json")), Some(local_path.join("tokenizer_config.json")), Some(local_path.join("preprocessor_config.json")), Some(local_path.join("processor_config.json")), None, ), Type::Api(api) => { let api_repo = api.repo(Repo::with_revision( tokenizer_name.to_string(), RepoType::Model, revision.clone().unwrap_or_else(|| "main".to_string()), )); let tokenizer_filename = match api_repo.get("tokenizer.json").await { Ok(tokenizer_filename) => Some(tokenizer_filename), Err(_) => get_base_tokenizer(&api, &api_repo).await, }; let config_filename = api_repo.get("config.json").await.ok(); let tokenizer_config_filename = api_repo.get("tokenizer_config.json").await.ok(); let preprocessor_config_filename = api_repo.get("preprocessor_config.json").await.ok(); let processor_config_filename = api_repo.get("processor_config.json").await.ok(); let model_info = if let Some(model_info) = get_hub_model_info(&api_repo).await { Some(model_info) } else { tracing::warn!("Could not retrieve model info from the Hugging Face hub."); None }; ( tokenizer_filename, config_filename, tokenizer_config_filename, preprocessor_config_filename, processor_config_filename, model_info, ) } Type::Cache(cache) => { let repo = cache.repo(Repo::with_revision( tokenizer_name.to_string(), RepoType::Model, revision.clone().unwrap_or_else(|| "main".to_string()), )); ( repo.get("tokenizer.json"), repo.get("config.json"), repo.get("tokenizer_config.json"), repo.get("preprocessor_config.json"), repo.get("processor_config.json"), None, ) } }; // Read the JSON contents of the file as an instance of 'HubTokenizerConfig'. let tokenizer_config: Option = if let Some(filename) = tokenizer_config_path { HubTokenizerConfig::from_file(filename) } else { tokenizer_config_filename.and_then(HubTokenizerConfig::from_file) }; let tokenizer_config = tokenizer_config.unwrap_or_else(|| { tracing::warn!("Could not find tokenizer config locally and no API specified"); HubTokenizerConfig::default() }); let tokenizer: Option = tokenizer_filename.and_then(|filename| { let mut tokenizer = Tokenizer::from_file(filename).ok(); if let Some(tokenizer) = &mut tokenizer { if let Some(class) = &tokenizer_config.tokenizer_class { if class == "LlamaTokenizer" || class == "LlamaTokenizerFast"{ if let Ok(post_processor) = create_post_processor(tokenizer, &tokenizer_config) { tracing::info!("Overriding LlamaTokenizer with TemplateProcessing to follow python override defined in https://github.com/huggingface/transformers/blob/4aa17d00690b7f82c95bb2949ea57e22c35b4336/src/transformers/models/llama/tokenization_llama_fast.py#L203-L205"); tokenizer.with_post_processor(post_processor); } } } } tokenizer }); let config: Option = config_filename.and_then(|filename| { std::fs::read_to_string(filename) .ok() .as_ref() .and_then(|c| { let config: Result = serde_json::from_str(c); if let Err(err) = &config { tracing::warn!("Could not parse config {err:?}"); } config.ok() }) }); let model_info = model_info.unwrap_or_else(|| HubModelInfo { model_id: tokenizer_name.to_string(), sha: None, pipeline_tag: None, }); let processor_config = processor_config_filename .and_then(HubProcessorConfig::from_file) .unwrap_or_default(); let preprocessor_config: Option = preprocessor_config_filename.and_then(HubPreprocessorConfig::from_file); tracing::info!("Using config {config:?}"); if tokenizer.is_none() { tracing::warn!("Could not find a fast tokenizer implementation for {tokenizer_name}"); tracing::warn!("Rust input length validation and truncation is disabled"); } // Only send usage stats when TGI is run in container and the function returns Some let is_container = matches!(usage_stats::is_container(), Ok(true)); let user_agent = match (usage_stats_level, is_container) { (usage_stats::UsageStatsLevel::On | usage_stats::UsageStatsLevel::NoStack, true) => { let reduced_args = usage_stats::Args::new( config.clone(), tokenizer_config.tokenizer_class.clone(), max_concurrent_requests, max_best_of, max_stop_sequences, max_top_n_tokens, max_input_tokens, max_total_tokens, // waiting_served_ratio, // max_batch_prefill_tokens, // max_batch_total_tokens, // max_waiting_tokens, // max_batch_size, revision.clone(), validation_workers, messages_api_enabled, disable_grammar_support, max_client_batch_size, usage_stats_level, ); Some(usage_stats::UserAgent::new(reduced_args)) } _ => None, }; if let Some(ref ua) = user_agent { let start_event = usage_stats::UsageStatsEvent::new(ua.clone(), usage_stats::EventType::Start, None); tokio::spawn(async move { start_event.send().await; }); }; let compat_return_full_text = match &model_info.pipeline_tag { None => { tracing::warn!("no pipeline tag found for model {tokenizer_name}"); true } Some(pipeline_tag) => pipeline_tag.as_str() == "text-generation", }; let result = start( backend, max_concurrent_requests, max_best_of, max_stop_sequences, max_top_n_tokens, max_input_tokens, max_total_tokens, validation_workers, api_key, config, (tokenizer, tokenizer_config), (preprocessor_config, processor_config), hostname, port, ngrok, _ngrok_authtoken, _ngrok_edge, messages_api_enabled, disable_grammar_support, max_client_batch_size, model_info, compat_return_full_text, allow_origin, ) .await; if let Some(ua) = user_agent { match result { Ok(_) => { let stop_event = usage_stats::UsageStatsEvent::new( ua.clone(), usage_stats::EventType::Stop, None, ); stop_event.send().await; Ok(()) } Err(e) => { let description = match usage_stats_level { usage_stats::UsageStatsLevel::On => Some(e.to_string()), usage_stats::UsageStatsLevel::NoStack => Some("unknow_error".to_string()), _ => None, }; let event = usage_stats::UsageStatsEvent::new( ua.clone(), usage_stats::EventType::Error, description, ); event.send().await; Err(e) } } } else { result } } #[allow(clippy::too_many_arguments)] async fn start( backend: impl Backend + Send + Sync + 'static, max_concurrent_requests: usize, max_best_of: usize, max_stop_sequences: usize, max_top_n_tokens: u32, max_input_tokens: usize, max_total_tokens: usize, validation_workers: usize, api_key: Option, config: Option, (tokenizer, tokenizer_config): (Option, HubTokenizerConfig), (preprocessor_config, processor_config): (Option, HubProcessorConfig), hostname: String, port: u16, ngrok: bool, _ngrok_authtoken: Option, _ngrok_edge: Option, messages_api_enabled: bool, disable_grammar_support: bool, max_client_batch_size: usize, model_info: HubModelInfo, compat_return_full_text: bool, allow_origin: Option, ) -> Result<(), WebServerError> { // Determine the server port based on the feature and environment variable. let port = if cfg!(feature = "google") { std::env::var("AIP_HTTP_PORT") .map(|aip_http_port| aip_http_port.parse::().unwrap_or(port)) .unwrap_or(port) } else { port }; let addr = match hostname.parse() { Ok(ip) => SocketAddr::new(ip, port), Err(_) => { tracing::warn!("Invalid hostname, defaulting to 0.0.0.0"); SocketAddr::new(IpAddr::V4(Ipv4Addr::new(0, 0, 0, 0)), port) } }; // Create state let validation = Validation::new( validation_workers, tokenizer, config, preprocessor_config, max_best_of, max_stop_sequences, max_top_n_tokens, max_input_tokens, max_total_tokens, disable_grammar_support, ); let infer = Infer::new( backend, validation, max_concurrent_requests, tokenizer_config, processor_config, ); // Duration buckets let duration_matcher = Matcher::Suffix(String::from("duration")); let n_duration_buckets = 35; let mut duration_buckets = Vec::with_capacity(n_duration_buckets); // Minimum duration in seconds let mut value = 0.0001; for _ in 0..n_duration_buckets { // geometric sequence value *= 1.5; duration_buckets.push(value); } // Input Length buckets let input_length_matcher = Matcher::Full(String::from("tgi_request_input_length")); let input_length_buckets: Vec = (0..100) .map(|x| (max_input_tokens as f64 / 100.0) * (x + 1) as f64) .collect(); // Generated tokens buckets let generated_tokens_matcher = Matcher::Full(String::from("tgi_request_generated_tokens")); let generated_tokens_buckets: Vec = (0..100) .map(|x| (max_total_tokens as f64 / 100.0) * (x + 1) as f64) .collect(); // Input Length buckets let max_new_tokens_matcher = Matcher::Full(String::from("tgi_request_max_new_tokens")); let max_new_tokens_buckets: Vec = (0..100) .map(|x| (max_total_tokens as f64 / 100.0) * (x + 1) as f64) .collect(); // Batch size buckets let batch_size_matcher = Matcher::Full(String::from("tgi_batch_next_size")); let batch_size_buckets: Vec = (0..1024).map(|x| (x + 1) as f64).collect(); // Speculated tokens buckets // let skipped_matcher = Matcher::Full(String::from("tgi_request_skipped_tokens")); // let skipped_buckets: Vec = (0..shard_info.speculate + 1).map(|x| x as f64).collect(); // Prometheus handler let builder = PrometheusBuilder::new() .set_buckets_for_metric(duration_matcher, &duration_buckets) .unwrap() .set_buckets_for_metric(input_length_matcher, &input_length_buckets) .unwrap() .set_buckets_for_metric(generated_tokens_matcher, &generated_tokens_buckets) .unwrap() .set_buckets_for_metric(max_new_tokens_matcher, &max_new_tokens_buckets) .unwrap() .set_buckets_for_metric(batch_size_matcher, &batch_size_buckets) .unwrap(); // .set_buckets_for_metric(skipped_matcher, &skipped_buckets) // .unwrap(); let prom_handle = builder .install_recorder() .expect("failed to install metrics recorder"); // CORS layer let allow_origin = allow_origin.unwrap_or(AllowOrigin::any()); let cors_layer = CorsLayer::new() .allow_methods([Method::GET, Method::POST]) .allow_headers([http::header::CONTENT_TYPE]) .allow_origin(allow_origin); // Endpoint info let info = Info { model_id: model_info.model_id, model_sha: model_info.sha, // model_dtype: shard_info.dtype, // model_device_type: shard_info.device_type, model_pipeline_tag: model_info.pipeline_tag, max_concurrent_requests, max_best_of, max_stop_sequences, max_input_tokens, max_total_tokens, // waiting_served_ratio, // max_batch_total_tokens, // max_waiting_tokens, // max_batch_size, validation_workers, max_client_batch_size, router: env!("CARGO_PKG_NAME"), version: env!("CARGO_PKG_VERSION"), sha: option_env!("VERGEN_GIT_SHA"), docker_label: option_env!("DOCKER_LABEL"), }; #[allow(unused_mut)] // mut is needed for conditional compilation let mut doc = ApiDoc::openapi(); #[cfg(feature = "google")] { use crate::VertexInstance; #[derive(OpenApi)] #[openapi( paths(vertex_compatibility), components(schemas(VertexInstance, VertexRequest, VertexResponse)) )] struct VertexApiDoc; doc.merge(VertexApiDoc::openapi()); } #[cfg(feature = "kserve")] { use crate::kserve::{ InferenceOutput, InferenceRequest, LiveResponse, MetadataServerResponse, OutputChunk, ReadyResponse, }; use crate::kserve::{ __path_kerve_server_metadata, __path_kserve_health_live, __path_kserve_health_ready, __path_kserve_model_infer, __path_kserve_model_metadata, __path_kserve_model_metadata_ready, }; #[derive(OpenApi)] #[openapi( paths( kserve_health_live, kserve_health_ready, kerve_server_metadata, kserve_model_metadata, kserve_model_metadata_ready, kserve_model_infer, ), components(schemas( InferenceOutput, InferenceRequest, LiveResponse, MetadataServerResponse, OutputChunk, ReadyResponse, )) )] struct KServeApiDoc; doc.merge(KServeApiDoc::openapi()); } // Configure Swagger UI let swagger_ui = SwaggerUi::new("/docs").url("/api-doc/openapi.json", doc); // Define base and health routes let mut base_routes = Router::new() .route("/", post(compat_generate)) .route("/generate", post(generate)) .route("/generate_stream", post(generate_stream)) .route("/v1/chat/completions", post(chat_completions)) .route("/v1/completions", post(completions)) .route("/vertex", post(vertex_compatibility)) .route("/tokenize", post(tokenize)); if let Some(api_key) = api_key { let mut prefix = "Bearer ".to_string(); prefix.push_str(&api_key); // Leak to allow FnMut let api_key: &'static str = prefix.leak(); let auth = move |headers: HeaderMap, request: axum::extract::Request, next: axum::middleware::Next| async move { match headers.get(AUTHORIZATION) { Some(token) => match token.to_str() { Ok(token_str) if token_str.to_lowercase() == api_key.to_lowercase() => { let response = next.run(request).await; Ok(response) } _ => Err(StatusCode::UNAUTHORIZED), }, None => Err(StatusCode::UNAUTHORIZED), } }; base_routes = base_routes.layer(axum::middleware::from_fn(auth)) } let info_routes = Router::new() .route("/", get(health)) .route("/chat_tokenize", post(get_chat_tokenize)) .route("/info", get(get_model_info)) .route("/health", get(health)) .route("/ping", get(health)) .route("/metrics", get(metrics)); // Conditional AWS Sagemaker route let aws_sagemaker_route = if messages_api_enabled { Router::new().route("/invocations", post(chat_completions)) // Use 'chat_completions' for OAI_ENABLED } else { Router::new().route("/invocations", post(compat_generate)) // Use 'compat_generate' otherwise }; let compute_type = ComputeType(std::env::var("COMPUTE_TYPE").unwrap_or("gpu+optimized".to_string())); // Combine routes and layers let mut app = Router::new() .merge(swagger_ui) .merge(base_routes) .merge(info_routes) .merge(aws_sagemaker_route); #[cfg(feature = "google")] { tracing::info!("Built with `google` feature"); tracing::info!( "Environment variables `AIP_PREDICT_ROUTE` and `AIP_HEALTH_ROUTE` will be respected." ); if let Ok(env_predict_route) = std::env::var("AIP_PREDICT_ROUTE") { app = app.route(&env_predict_route, post(vertex_compatibility)); } if let Ok(env_health_route) = std::env::var("AIP_HEALTH_ROUTE") { app = app.route(&env_health_route, get(health)); } } #[cfg(feature = "kserve")] { tracing::info!("Built with `kserve` feature"); app = app .route( "/v2/models/:model_name/versions/:model_version/infer", post(kserve_model_infer), ) .route( "/v2/models/:model_name/versions/:model_version", get(kserve_model_metadata), ) .route("/v2/health/ready", get(kserve_health_ready)) .route("/v2/health/live", get(kserve_health_live)) .route("/v2", get(kerve_server_metadata)) .route( "/v2/models/:model_name/versions/:model_version/ready", get(kserve_model_metadata_ready), ); } // add layers after routes app = app .layer(Extension(info)) .layer(Extension(compat_return_full_text)) .layer(Extension(infer)) .layer(Extension(compute_type)) .layer(Extension(prom_handle.clone())) .layer(OtelAxumLayer::default()) .layer(cors_layer); tracing::info!("Connected"); if ngrok { #[cfg(feature = "ngrok")] { panic!("ngrok feature is not functional with axum=0.7 and hyper=1, waiting on https://github.com/ngrok/ngrok-rust/pull/137/files to re-enable."); // Run server } #[cfg(not(feature = "ngrok"))] { let _ngrok_authtoken = ngrok_authtoken; let _ngrok_domain = ngrok_domain; let _ngrok_username = ngrok_username; let _ngrok_password = ngrok_password; panic!("`text-generation-router` was compiled without the `ngrok` feature"); } } else { // Run server let listener = tokio::net::TcpListener::bind(&addr).await.unwrap(); axum::serve(listener, app) .with_graceful_shutdown(shutdown_signal()) .await .map_err(|err| WebServerError::Axum(Box::new(err)))?; } Ok(()) } /// get model info from the Huggingface Hub pub async fn get_hub_model_info(api: &ApiRepo) -> Option { let response = api.info_request().send().await.ok()?; if response.status().is_success() { let hub_model_info: HubModelInfo = serde_json::from_str(&response.text().await.ok()?).ok()?; if let Some(sha) = &hub_model_info.sha { tracing::info!( "Serving revision {sha} of model {}", hub_model_info.model_id ); } Some(hub_model_info) } else { None } } /// get base tokenizer pub async fn get_base_tokenizer(api: &Api, api_repo: &ApiRepo) -> Option { let config_filename = api_repo.get("config.json").await.ok()?; // Open the file in read-only mode with buffer. let file = File::open(config_filename).ok()?; let reader = BufReader::new(file); // Read the JSON contents of the file as an instance of `User`. let config: serde_json::Value = serde_json::from_reader(reader).ok()?; if let Some(serde_json::Value::String(base_model_id)) = config.get("base_model_name_or_path") { let api_base_repo = api.repo(Repo::with_revision( base_model_id.to_string(), RepoType::Model, "main".to_string(), )); api_base_repo.get("tokenizer.json").await.ok() } else { None } } /// get tokenizer_config from the Huggingface Hub pub async fn get_tokenizer_config(api_repo: &ApiRepo) -> Option { let tokenizer_config_filename = api_repo.get("tokenizer_config.json").await.ok()?; // Open the file in read-only mode with buffer. let file = File::open(tokenizer_config_filename).ok()?; let reader = BufReader::new(file); // Read the JSON contents of the file as an instance of 'HubTokenizerConfig'. let tokenizer_config: HubTokenizerConfig = serde_json::from_reader(reader) .map_err(|e| { tracing::warn!("Unable to parse tokenizer config: {}", e); e }) .ok()?; Some(tokenizer_config) } /// Shutdown signal handler async fn shutdown_signal() { let ctrl_c = async { signal::ctrl_c() .await .expect("failed to install Ctrl+C handler"); }; #[cfg(unix)] let terminate = async { signal::unix::signal(signal::unix::SignalKind::terminate()) .expect("failed to install signal handler") .recv() .await; }; #[cfg(not(unix))] let terminate = std::future::pending::<()>(); tokio::select! { _ = ctrl_c => {}, _ = terminate => {}, } tracing::info!("signal received, starting graceful shutdown"); opentelemetry::global::shutdown_tracer_provider(); } /// Convert to Axum supported formats impl From for (StatusCode, Json) { fn from(err: InferError) -> Self { let status_code = match err { InferError::GenerationError(_) => StatusCode::FAILED_DEPENDENCY, InferError::Overloaded(_) => StatusCode::TOO_MANY_REQUESTS, InferError::ValidationError(_) => StatusCode::UNPROCESSABLE_ENTITY, InferError::IncompleteGeneration => StatusCode::INTERNAL_SERVER_ERROR, InferError::TemplateError(_) => StatusCode::UNPROCESSABLE_ENTITY, InferError::MissingTemplateVariable(_) => StatusCode::UNPROCESSABLE_ENTITY, InferError::ToolError(_) => StatusCode::UNPROCESSABLE_ENTITY, }; ( status_code, Json(ErrorResponse { error: err.to_string(), error_type: err.error_type().to_string(), }), ) } } impl From for Event { fn from(err: InferError) -> Self { Event::default() .json_data(ErrorResponse { error: err.to_string(), error_type: err.error_type().to_string(), }) .unwrap() } } #[derive(Debug, Error)] pub enum WebServerError { #[error("Axum error: {0}")] Axum(#[from] axum::BoxError), } /// Create a post_processor for the LlamaTokenizer fn create_post_processor( tokenizer: &Tokenizer, tokenizer_config: &HubTokenizerConfig, ) -> Result { let add_bos_token = tokenizer_config.add_bos_token.unwrap_or(true); let add_eos_token = tokenizer_config.add_eos_token.unwrap_or(false); let bos_token = tokenizer_config.bos_token.as_ref(); let eos_token = tokenizer_config.eos_token.as_ref(); if add_bos_token && bos_token.is_none() { panic!("add_bos_token = true but bos_token is None"); } if add_eos_token && eos_token.is_none() { panic!("add_eos_token = true but eos_token is None"); } let mut single = Vec::new(); let mut pair = Vec::new(); let mut special_tokens = Vec::new(); if add_bos_token { if let Some(bos) = bos_token { let bos_token_id = tokenizer .token_to_id(bos.as_str()) .expect("Should have found the bos token id"); special_tokens.push((bos.as_str(), bos_token_id)); single.push(format!("{}:0", bos.as_str())); pair.push(format!("{}:0", bos.as_str())); } } single.push("$A:0".to_string()); pair.push("$A:0".to_string()); if add_eos_token { if let Some(eos) = eos_token { let eos_token_id = tokenizer .token_to_id(eos.as_str()) .expect("Should have found the eos token id"); special_tokens.push((eos.as_str(), eos_token_id)); single.push(format!("{}:0", eos.as_str())); pair.push(format!("{}:0", eos.as_str())); } } if add_bos_token { if let Some(bos) = bos_token { pair.push(format!("{}:1", bos.as_str())); } } pair.push("$B:1".to_string()); if add_eos_token { if let Some(eos) = eos_token { pair.push(format!("{}:1", eos.as_str())); } } let post_processor = TemplateProcessing::builder() .try_single(single)? .try_pair(pair)? .special_tokens(special_tokens) .build()?; Ok(post_processor) } type PreparedInput = (String, Option, Option); fn prepare_chat_input( infer: &Infer, response_format: Option, tools: Option>, tool_choice: ToolChoice, tool_prompt: &str, guideline: Option, messages: Vec, ) -> Result { if response_format.is_some() && tools.is_some() { return Err(InferError::ToolError( "Grammar and tools are mutually exclusive".into(), )); } if let Some(format) = response_format { let inputs = infer.apply_chat_template(guideline, messages, None)?; return Ok((inputs, Some(format), None)); } // if tools are set, apply the tool grammar and then the chat template let tool_grammar: Option = ToolGrammar::apply(tools, tool_choice)?; let grammar = tool_grammar .as_ref() .map(|t| GrammarType::Json(serde_json::json!(t))); let tools_grammar_prompt = tool_grammar .as_ref() .map(|t| (GrammarType::Json(serde_json::json!(t)), tool_prompt.into())); let inputs = infer.apply_chat_template(guideline, messages, tools_grammar_prompt)?; Ok((inputs, grammar, tool_grammar)) }