2025-02-14 12:40:57 +00:00
|
|
|
mod backend;
|
2025-03-11 08:19:01 +00:00
|
|
|
mod llamacpp;
|
|
|
|
mod quantize;
|
|
|
|
|
|
|
|
use quantize::QuantizeType;
|
2025-02-14 12:40:57 +00:00
|
|
|
|
|
|
|
use backend::{
|
|
|
|
BackendError, LlamacppBackend, LlamacppConfig, LlamacppGGMLType, LlamacppNuma,
|
|
|
|
LlamacppSplitMode,
|
|
|
|
};
|
|
|
|
use clap::Parser;
|
2025-03-11 08:19:01 +00:00
|
|
|
use hf_hub::api::tokio::ApiBuilder;
|
|
|
|
use hf_hub::{Repo, RepoType};
|
|
|
|
use std::path::Path;
|
2025-02-14 12:40:57 +00:00
|
|
|
use text_generation_router::{logging, server, usage_stats};
|
|
|
|
use thiserror::Error;
|
2025-03-11 08:19:01 +00:00
|
|
|
use tokenizers::Tokenizer;
|
|
|
|
use tokio::process::Command;
|
2025-02-14 12:40:57 +00:00
|
|
|
use tokio::sync::oneshot::error::RecvError;
|
|
|
|
use tracing::{error, warn};
|
|
|
|
|
|
|
|
/// Backend Configuration
|
|
|
|
#[derive(Parser, Debug)]
|
|
|
|
#[clap(author, version, about, long_about = None)]
|
|
|
|
struct Args {
|
|
|
|
/// Name of the model to load.
|
|
|
|
#[clap(long, env)]
|
|
|
|
model_id: String,
|
|
|
|
|
|
|
|
/// Revision of the model.
|
|
|
|
#[clap(default_value = "main", long, env)]
|
|
|
|
revision: String,
|
|
|
|
|
|
|
|
/// Path to the GGUF model file for inference.
|
|
|
|
#[clap(long, env)]
|
2025-03-11 08:19:01 +00:00
|
|
|
model_gguf: Option<String>,
|
2025-02-14 12:40:57 +00:00
|
|
|
|
|
|
|
/// Number of threads to use for generation.
|
|
|
|
#[clap(long, env)]
|
|
|
|
n_threads: Option<usize>,
|
|
|
|
|
|
|
|
/// Number of threads to use for batch processing.
|
|
|
|
#[clap(long, env)]
|
|
|
|
n_threads_batch: Option<usize>,
|
|
|
|
|
|
|
|
/// Number of layers to store in VRAM.
|
|
|
|
#[clap(default_value = "0", long, env)]
|
|
|
|
n_gpu_layers: usize,
|
|
|
|
|
|
|
|
/// Split the model across multiple GPUs.
|
|
|
|
#[clap(default_value = "layer", long, env)]
|
|
|
|
split_mode: LlamacppSplitMode,
|
|
|
|
|
|
|
|
/// Defragment the KV cache if holes/size > threshold.
|
|
|
|
#[clap(default_value = "-1.0", long, env)]
|
|
|
|
defrag_threshold: f32,
|
|
|
|
|
|
|
|
/// Enable NUMA optimizations.
|
|
|
|
#[clap(default_value = "disabled", value_enum, long, env)]
|
|
|
|
numa: LlamacppNuma,
|
|
|
|
|
|
|
|
/// Use memory mapping for the model.
|
|
|
|
#[clap(long, env)]
|
2025-03-11 08:19:01 +00:00
|
|
|
disable_mmap: bool,
|
2025-02-14 12:40:57 +00:00
|
|
|
|
|
|
|
/// Use memory locking to prevent swapping.
|
|
|
|
#[clap(long, env)]
|
|
|
|
use_mlock: bool,
|
|
|
|
|
|
|
|
/// Enable offloading of KQV operations to the GPU.
|
|
|
|
#[clap(long, env)]
|
2025-03-11 08:19:01 +00:00
|
|
|
disable_offload_kqv: bool,
|
2025-02-14 12:40:57 +00:00
|
|
|
|
|
|
|
/// Enable flash attention for faster inference. (EXPERIMENTAL)
|
|
|
|
#[clap(long, env)]
|
2025-03-11 08:19:01 +00:00
|
|
|
disable_flash_attention: bool,
|
2025-02-14 12:40:57 +00:00
|
|
|
|
|
|
|
/// Data type used for K cache.
|
|
|
|
#[clap(default_value = "f16", value_enum, long, env)]
|
|
|
|
type_k: LlamacppGGMLType,
|
|
|
|
|
|
|
|
/// Data type used for V cache.
|
|
|
|
#[clap(default_value = "f16", value_enum, long, env)]
|
|
|
|
type_v: LlamacppGGMLType,
|
|
|
|
|
|
|
|
/// Number of tokenizer workers used for payload validation and truncation.
|
|
|
|
#[clap(default_value = "2", long, env)]
|
|
|
|
validation_workers: usize,
|
|
|
|
|
|
|
|
/// Maximum number of concurrent requests.
|
|
|
|
#[clap(long, env)]
|
|
|
|
max_concurrent_requests: Option<usize>,
|
|
|
|
|
|
|
|
/// Maximum number of input tokens per request.
|
|
|
|
#[clap(default_value = "1024", long, env)]
|
|
|
|
max_input_tokens: usize,
|
|
|
|
|
|
|
|
/// Maximum number of total tokens (input + output) per request.
|
|
|
|
#[clap(default_value = "2048", long, env)]
|
|
|
|
max_total_tokens: usize,
|
|
|
|
|
|
|
|
/// Maximum number of tokens in a batch.
|
|
|
|
#[clap(long, env)]
|
|
|
|
max_batch_total_tokens: Option<usize>,
|
|
|
|
|
|
|
|
/// Maximum number of tokens in a physical batch.
|
|
|
|
#[clap(long, env)]
|
|
|
|
max_physical_batch_total_tokens: Option<usize>,
|
|
|
|
|
|
|
|
/// Maximum number of requests per batch.
|
|
|
|
#[clap(long, env)]
|
|
|
|
max_batch_size: Option<usize>,
|
|
|
|
|
|
|
|
/// IP address to listen on.
|
|
|
|
#[clap(default_value = "0.0.0.0", long)]
|
|
|
|
hostname: String,
|
|
|
|
|
|
|
|
/// Port to listen on.
|
|
|
|
#[clap(default_value = "3000", long, short, env)]
|
|
|
|
port: u16,
|
|
|
|
|
|
|
|
/// Enable JSON output format.
|
|
|
|
#[clap(long, env)]
|
|
|
|
json_output: bool,
|
|
|
|
|
|
|
|
/// OTLP endpoint for telemetry data.
|
|
|
|
#[clap(long, env)]
|
|
|
|
otlp_endpoint: Option<String>,
|
|
|
|
|
|
|
|
/// Service name for OTLP telemetry.
|
|
|
|
#[clap(default_value = "text-generation-inference.router", long, env)]
|
|
|
|
otlp_service_name: String,
|
|
|
|
|
|
|
|
/// Allowed origins for CORS.
|
|
|
|
#[clap(long, env)]
|
|
|
|
cors_allow_origin: Option<Vec<String>>,
|
|
|
|
|
|
|
|
/// Path to the tokenizer configuration file.
|
|
|
|
#[clap(long, env)]
|
|
|
|
tokenizer_config_path: Option<String>,
|
|
|
|
|
|
|
|
/// Disable grammar support.
|
|
|
|
#[clap(long, env)]
|
|
|
|
disable_grammar_support: bool,
|
|
|
|
|
|
|
|
/// Maximum number of inputs per request.
|
|
|
|
#[clap(default_value = "4", long, env)]
|
|
|
|
max_client_batch_size: usize,
|
|
|
|
|
|
|
|
/// Level of usage statistics collection.
|
|
|
|
#[clap(default_value = "on", long, env)]
|
|
|
|
usage_stats: usage_stats::UsageStatsLevel,
|
|
|
|
|
|
|
|
/// Maximum payload size in bytes.
|
|
|
|
#[clap(default_value = "2000000", long, env)]
|
|
|
|
payload_limit: usize,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[tokio::main]
|
|
|
|
async fn main() -> Result<(), RouterError> {
|
|
|
|
let args = Args::parse();
|
|
|
|
|
|
|
|
logging::init_logging(args.otlp_endpoint, args.otlp_service_name, args.json_output);
|
|
|
|
|
|
|
|
let n_threads = match args.n_threads {
|
|
|
|
Some(0) | None => num_cpus::get(),
|
|
|
|
Some(threads) => threads,
|
|
|
|
};
|
|
|
|
let n_threads_batch = match args.n_threads_batch {
|
|
|
|
Some(0) | None => n_threads,
|
|
|
|
Some(threads) => threads,
|
|
|
|
};
|
|
|
|
let max_batch_size = match args.max_batch_size {
|
|
|
|
Some(0) | None => n_threads_batch,
|
|
|
|
Some(threads) => threads,
|
|
|
|
};
|
|
|
|
let max_batch_total_tokens = match args.max_batch_total_tokens {
|
|
|
|
None => max_batch_size * args.max_total_tokens,
|
|
|
|
Some(size) => size,
|
|
|
|
};
|
|
|
|
let max_physical_batch_total_tokens = match args.max_physical_batch_total_tokens {
|
|
|
|
None => max_batch_total_tokens,
|
|
|
|
Some(size) => size,
|
|
|
|
};
|
|
|
|
let max_concurrent_requests = match args.max_concurrent_requests {
|
|
|
|
None => max_batch_size * 2,
|
|
|
|
Some(size) => size,
|
|
|
|
};
|
|
|
|
if args.max_input_tokens >= args.max_total_tokens {
|
|
|
|
return Err(RouterError::ArgumentValidation(
|
|
|
|
"`max_input_tokens` must be < `max_total_tokens`".to_string(),
|
|
|
|
));
|
|
|
|
}
|
|
|
|
if args.max_total_tokens > max_batch_total_tokens {
|
|
|
|
return Err(RouterError::ArgumentValidation(
|
|
|
|
"`max_total_tokens` must be <= `max_batch_total_tokens`".to_string(),
|
|
|
|
));
|
|
|
|
}
|
|
|
|
if max_batch_size * args.max_total_tokens > max_batch_total_tokens {
|
|
|
|
return Err(RouterError::ArgumentValidation(
|
|
|
|
"`max_batch_size` * `max_total_tokens` must be <= `max_batch_total_tokens`".to_string(),
|
|
|
|
));
|
|
|
|
}
|
|
|
|
|
2025-03-11 08:19:01 +00:00
|
|
|
let api_builder = || {
|
|
|
|
let mut builder = ApiBuilder::new().with_progress(true);
|
|
|
|
|
|
|
|
if let Ok(cache_dir) = std::env::var("HUGGINGFACE_HUB_CACHE") {
|
|
|
|
builder = builder.with_cache_dir(cache_dir.into());
|
|
|
|
}
|
|
|
|
if let Ok(token) = std::env::var("HF_TOKEN") {
|
|
|
|
builder = builder.with_token(token.into());
|
|
|
|
}
|
|
|
|
if let Ok(origin) = std::env::var("HF_HUB_USER_AGENT_ORIGIN") {
|
|
|
|
builder = builder.with_user_agent("origin", origin.as_str());
|
|
|
|
}
|
|
|
|
builder
|
|
|
|
};
|
|
|
|
let api_repo = api_builder().build()?.repo(Repo::with_revision(
|
|
|
|
args.model_id.clone(),
|
|
|
|
RepoType::Model,
|
|
|
|
args.revision.clone(),
|
|
|
|
));
|
|
|
|
|
|
|
|
let tokenizer_path = api_repo.get("tokenizer.json").await?;
|
|
|
|
let tokenizer = Tokenizer::from_file(&tokenizer_path)?;
|
|
|
|
|
|
|
|
let model_gguf = if let Some(model_gguf) = args.model_gguf {
|
|
|
|
model_gguf
|
|
|
|
} else {
|
|
|
|
let model_gguf = format!("models/{}/model.gguf", args.model_id);
|
|
|
|
let model_gguf_path = Path::new(&model_gguf);
|
|
|
|
|
|
|
|
if !model_gguf_path.exists() {
|
|
|
|
let tmp_gguf = "models/tmp.gguf";
|
|
|
|
|
|
|
|
if let Some(parent) = Path::new(model_gguf_path).parent() {
|
|
|
|
std::fs::create_dir_all(parent)?;
|
|
|
|
}
|
|
|
|
let cache_path = tokenizer_path.parent().unwrap();
|
|
|
|
|
|
|
|
for sibling in api_repo.info().await?.siblings {
|
|
|
|
let _ = api_repo.get(&sibling.rfilename).await?;
|
|
|
|
}
|
|
|
|
let status = Command::new("convert_hf_to_gguf.py")
|
|
|
|
.arg("--outfile")
|
|
|
|
.arg(tmp_gguf)
|
|
|
|
.arg(cache_path)
|
|
|
|
.spawn()?
|
|
|
|
.wait()
|
|
|
|
.await?;
|
|
|
|
|
|
|
|
if !status.success() {
|
|
|
|
let exit_code = status.code().unwrap_or(-1);
|
|
|
|
error!("Failed to generate GGUF, exit code: {}", exit_code);
|
|
|
|
return Err(RouterError::CommandError(exit_code));
|
|
|
|
}
|
|
|
|
quantize::model(tmp_gguf, &model_gguf, QuantizeType::MostlyQ4_0, n_threads)
|
|
|
|
.map_err(RouterError::QuantizeError)?;
|
|
|
|
}
|
|
|
|
model_gguf
|
2025-02-14 12:40:57 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
let (backend, ok, shutdown) = LlamacppBackend::new(
|
|
|
|
LlamacppConfig {
|
2025-03-11 08:19:01 +00:00
|
|
|
model_gguf,
|
2025-02-14 12:40:57 +00:00
|
|
|
n_threads,
|
|
|
|
n_threads_batch,
|
|
|
|
n_gpu_layers: args.n_gpu_layers,
|
|
|
|
split_mode: args.split_mode,
|
|
|
|
defrag_threshold: args.defrag_threshold,
|
|
|
|
numa: args.numa,
|
2025-03-11 08:19:01 +00:00
|
|
|
use_mmap: !args.disable_mmap,
|
2025-02-14 12:40:57 +00:00
|
|
|
use_mlock: args.use_mlock,
|
2025-03-11 08:19:01 +00:00
|
|
|
flash_attention: !args.disable_flash_attention,
|
2025-02-14 12:40:57 +00:00
|
|
|
type_k: args.type_k,
|
|
|
|
type_v: args.type_v,
|
2025-03-11 08:19:01 +00:00
|
|
|
offload_kqv: !args.disable_offload_kqv,
|
2025-02-14 12:40:57 +00:00
|
|
|
max_batch_total_tokens,
|
|
|
|
max_physical_batch_total_tokens,
|
|
|
|
max_batch_size,
|
|
|
|
batch_timeout: tokio::time::Duration::from_millis(5),
|
|
|
|
},
|
|
|
|
tokenizer,
|
|
|
|
);
|
|
|
|
ok.await??;
|
|
|
|
|
|
|
|
if cfg!(debug_assertions) {
|
|
|
|
warn!("Graceful shutdown disabled!");
|
|
|
|
let _ = tokio::task::spawn(async move {
|
|
|
|
let _ = tokio::signal::ctrl_c().await;
|
|
|
|
let _ = shutdown.send(true);
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
|
|
|
server::run(
|
|
|
|
backend,
|
|
|
|
max_concurrent_requests,
|
|
|
|
0, // max_best_of
|
|
|
|
0, // max_stop_sequences
|
|
|
|
0, // max_top_n_tokens
|
|
|
|
args.max_input_tokens,
|
|
|
|
args.max_total_tokens,
|
|
|
|
args.validation_workers,
|
|
|
|
None, // api_key
|
|
|
|
args.model_id, // tokenizer_name
|
|
|
|
args.tokenizer_config_path,
|
|
|
|
Some(args.revision),
|
|
|
|
false, // trust_remote_code
|
|
|
|
args.hostname,
|
|
|
|
args.port,
|
|
|
|
args.cors_allow_origin,
|
|
|
|
false, // ngrok,
|
|
|
|
None, // ngrok_authtoken,
|
|
|
|
None, // ngrok_edge,
|
|
|
|
args.disable_grammar_support,
|
|
|
|
args.max_client_batch_size,
|
|
|
|
args.usage_stats,
|
|
|
|
args.payload_limit,
|
|
|
|
)
|
|
|
|
.await?;
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Debug, Error)]
|
|
|
|
enum RouterError {
|
|
|
|
#[error("Argument validation error: {0}")]
|
|
|
|
ArgumentValidation(String),
|
|
|
|
#[error("Tokenizer error: {0}")]
|
|
|
|
Tokenizer(#[from] tokenizers::Error),
|
|
|
|
#[error("Backend error: {0}")]
|
|
|
|
Backend(#[from] BackendError),
|
|
|
|
#[error("WebServer error: {0}")]
|
|
|
|
WebServer(#[from] server::WebServerError),
|
|
|
|
#[error("Recv error: {0}")]
|
|
|
|
RecvError(#[from] RecvError),
|
2025-03-11 08:19:01 +00:00
|
|
|
#[error("Io error: {0}")]
|
|
|
|
IoError(#[from] std::io::Error),
|
|
|
|
#[error("Var error: {0}")]
|
|
|
|
VarError(#[from] std::env::VarError),
|
|
|
|
#[error("Quantize error: {0}")]
|
|
|
|
QuantizeError(String),
|
|
|
|
#[error("Command error: {0}")]
|
|
|
|
CommandError(i32),
|
|
|
|
#[error("HF hub error: {0}")]
|
|
|
|
HubError(#[from] hf_hub::api::tokio::ApiError),
|
2025-02-14 12:40:57 +00:00
|
|
|
}
|