mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 12:24:53 +00:00
feat: add tokenizer-config-path to launcher args
This commit is contained in:
parent
650fea1834
commit
f660fcf715
@ -371,6 +371,10 @@ struct Args {
|
|||||||
/// Display a lot of information about your runtime environment
|
/// Display a lot of information about your runtime environment
|
||||||
#[clap(long, short, action)]
|
#[clap(long, short, action)]
|
||||||
env: bool,
|
env: bool,
|
||||||
|
|
||||||
|
/// path to the tokenizer config file
|
||||||
|
#[clap(long, env)]
|
||||||
|
tokenizer_config_path: Option<String>,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
@ -1016,6 +1020,12 @@ fn spawn_webserver(
|
|||||||
args.model_id,
|
args.model_id,
|
||||||
];
|
];
|
||||||
|
|
||||||
|
// Tokenizer config path
|
||||||
|
if let Some(ref tokenizer_config_path) = args.tokenizer_config_path {
|
||||||
|
router_args.push("--tokenizer-config-path".to_string());
|
||||||
|
router_args.push(tokenizer_config_path.to_string());
|
||||||
|
}
|
||||||
|
|
||||||
// Model optional max batch total tokens
|
// Model optional max batch total tokens
|
||||||
if let Some(max_batch_total_tokens) = args.max_batch_total_tokens {
|
if let Some(max_batch_total_tokens) = args.max_batch_total_tokens {
|
||||||
router_args.push("--max-batch-total-tokens".to_string());
|
router_args.push("--max-batch-total-tokens".to_string());
|
||||||
|
Loading…
Reference in New Issue
Block a user