From 269792094b9fc11fa725c71189433a978f50493f Mon Sep 17 00:00:00 2001 From: Nicolas Patry Date: Mon, 4 Dec 2023 14:56:38 +0000 Subject: [PATCH] Cargo fmt --- launcher/src/main.rs | 1 - router/client/src/lib.rs | 2 +- router/src/infer.rs | 63 +++++++++++++++++++++------------------- 3 files changed, 34 insertions(+), 32 deletions(-) diff --git a/launcher/src/main.rs b/launcher/src/main.rs index 3545c005..4e230205 100644 --- a/launcher/src/main.rs +++ b/launcher/src/main.rs @@ -162,7 +162,6 @@ struct Args { #[clap(long, env)] speculate: Option, - /// The dtype to be forced upon the model. This option cannot be used with `--quantize`. #[clap(long, env, value_enum)] dtype: Option, diff --git a/router/client/src/lib.rs b/router/client/src/lib.rs index 0bb61568..1ea5e365 100644 --- a/router/client/src/lib.rs +++ b/router/client/src/lib.rs @@ -10,7 +10,7 @@ pub use pb::generate::v1::HealthResponse; pub use pb::generate::v1::InfoResponse as ShardInfo; pub use pb::generate::v1::{ Batch, CachedBatch, FinishReason, GeneratedText, Generation, NextTokenChooserParameters, - Tokens, Request, StoppingCriteriaParameters, + Request, StoppingCriteriaParameters, Tokens, }; pub use sharded_client::ShardedClient; use thiserror::Error; diff --git a/router/src/infer.rs b/router/src/infer.rs index 5972e67b..53d71f89 100644 --- a/router/src/infer.rs +++ b/router/src/infer.rs @@ -9,7 +9,7 @@ use std::sync::{ Arc, }; use text_generation_client::{ - Batch, CachedBatch, ClientError, GeneratedText, Generation, Tokens, ShardedClient, + Batch, CachedBatch, ClientError, GeneratedText, Generation, ShardedClient, Tokens, }; use thiserror::Error; use tokio::sync::mpsc::error::SendError; @@ -524,45 +524,48 @@ fn send_responses( } // Create last Token - let tokens: Vec = if let Some(tokens_) = generation.tokens{ - tokens_.ids.into_iter() - .zip(tokens_.logprobs.into_iter()) - .zip(tokens_.texts.into_iter()) - .zip(tokens_.is_special.into_iter()) - .map(|(((id, logprob), text), special)| Token { - id, - text, - logprob, - special, - }).collect() - }else{ + let tokens: Vec = if let Some(tokens_) = generation.tokens { + tokens_ + .ids + .into_iter() + .zip(tokens_.logprobs.into_iter()) + .zip(tokens_.texts.into_iter()) + .zip(tokens_.is_special.into_iter()) + .map(|(((id, logprob), text), special)| Token { + id, + text, + logprob, + special, + }) + .collect() + } else { vec![] }; // generation.top_tokens let mut top_tokens = Vec::new(); - for top_tokens_ in generation.top_tokens{ + for top_tokens_ in generation.top_tokens { let mut local_top_tokens = Vec::new(); - local_top_tokens.extend( - top_tokens_ - .ids - .into_iter() - .zip(top_tokens_.logprobs.into_iter()) - .zip(top_tokens_.texts.into_iter()) - .zip(top_tokens_.is_special.into_iter()) - .map(|(((id, logprob), text), special)| Token { - id, - text, - logprob, - special, - }), - ); + local_top_tokens.extend( + top_tokens_ + .ids + .into_iter() + .zip(top_tokens_.logprobs.into_iter()) + .zip(top_tokens_.texts.into_iter()) + .zip(top_tokens_.is_special.into_iter()) + .map(|(((id, logprob), text), special)| Token { + id, + text, + logprob, + special, + }), + ); top_tokens.push(local_top_tokens); } - // Force top_tokens to be the same size as tokens, both are going to be + // Force top_tokens to be the same size as tokens, both are going to be // zipped later - if top_tokens.len() != tokens.len(){ + if top_tokens.len() != tokens.len() { top_tokens = (0..tokens.len()).map(|_| Vec::new()).collect(); }