mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-24 00:12:08 +00:00
misc(backend): decrease log verbosity in callback
This commit is contained in:
parent
1149186794
commit
52208f5b78
@ -17,7 +17,7 @@ use tokenizers::Tokenizer;
|
|||||||
use tokio::sync::mpsc::{unbounded_channel, UnboundedSender};
|
use tokio::sync::mpsc::{unbounded_channel, UnboundedSender};
|
||||||
use tokio::time::Instant;
|
use tokio::time::Instant;
|
||||||
use tokio_stream::wrappers::UnboundedReceiverStream;
|
use tokio_stream::wrappers::UnboundedReceiverStream;
|
||||||
use tracing::{error, info};
|
use tracing::{debug, error, info};
|
||||||
|
|
||||||
type InferResult = Result<InferStreamResponse, InferError>;
|
type InferResult = Result<InferStreamResponse, InferError>;
|
||||||
|
|
||||||
@ -113,7 +113,7 @@ fn llama_generate_callback(
|
|||||||
is_final: bool,
|
is_final: bool,
|
||||||
n_generated_tokens: usize,
|
n_generated_tokens: usize,
|
||||||
) -> bool {
|
) -> bool {
|
||||||
info!("Generated token: {new_token_id} -> logits={new_token_logit}, is_final={is_final} ({n_generated_tokens})");
|
debug!("Generated token: {new_token_id} -> logits={new_token_logit}, is_final={is_final} ({n_generated_tokens})");
|
||||||
|
|
||||||
let ctx = unsafe { &mut *ctx };
|
let ctx = unsafe { &mut *ctx };
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user