From 4ddea01c6e4f057caf2233fe1e9e82d57732bb21 Mon Sep 17 00:00:00 2001 From: OlivierDehaene <23298448+OlivierDehaene@users.noreply.github.com> Date: Mon, 7 Oct 2024 12:11:50 +0200 Subject: [PATCH] remove log --- server/text_generation_server/models/flash_causal_lm.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/server/text_generation_server/models/flash_causal_lm.py b/server/text_generation_server/models/flash_causal_lm.py index 65552ff7..4c3024f4 100644 --- a/server/text_generation_server/models/flash_causal_lm.py +++ b/server/text_generation_server/models/flash_causal_lm.py @@ -1712,8 +1712,6 @@ class FlashCausalLM(Model): if prefill: batch.prepare_for_prefill() - log_master(logger.info, f"Tokens in this forward: {len(batch.input_ids)}") - prefill_logprobs = batch.prefill_next_token_indices is not None # Update adapter indices for speculative tokens (if present)