From 291722cb48516aaab9e3559cc15036dcc33de6af Mon Sep 17 00:00:00 2001 From: OlivierDehaene <23298448+OlivierDehaene@users.noreply.github.com> Date: Thu, 15 Dec 2022 16:59:37 +0100 Subject: [PATCH] remove print --- server/text_generation/models/seq2seq_lm.py | 1 - 1 file changed, 1 deletion(-) diff --git a/server/text_generation/models/seq2seq_lm.py b/server/text_generation/models/seq2seq_lm.py index 065ecbff..e51ce60b 100644 --- a/server/text_generation/models/seq2seq_lm.py +++ b/server/text_generation/models/seq2seq_lm.py @@ -448,7 +448,6 @@ class Seq2SeqLM(Model): token_ids = decoder_input_ids[-new_decoder_input_length:] output_text = self.tokenizer.decode(token_ids, skip_special_tokens=True) tokens = self.tokenizer.batch_decode(token_ids) - print(tokens) # Add NaN for the bos token logprobs = [float("nan")] + decoder_logprobs[ -new_decoder_input_length: