From f6324ffb3a81218fde18f994a9bc1898b6ce3c22 Mon Sep 17 00:00:00 2001 From: erikkaum Date: Thu, 25 Jul 2024 17:31:04 +0200 Subject: [PATCH] delete the last no repeat processor from warpers --- server/text_generation_server/utils/logits_process.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/server/text_generation_server/utils/logits_process.py b/server/text_generation_server/utils/logits_process.py index 452691a2..2c1531fe 100644 --- a/server/text_generation_server/utils/logits_process.py +++ b/server/text_generation_server/utils/logits_process.py @@ -18,7 +18,6 @@ from transformers import ( TopKLogitsWarper, TopPLogitsWarper, TypicalLogitsWarper, - NoRepeatNGramLogitsProcessor ) mempool = torch.cuda.graph_pool_handle() if torch.cuda.is_available() else None @@ -44,8 +43,6 @@ class StaticWarper: self.warpers.append(TopPLogitsWarper(top_p=top_p)) if typical_p is not None and typical_p < 1.0: self.warpers.append(TypicalLogitsWarper(mass=typical_p)) - if no_repeat_ngram_size is not None and no_repeat_ngram_size > 0: - self.warpers.append(NoRepeatNGramLogitsProcessor(no_repeat_ngram_size)) self.cuda_graph = None self.static_scores = None