From 6a7a6b06617871389c523182dea07ef3ee82f47e Mon Sep 17 00:00:00 2001 From: OlivierDehaene <23298448+OlivierDehaene@users.noreply.github.com> Date: Wed, 3 May 2023 09:48:05 +0200 Subject: [PATCH] fix(server): fix typo in tokenizers decode --- server/text_generation_server/models/causal_lm.py | 2 +- server/text_generation_server/models/flash_causal_lm.py | 2 +- server/text_generation_server/models/flash_santacoder.py | 2 +- server/text_generation_server/models/galactica.py | 2 +- server/text_generation_server/models/santacoder.py | 2 +- 5 files changed, 5 insertions(+), 5 deletions(-) diff --git a/server/text_generation_server/models/causal_lm.py b/server/text_generation_server/models/causal_lm.py index 7dc7fb85..26a9a661 100644 --- a/server/text_generation_server/models/causal_lm.py +++ b/server/text_generation_server/models/causal_lm.py @@ -490,7 +490,7 @@ class CausalLM(Model): def decode(self, generated_ids: List[int]) -> str: return self.tokenizer.decode( - generated_ids, skip_special_tokens=True, cleanup_tokenization_spaces=False + generated_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False ) def forward( diff --git a/server/text_generation_server/models/flash_causal_lm.py b/server/text_generation_server/models/flash_causal_lm.py index 61ccca84..413866d1 100644 --- a/server/text_generation_server/models/flash_causal_lm.py +++ b/server/text_generation_server/models/flash_causal_lm.py @@ -402,7 +402,7 @@ class FlashCausalLM(Model): def decode(self, generated_ids: Union[torch.Tensor, List[int]]) -> str: return self.tokenizer.decode( - generated_ids, skip_special_tokens=True, cleanup_tokenization_spaces=False + generated_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False ) def forward( diff --git a/server/text_generation_server/models/flash_santacoder.py b/server/text_generation_server/models/flash_santacoder.py index aa1bdfb5..550be956 100644 --- a/server/text_generation_server/models/flash_santacoder.py +++ b/server/text_generation_server/models/flash_santacoder.py @@ -165,7 +165,7 @@ class FlashSantacoder(FlashCausalLM): def decode(self, generated_ids: List[int]) -> str: # Do not skip special tokens as they are used for custom parsing rules of the generated text return self.tokenizer.decode( - generated_ids, skip_special_tokens=False, cleanup_tokenization_spaces=False + generated_ids, skip_special_tokens=False, clean_up_tokenization_spaces=False ) diff --git a/server/text_generation_server/models/galactica.py b/server/text_generation_server/models/galactica.py index c1ec1566..78e9bfe4 100644 --- a/server/text_generation_server/models/galactica.py +++ b/server/text_generation_server/models/galactica.py @@ -173,7 +173,7 @@ class Galactica(OPT): def decode(self, generated_ids: List[int]) -> str: # Do not skip special tokens as they are used for custom parsing rules of the generated text return self.tokenizer.decode( - generated_ids, skip_special_tokens=False, cleanup_tokenization_spaces=False + generated_ids, skip_special_tokens=False, clean_up_tokenization_spaces=False ) def forward( diff --git a/server/text_generation_server/models/santacoder.py b/server/text_generation_server/models/santacoder.py index 796c33e3..a7b09a82 100644 --- a/server/text_generation_server/models/santacoder.py +++ b/server/text_generation_server/models/santacoder.py @@ -64,5 +64,5 @@ class SantaCoder(CausalLM): def decode(self, generated_ids: List[int]) -> str: # Do not skip special tokens as they are used for custom parsing rules of the generated text return self.tokenizer.decode( - generated_ids, skip_special_tokens=False, cleanup_tokenization_spaces=False + generated_ids, skip_special_tokens=False, clean_up_tokenization_spaces=False )