mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 04:14:52 +00:00
fix error if top_n_tokens is 0 or null
This commit is contained in:
parent
fd8b42678d
commit
1f7042d165
@ -719,7 +719,7 @@ class CausalLM(Model):
|
||||
[next_token_id_squeezed.item() in self.all_special_ids],
|
||||
),
|
||||
generated_text,
|
||||
[top_tokens],
|
||||
[top_tokens] if top_tokens is not None else None,
|
||||
)
|
||||
|
||||
generations.append(generation)
|
||||
|
@ -1089,7 +1089,7 @@ class FlashCausalLM(Model):
|
||||
[nid in self.all_special_ids for nid in _next_token_ids],
|
||||
),
|
||||
generated_text,
|
||||
[top_tokens],
|
||||
[top_tokens] if top_tokens is not None else None,
|
||||
)
|
||||
|
||||
generations.append(generation)
|
||||
|
@ -809,7 +809,7 @@ class IdeficsCausalLM(Model):
|
||||
[next_token_id_squeezed.item() in self.all_special_ids],
|
||||
),
|
||||
generated_text,
|
||||
[top_tokens],
|
||||
[top_tokens] if top_tokens is not None else None,
|
||||
)
|
||||
|
||||
generations.append(generation)
|
||||
|
@ -773,7 +773,7 @@ class Seq2SeqLM(Model):
|
||||
[next_token_id_squeezed.item() in self.all_special_ids],
|
||||
),
|
||||
generated_text,
|
||||
[top_tokens],
|
||||
[top_tokens] if top_tokens is not None else None,
|
||||
)
|
||||
|
||||
generations.append(generation)
|
||||
|
Loading…
Reference in New Issue
Block a user