mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 12:24:53 +00:00
fix error if top_n_tokens is 0 or null
This commit is contained in:
parent
fd8b42678d
commit
1f7042d165
@ -719,7 +719,7 @@ class CausalLM(Model):
|
|||||||
[next_token_id_squeezed.item() in self.all_special_ids],
|
[next_token_id_squeezed.item() in self.all_special_ids],
|
||||||
),
|
),
|
||||||
generated_text,
|
generated_text,
|
||||||
[top_tokens],
|
[top_tokens] if top_tokens is not None else None,
|
||||||
)
|
)
|
||||||
|
|
||||||
generations.append(generation)
|
generations.append(generation)
|
||||||
|
@ -1089,7 +1089,7 @@ class FlashCausalLM(Model):
|
|||||||
[nid in self.all_special_ids for nid in _next_token_ids],
|
[nid in self.all_special_ids for nid in _next_token_ids],
|
||||||
),
|
),
|
||||||
generated_text,
|
generated_text,
|
||||||
[top_tokens],
|
[top_tokens] if top_tokens is not None else None,
|
||||||
)
|
)
|
||||||
|
|
||||||
generations.append(generation)
|
generations.append(generation)
|
||||||
|
@ -809,7 +809,7 @@ class IdeficsCausalLM(Model):
|
|||||||
[next_token_id_squeezed.item() in self.all_special_ids],
|
[next_token_id_squeezed.item() in self.all_special_ids],
|
||||||
),
|
),
|
||||||
generated_text,
|
generated_text,
|
||||||
[top_tokens],
|
[top_tokens] if top_tokens is not None else None,
|
||||||
)
|
)
|
||||||
|
|
||||||
generations.append(generation)
|
generations.append(generation)
|
||||||
|
@ -773,7 +773,7 @@ class Seq2SeqLM(Model):
|
|||||||
[next_token_id_squeezed.item() in self.all_special_ids],
|
[next_token_id_squeezed.item() in self.all_special_ids],
|
||||||
),
|
),
|
||||||
generated_text,
|
generated_text,
|
||||||
[top_tokens],
|
[top_tokens] if top_tokens is not None else None,
|
||||||
)
|
)
|
||||||
|
|
||||||
generations.append(generation)
|
generations.append(generation)
|
||||||
|
Loading…
Reference in New Issue
Block a user