mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-06-19 15:52:08 +00:00
use all tokens
This commit is contained in:
parent
b5233f9c3c
commit
70637b4170
@ -355,7 +355,7 @@ class FlashCausalLM(Model):
|
|||||||
# Generated token
|
# Generated token
|
||||||
next_token_logprob = logprobs[-1, next_token_id_item]
|
next_token_logprob = logprobs[-1, next_token_id_item]
|
||||||
next_token_text, offset, token_offset = self.decode_token(
|
next_token_text, offset, token_offset = self.decode_token(
|
||||||
all_input_ids[-(stopping_criteria.current_tokens + 1) :],
|
all_input_ids,
|
||||||
offset,
|
offset,
|
||||||
token_offset,
|
token_offset,
|
||||||
)
|
)
|
||||||
|
Loading…
Reference in New Issue
Block a user