mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-25 20:12:07 +00:00
fix.
This commit is contained in:
parent
5c3efbc763
commit
82a6cb82e1
@ -1418,9 +1418,9 @@ class FlashCausalLM(Model):
|
||||
)
|
||||
max_total_tokens = available_blocks
|
||||
else:
|
||||
max_total_tokens = len(batch.input_ids)
|
||||
max_total_tokens = sum(len(input_ids) for input_ids in batch.input_ids)
|
||||
max_input_tokens = (
|
||||
batch.num_blocks - 1
|
||||
max_total_tokens - 1
|
||||
if max_input_tokens is None
|
||||
else max_input_tokens
|
||||
)
|
||||
|
Loading…
Reference in New Issue
Block a user