mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-12 04:44:52 +00:00
fix: adds causal to attention params to check when using flash attn v1
This commit is contained in:
parent
8a7749b8fb
commit
519e5ac05b
@ -293,6 +293,7 @@ else:
|
|||||||
max_s,
|
max_s,
|
||||||
softmax_scale,
|
softmax_scale,
|
||||||
window_size_left=-1,
|
window_size_left=-1,
|
||||||
|
causal=None,
|
||||||
softcap=None,
|
softcap=None,
|
||||||
):
|
):
|
||||||
if window_size_left != -1:
|
if window_size_left != -1:
|
||||||
|
Loading…
Reference in New Issue
Block a user