mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 04:14:52 +00:00
Fix flash attention ?
This commit is contained in:
parent
5b736e6d48
commit
bcf98a8b81
@ -1,4 +1,4 @@
|
|||||||
flash_att_commit := 3a9bfd076f98746c73362328958dbc68d145fbec
|
flash_att_commit := 1a7f4dfa9e51f6a90177a3244a5bc9c687894cdd
|
||||||
|
|
||||||
build-flash-attention:
|
build-flash-attention:
|
||||||
if [ ! -d 'flash-attention' ]; then \
|
if [ ! -d 'flash-attention' ]; then \
|
||||||
|
Loading…
Reference in New Issue
Block a user