mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 12:24:53 +00:00
fix: add back typo removed variable
This commit is contained in:
parent
3e22cdd14c
commit
66f89120b5
@ -43,6 +43,8 @@ __all__ = [
|
||||
|
||||
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
|
||||
|
||||
FLASH_ATTENTION = True
|
||||
|
||||
# FlashCausalLM reqiures CUDA Graphs to be enabled on the system. This will throw a RuntimeError
|
||||
# if CUDA Graphs are not available when calling `torch.cuda.graph_pool_handle()` in the FlashCausalLM
|
||||
HAS_CUDA_GRAPH = False
|
||||
|
Loading…
Reference in New Issue
Block a user