mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 12:24:53 +00:00
fix: add back typo removed variable
This commit is contained in:
parent
3e22cdd14c
commit
66f89120b5
@ -43,6 +43,8 @@ __all__ = [
|
|||||||
|
|
||||||
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
|
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
|
||||||
|
|
||||||
|
FLASH_ATTENTION = True
|
||||||
|
|
||||||
# FlashCausalLM reqiures CUDA Graphs to be enabled on the system. This will throw a RuntimeError
|
# FlashCausalLM reqiures CUDA Graphs to be enabled on the system. This will throw a RuntimeError
|
||||||
# if CUDA Graphs are not available when calling `torch.cuda.graph_pool_handle()` in the FlashCausalLM
|
# if CUDA Graphs are not available when calling `torch.cuda.graph_pool_handle()` in the FlashCausalLM
|
||||||
HAS_CUDA_GRAPH = False
|
HAS_CUDA_GRAPH = False
|
||||||
|
Loading…
Reference in New Issue
Block a user