mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-09 03:14:53 +00:00
Fix the import error
Signed-off-by: yuanwu <yuan.wu@intel.com>
This commit is contained in:
parent
cfcbd80fb4
commit
f0acbbf10c
@ -38,6 +38,7 @@ __all__ = [
|
|||||||
]
|
]
|
||||||
from text_generation_server.models.globals import ATTENTION
|
from text_generation_server.models.globals import ATTENTION
|
||||||
|
|
||||||
|
VLM_BATCH_TYPES = set()
|
||||||
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
|
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
|
||||||
|
|
||||||
FLASH_ATTENTION = False
|
FLASH_ATTENTION = False
|
||||||
@ -155,6 +156,9 @@ if FLASH_ATTENTION:
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
__all__.append(VLM_BATCH_TYPES)
|
||||||
|
|
||||||
|
|
||||||
class ModelType(enum.Enum):
|
class ModelType(enum.Enum):
|
||||||
DEEPSEEK_V2 = {
|
DEEPSEEK_V2 = {
|
||||||
"type": "deepseek_v2",
|
"type": "deepseek_v2",
|
||||||
|
Loading…
Reference in New Issue
Block a user