Fix the import error

Signed-off-by: yuanwu <yuan.wu@intel.com>
This commit is contained in:
yuanwu 2025-05-11 23:05:37 +00:00
parent cfcbd80fb4
commit f0acbbf10c

View File

@ -38,6 +38,7 @@ __all__ = [
]
from text_generation_server.models.globals import ATTENTION
VLM_BATCH_TYPES = set()
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
FLASH_ATTENTION = False
@ -155,6 +156,9 @@ if FLASH_ATTENTION:
}
__all__.append(VLM_BATCH_TYPES)
class ModelType(enum.Enum):
DEEPSEEK_V2 = {
"type": "deepseek_v2",