mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 04:14:52 +00:00
Fix imports.
This commit is contained in:
parent
ff51589332
commit
b041bf15ae
@ -45,15 +45,6 @@ __all__ = [
|
|||||||
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
|
FLASH_ATT_ERROR_MESSAGE = "{} requires Flash Attention enabled models."
|
||||||
|
|
||||||
FLASH_ATTENTION = True
|
FLASH_ATTENTION = True
|
||||||
from text_generation_server.models.flash_rw import FlashRWSharded
|
|
||||||
from text_generation_server.models.flash_neox import FlashNeoXSharded
|
|
||||||
from text_generation_server.models.flash_llama import (
|
|
||||||
FlashLlama,
|
|
||||||
)
|
|
||||||
from text_generation_server.models.flash_santacoder import (
|
|
||||||
FlashSantacoderSharded,
|
|
||||||
)
|
|
||||||
from text_generation_server.models.idefics import IDEFICSSharded
|
|
||||||
try:
|
try:
|
||||||
from text_generation_server.models.flash_rw import FlashRWSharded
|
from text_generation_server.models.flash_rw import FlashRWSharded
|
||||||
from text_generation_server.models.flash_neox import FlashNeoXSharded
|
from text_generation_server.models.flash_neox import FlashNeoXSharded
|
||||||
|
Loading…
Reference in New Issue
Block a user