From f82ae76dffd9044563ce570f11c8ba4f077077d0 Mon Sep 17 00:00:00 2001 From: fxmarty <9808326+fxmarty@users.noreply.github.com> Date: Fri, 17 May 2024 16:51:31 +0000 Subject: [PATCH] add back warning --- server/text_generation_server/models/__init__.py | 1 + 1 file changed, 1 insertion(+) diff --git a/server/text_generation_server/models/__init__.py b/server/text_generation_server/models/__init__.py index ff75a635..9e5676f5 100644 --- a/server/text_generation_server/models/__init__.py +++ b/server/text_generation_server/models/__init__.py @@ -84,6 +84,7 @@ try: HAS_FLASH_ATTN_V2_ROCM, ) except ImportError as e: + logger.warning(f"Could not import Flash Attention enabled models: {e}") FLASH_ATTENTION = False HAS_FLASH_ATTN_V2_CUDA = False HAS_FLASH_ATTN_V2_ROCM = False