From ad942a1d79fffb76c70473669e0ddb59c2f1bc9e Mon Sep 17 00:00:00 2001 From: drbh Date: Fri, 2 Aug 2024 21:46:58 +0000 Subject: [PATCH] fix: avoid changing conditional --- server/text_generation_server/layers/attention/cuda.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/server/text_generation_server/layers/attention/cuda.py b/server/text_generation_server/layers/attention/cuda.py index c84cc7da..7e42eddf 100644 --- a/server/text_generation_server/layers/attention/cuda.py +++ b/server/text_generation_server/layers/attention/cuda.py @@ -257,7 +257,7 @@ else: window_size_left=-1, softcap=None, ): - if window_size_left is not None and window_size_left != -1: + if window_size_left != -1: raise NotImplementedError( "window_size_left is only available with flash attn v2" )