From ca20c304b3534605c28ebd86541a77a5e1e3ca12 Mon Sep 17 00:00:00 2001 From: OlivierDehaene <23298448+OlivierDehaene@users.noreply.github.com> Date: Wed, 10 Jan 2024 17:17:48 +0100 Subject: [PATCH] add log --- server/text_generation_server/models/flash_causal_lm.py | 1 + 1 file changed, 1 insertion(+) diff --git a/server/text_generation_server/models/flash_causal_lm.py b/server/text_generation_server/models/flash_causal_lm.py index 21ed4f6c..ab2d3313 100644 --- a/server/text_generation_server/models/flash_causal_lm.py +++ b/server/text_generation_server/models/flash_causal_lm.py @@ -779,6 +779,7 @@ class FlashCausalLM(Model): if os.getenv("ENABLE_CUDA_GRAPHS", "False") == "True": try: + logger.info("Experimental support for Cuda Graphs is enabled") # Warmup cuda graphs for all power of twos until 64 for i in range(6): self.cuda_graph_warmup(2**i, max_s, max_bt)