From 999eba8096b38045ca8208be82fbbccefe0590d1 Mon Sep 17 00:00:00 2001 From: Mohit Sharma Date: Wed, 11 Dec 2024 12:10:35 +0000 Subject: [PATCH] (nit) remove log --- server/text_generation_server/layers/attention/rocm.py | 1 - 1 file changed, 1 deletion(-) diff --git a/server/text_generation_server/layers/attention/rocm.py b/server/text_generation_server/layers/attention/rocm.py index a401b589..0cfac25b 100644 --- a/server/text_generation_server/layers/attention/rocm.py +++ b/server/text_generation_server/layers/attention/rocm.py @@ -139,7 +139,6 @@ def paged_attention( max_logits = torch.empty_like(exp_sums) if not use_custom: - logger.info("Using PagedAttention V2") ops.paged_attention_v2( out, exp_sums,