From 4c8f5cdc359fd9167c9ecdb226ba0fd9415ed67a Mon Sep 17 00:00:00 2001 From: drbh Date: Mon, 23 Dec 2024 14:39:55 +0000 Subject: [PATCH] fix: lint with ruff --- .../models/custom_modeling/flash_llama_modeling.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py b/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py index 43c5dfb4..20bab01b 100644 --- a/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py +++ b/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py @@ -647,7 +647,7 @@ class FlashLlamaForCausalLM(torch.nn.Module): ) self.model = FlashLlamaModel(prefix, config, weights) if config.tie_word_embeddings: - suffix = f"model.embed_tokens" + suffix = "model.embed_tokens" else: suffix = "lm_head"