diff --git a/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py b/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py index f7980d2d..3e8e67ab 100644 --- a/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py +++ b/server/text_generation_server/models/custom_modeling/flash_llama_modeling.py @@ -424,7 +424,7 @@ class FlashLlamaModel(torch.nn.Module): FlashLlamaLayer( index=0, prefix=( - "model.layers.0" if not prefix else "{prefix}.model.layers.0" + "model.layers.0" if not prefix else f"{prefix}.model.layers.0" ), config=config, weights=weights,