From f8521900601578a070ab5bb4275cbb2cd45b8e01 Mon Sep 17 00:00:00 2001 From: drbh Date: Thu, 8 Aug 2024 14:08:56 -0400 Subject: [PATCH] fix: prefer hidden_activation over hidden_act in gemma2 (#2381) --- .../models/custom_modeling/flash_gemma2_modeling.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/server/text_generation_server/models/custom_modeling/flash_gemma2_modeling.py b/server/text_generation_server/models/custom_modeling/flash_gemma2_modeling.py index de86f514..54d212e6 100644 --- a/server/text_generation_server/models/custom_modeling/flash_gemma2_modeling.py +++ b/server/text_generation_server/models/custom_modeling/flash_gemma2_modeling.py @@ -265,7 +265,7 @@ class FlashGemma2Attention(torch.nn.Module): class Gemma2MLP(nn.Module): def __init__(self, prefix, config, weights): super().__init__() - act = config.hidden_act + act = config.hidden_activation self.act = ( ACT2FN[act] if "gelu" not in act