From 22dbc449cd0b8f9c93cd43a405fe60252eb464d4 Mon Sep 17 00:00:00 2001 From: Nicolas Patry Date: Tue, 10 Dec 2024 12:15:13 +0100 Subject: [PATCH] Fixing max_position_embeddings for falcon. --- .../models/custom_modeling/flash_rw_modeling.py | 1 + 1 file changed, 1 insertion(+) diff --git a/server/text_generation_server/models/custom_modeling/flash_rw_modeling.py b/server/text_generation_server/models/custom_modeling/flash_rw_modeling.py index 2dcd1bf3..fbf1a597 100644 --- a/server/text_generation_server/models/custom_modeling/flash_rw_modeling.py +++ b/server/text_generation_server/models/custom_modeling/flash_rw_modeling.py @@ -78,6 +78,7 @@ class RWConfig(PretrainedConfig): self.alibi = False self.rotary = True self.rope_theta = rope_theta + self.max_position_embeddings = 2048 self.vocab_size = vocab_size # Backward compatibility with n_embed kwarg