From e01e1b7ca6c9206626190b7f78eb1af831ff0160 Mon Sep 17 00:00:00 2001 From: drbh Date: Thu, 8 Aug 2024 01:35:42 +0000 Subject: [PATCH] fix: run lints --- integration-tests/models/test_opt.py | 2 +- .../models/custom_modeling/opt_modeling.py | 10 +++++++--- 2 files changed, 8 insertions(+), 4 deletions(-) diff --git a/integration-tests/models/test_opt.py b/integration-tests/models/test_opt.py index edf89b51..cbeb6376 100644 --- a/integration-tests/models/test_opt.py +++ b/integration-tests/models/test_opt.py @@ -16,4 +16,4 @@ async def opt_sharded(opt_sharded_handle): @pytest.mark.release @pytest.mark.asyncio async def test_opt(opt_sharded): - pass \ No newline at end of file + pass diff --git a/server/text_generation_server/models/custom_modeling/opt_modeling.py b/server/text_generation_server/models/custom_modeling/opt_modeling.py index 747a9841..b0d05f43 100644 --- a/server/text_generation_server/models/custom_modeling/opt_modeling.py +++ b/server/text_generation_server/models/custom_modeling/opt_modeling.py @@ -98,7 +98,9 @@ class OPTLearnedPositionalEmbedding(nn.Module): super().__init__() self.offset = 2 self.weight = nn.Parameter( - weights.get_tensor(f"{prefix and prefix + '.'}decoder.embed_positions.weight") + weights.get_tensor( + f"{prefix and prefix + '.'}decoder.embed_positions.weight" + ) ) def forward( @@ -437,7 +439,7 @@ class OPTDecoder(OPTPreTrainedModel): self.max_target_positions = config.max_position_embeddings self.vocab_size = config.vocab_size - prefix = prefix and prefix + '.' + prefix = prefix and prefix + "." self.embed_tokens = TensorParallelEmbedding( prefix=f"{prefix}decoder.embed_tokens", weights=weights @@ -757,7 +759,9 @@ class OPTForCausalLM(OPTPreTrainedModel): self.model = OPTModel(prefix, config, weights) self.lm_head = SpeculativeHead.load( - config, prefix=f"{prefix and prefix + '.'}decoder.embed_tokens", weights=weights + config, + prefix=f"{prefix and prefix + '.'}decoder.embed_tokens", + weights=weights, ) def forward(