diff --git a/backends/gaudi/server/text_generation_server/models/custom_modeling/flash_llava_next.py b/backends/gaudi/server/text_generation_server/models/custom_modeling/flash_llava_next.py index c4d4f728..d884f413 100644 --- a/backends/gaudi/server/text_generation_server/models/custom_modeling/flash_llava_next.py +++ b/backends/gaudi/server/text_generation_server/models/custom_modeling/flash_llava_next.py @@ -12,7 +12,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -""" PyTorch Llava-NeXT model.""" +"""PyTorch Llava-NeXT model.""" from typing import List, Optional, Tuple diff --git a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics2.py b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics2.py index 41a45373..0579ca5d 100644 --- a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics2.py +++ b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics2.py @@ -12,7 +12,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -""" PyTorch Idefics2 model.""" +"""PyTorch Idefics2 model.""" from typing import List, Optional, Tuple diff --git a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics3.py b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics3.py index 6dd44c11..e12f2209 100644 --- a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics3.py +++ b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics3.py @@ -12,7 +12,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -""" PyTorch Idefics3 model.""" +"""PyTorch Idefics3 model.""" from typing import List, Optional, Tuple diff --git a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_config.py b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_config.py index a5565819..6ce2054e 100644 --- a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_config.py +++ b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_config.py @@ -17,7 +17,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -""" Idefics model configuration""" +"""Idefics model configuration""" import copy from transformers import PretrainedConfig diff --git a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_modeling.py b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_modeling.py index a130dbc1..910e9bcd 100644 --- a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_modeling.py +++ b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_modeling.py @@ -17,7 +17,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -""" PyTorch Idefics model.""" +"""PyTorch Idefics model.""" from typing import List, Optional, Tuple, Union import torch diff --git a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_vision.py b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_vision.py index dd8f76bc..7d2051e0 100644 --- a/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_vision.py +++ b/backends/gaudi/server/text_generation_server/models/custom_modeling/idefics_vision.py @@ -12,7 +12,7 @@ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. -""" PyTorch IdeficsVision model: a copy of CLIPVisionModel using a simpler config object""" +"""PyTorch IdeficsVision model: a copy of CLIPVisionModel using a simpler config object""" from dataclasses import dataclass diff --git a/backends/gaudi/server/text_generation_server/models/flash_causal_lm.py b/backends/gaudi/server/text_generation_server/models/flash_causal_lm.py index e380ed53..fd26a5ef 100644 --- a/backends/gaudi/server/text_generation_server/models/flash_causal_lm.py +++ b/backends/gaudi/server/text_generation_server/models/flash_causal_lm.py @@ -1798,7 +1798,9 @@ class FlashCausalLM(Model): total_batch_seq = 0.001 total_mem = 0 available_mem = free_mem - self.mem_reserved - log_master(logger.info, f"Decode batch size list:{[bsz[0] for bsz in buckets]}\n") + log_master( + logger.info, f"Decode batch size list:{[bsz[0] for bsz in buckets]}\n" + ) for i, (batch_size, block_num) in enumerate(buckets): if batch_size > block_num: continue diff --git a/backends/gaudi/server/text_generation_server/models/flash_vlm_causal_lm.py b/backends/gaudi/server/text_generation_server/models/flash_vlm_causal_lm.py index fe28d067..ee6495cd 100644 --- a/backends/gaudi/server/text_generation_server/models/flash_vlm_causal_lm.py +++ b/backends/gaudi/server/text_generation_server/models/flash_vlm_causal_lm.py @@ -822,7 +822,9 @@ class FlashVlmCausalLM(FlashCausalLM): total_batch_seq = 0.001 total_mem = 0 available_mem = decode_available_memory - log_master(logger.info, f"Decode batch size list:{[bsz[0] for bsz in buckets]}\n") + log_master( + logger.info, f"Decode batch size list:{[bsz[0] for bsz in buckets]}\n" + ) for i, (batch_size, block_num) in enumerate(buckets): if batch_size > block_num: continue diff --git a/backends/gaudi/server/text_generation_server/models/mllama_causal_lm.py b/backends/gaudi/server/text_generation_server/models/mllama_causal_lm.py index ec9e149b..995b456f 100644 --- a/backends/gaudi/server/text_generation_server/models/mllama_causal_lm.py +++ b/backends/gaudi/server/text_generation_server/models/mllama_causal_lm.py @@ -442,7 +442,9 @@ class FlashMllamaCausalLM(FlashVlmCausalLM): total_batch_seq = 0.001 total_mem = 0 available_mem = free_mem - self.mem_reserved - log_master(logger.info, f"Decode batch size list:{[bsz[0] for bsz in buckets]}\n") + log_master( + logger.info, f"Decode batch size list:{[bsz[0] for bsz in buckets]}\n" + ) for i, (batch_size, block_num) in enumerate(buckets): if batch_size > block_num: continue diff --git a/backends/gaudi/server/text_generation_server/utils/segments.py b/backends/gaudi/server/text_generation_server/utils/segments.py index f5961102..133049be 100644 --- a/backends/gaudi/server/text_generation_server/utils/segments.py +++ b/backends/gaudi/server/text_generation_server/utils/segments.py @@ -8,7 +8,7 @@ import torch def find_segments( - adapter_indices: Union[torch.Tensor, List[int]] + adapter_indices: Union[torch.Tensor, List[int]], ) -> Tuple[List[int], List[int]]: segments = [0] segment_indices = []