diff --git a/backends/gaudi/server/text_generation_server/models/custom_modeling/qwen2_5_vl.py b/backends/gaudi/server/text_generation_server/models/custom_modeling/qwen2_5_vl.py index 441b0016..5f180458 100644 --- a/backends/gaudi/server/text_generation_server/models/custom_modeling/qwen2_5_vl.py +++ b/backends/gaudi/server/text_generation_server/models/custom_modeling/qwen2_5_vl.py @@ -49,7 +49,8 @@ from text_generation_server.models.custom_modeling.flash_qwen2_modeling import ( # Copied from: https://github.com/huggingface/transformers/blob/main/src/transformers/models/qwen2_5_vl/processing_qwen2_5_vl.py from typing import Union from transformers.feature_extraction_utils import BatchFeature -from transformers.image_utils import ImageInput, VideoInput +from transformers.image_utils import ImageInput +from transformers.video_utils import VideoInput from transformers.processing_utils import ( ProcessingKwargs, ProcessorMixin, diff --git a/backends/gaudi/tgi-entrypoint.sh b/backends/gaudi/tgi-entrypoint.sh index d787ea8e..3153870d 100644 --- a/backends/gaudi/tgi-entrypoint.sh +++ b/backends/gaudi/tgi-entrypoint.sh @@ -12,7 +12,7 @@ if [[ "$ATTENTION" == "paged" ]]; then # Check if Llama-4 is in the command line arguments if [[ "$*" == *"Llama-4"* || "$*" == *"Qwen3"* ]]; then echo 'ATTENTION=paged and Llama-4 or Qwen3 detected' - pip install git+https://github.com/huggingface/transformers.git@29338949 + pip install transformers==4.52.1 fi fi