From 996473164a2a101f4e0b8f5ccb69f06fc482a193 Mon Sep 17 00:00:00 2001 From: Mohit Sharma Date: Wed, 30 Apr 2025 09:57:22 +0000 Subject: [PATCH] fix qwen test --- server/text_generation_server/models/vlm_causal_lm.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/server/text_generation_server/models/vlm_causal_lm.py b/server/text_generation_server/models/vlm_causal_lm.py index 8ac3d65a..09cdb8a9 100644 --- a/server/text_generation_server/models/vlm_causal_lm.py +++ b/server/text_generation_server/models/vlm_causal_lm.py @@ -886,6 +886,7 @@ class VlmCausalLM(FlashCausalLM): ) def encode_images(self, batch): + image_grid_thw = None if batch.pixel_values is not None: device = batch.input_ids.device for request_id, image_id, image_input in batch.pixel_values: @@ -923,7 +924,7 @@ class VlmCausalLM(FlashCausalLM): batch.pixel_values = None batch.pixel_attention_mask = None batch.image_sizes = None - batch.image_grid_thw = None + batch.image_grid_thw = image_grid_thw def set_inputs_embeds(self, batch): if batch.has_image_inputs: @@ -1066,6 +1067,7 @@ class VlmCausalLM(FlashCausalLM): ) if batch.prefill_cache_indices is not None: batch.prefill_cache_indices = None + batch.image_grid_thw = None batch.free_encoder_cache() return logits, speculative_logits