mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-06-19 15:52:08 +00:00
disable chunking for qwen
This commit is contained in:
parent
dd91b60998
commit
15926210d3
@ -1522,6 +1522,8 @@ def get_model(
|
||||
kv_cache_dtype=kv_cache_dtype,
|
||||
trust_remote_code=trust_remote_code,
|
||||
lora_adapter_ids=lora_adapter_ids,
|
||||
# TODO: Fix bug in rust image_text_replacement implementation
|
||||
support_chunking=False,
|
||||
)
|
||||
# TODO: Uncomment when transformers is refactored
|
||||
# elif FLASH_TRANSFORMERS_BACKEND:
|
||||
@ -1553,6 +1555,8 @@ def get_model(
|
||||
lora_adapter_ids=lora_adapter_ids,
|
||||
config_class=Qwen2_5_VLConfig,
|
||||
processor_class=Qwen2_5_VLProcessor,
|
||||
# TODO: Fix bug in rust image_text_replacement implementation
|
||||
support_chunking=False,
|
||||
)
|
||||
# TODO: Uncomment when transformers is refactored
|
||||
# elif FLASH_TRANSFORMERS_BACKEND:
|
||||
|
Loading…
Reference in New Issue
Block a user