Hotfix: pass through model revision in VlmCausalLM

This commit is contained in:
Daniël de Kok 2024-07-19 12:51:57 +00:00
parent 80adb5be16
commit c3e65f575a

View File

@ -261,7 +261,12 @@ class VlmCausalLM(FlashCausalLM):
**processor_kwargs,
)
self.batch_class = batch_class
super().__init__(model_id=model_id, **kwargs)
super().__init__(
model_id=model_id,
revision=revision,
trust_remote_code=trust_remote_code,
**kwargs,
)
@property
def batch_type(self) -> Type[VlmCausalLMBatch]: