mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-06-18 15:22:09 +00:00
fix: prefer patch to be vlm specific
This commit is contained in:
parent
b32cd97b71
commit
783ca66926
@ -1267,15 +1267,6 @@ class FlashCausalLM(Model):
|
|||||||
|
|
||||||
prefix = None
|
prefix = None
|
||||||
model = model_class(prefix, config, weights)
|
model = model_class(prefix, config, weights)
|
||||||
|
|
||||||
if model.config.vocab_size != tokenizer.vocab_size:
|
|
||||||
logger.warning(
|
|
||||||
f"Tokenizer vocab size {tokenizer.vocab_size} does not match model vocab size {model.config.vocab_size}. Updating tokenizer vocab size."
|
|
||||||
)
|
|
||||||
# TODO: HUGE HACK! This is a workaround for the fact that Qwen2TokenizerFast
|
|
||||||
# returns the incorrect vocab size for the 2B model.
|
|
||||||
tokenizer._vocab_size = model.config.vocab_size
|
|
||||||
|
|
||||||
torch.distributed.barrier(group=self.process_group)
|
torch.distributed.barrier(group=self.process_group)
|
||||||
|
|
||||||
# VLM models define the config we care about in their text_config
|
# VLM models define the config we care about in their text_config
|
||||||
|
@ -414,6 +414,15 @@ class VlmCausalLM(FlashCausalLM):
|
|||||||
**kwargs,
|
**kwargs,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if self.config.vocab_size != self.tokenizer.vocab_size:
|
||||||
|
logger.warning(
|
||||||
|
f"Tokenizer vocab size {self.tokenizer.vocab_size} does not match model vocab size {self.config.vocab_size}. Updating tokenizer vocab size."
|
||||||
|
)
|
||||||
|
# TODO: HUGE HACK! This is a workaround to update the vocab size
|
||||||
|
# in the tokenizer. When the tokenizer is updated within the model
|
||||||
|
# the vocab size is not updated in the tokenizer.
|
||||||
|
self.tokenizer._vocab_size = self.config.vocab_size
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def batch_type(self) -> Type[VlmCausalLMBatch]:
|
def batch_type(self) -> Type[VlmCausalLMBatch]:
|
||||||
return self.batch_class
|
return self.batch_class
|
||||||
|
Loading…
Reference in New Issue
Block a user