mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 12:24:53 +00:00
fix: improve text model loading
This commit is contained in:
parent
34174af8c8
commit
064e040ee3
@ -723,7 +723,7 @@ class Idefics3ForConditionalGeneration(nn.Module):
|
||||
|
||||
vision_config = config.vision_config
|
||||
self.text_model = load_text_model(
|
||||
prefix="model" if not prefix else f"{prefix}.model",
|
||||
prefix=f"{prefix}.model.text_model" if prefix else "model.text_model",
|
||||
config=config.text_config,
|
||||
weights=weights,
|
||||
name="text_model",
|
||||
|
@ -4,7 +4,7 @@ def load_text_model(prefix, config, weights, name=None):
|
||||
FlashLlamaForCausalLM,
|
||||
)
|
||||
|
||||
return FlashLlamaForCausalLM(f"{prefix}.text_model", config, weights)
|
||||
return FlashLlamaForCausalLM(prefix, config, weights)
|
||||
elif config.model_type == "mistral":
|
||||
from text_generation_server.models.custom_modeling.flash_mistral_modeling import (
|
||||
FlashMistralForCausalLM,
|
||||
|
Loading…
Reference in New Issue
Block a user