mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 12:24:53 +00:00
feat: support loading gemma2 as vlm text model
This commit is contained in:
parent
5df8059037
commit
0f0fe9a998
@ -17,6 +17,12 @@ def load_text_model(prefix, config, weights, name=None):
|
|||||||
)
|
)
|
||||||
|
|
||||||
return FlashGemmaForCausalLM(prefix, config, weights, causal=False)
|
return FlashGemmaForCausalLM(prefix, config, weights, causal=False)
|
||||||
|
elif config.model_type == "gemma2":
|
||||||
|
from text_generation_server.models.custom_modeling.flash_gemma2_modeling import (
|
||||||
|
FlashGemma2ForCausalLM,
|
||||||
|
)
|
||||||
|
|
||||||
|
return FlashGemma2ForCausalLM(prefix, config, weights)
|
||||||
elif config.model_type == "paligemma":
|
elif config.model_type == "paligemma":
|
||||||
from text_generation_server.models.custom_modeling.flash_gemma_modeling import (
|
from text_generation_server.models.custom_modeling.flash_gemma_modeling import (
|
||||||
FlashGemmaForCausalLM,
|
FlashGemmaForCausalLM,
|
||||||
|
Loading…
Reference in New Issue
Block a user