Updating model link.

This commit is contained in:
Nicolas Patry 2024-09-30 11:53:21 +02:00
parent af677caf4f
commit 933060cc3f
No known key found for this signature in database
GPG Key ID: 64AF4752B2967863
2 changed files with 2 additions and 1 deletions

View File

@ -35,6 +35,7 @@ Text Generation Inference enables serving optimized models on specific hardware
- [Gpt Neox](https://huggingface.co/EleutherAI/gpt-neox-20b) - [Gpt Neox](https://huggingface.co/EleutherAI/gpt-neox-20b)
- [Gptj](https://huggingface.co/EleutherAI/gpt-j-6b) - [Gptj](https://huggingface.co/EleutherAI/gpt-j-6b)
- [Idefics](https://huggingface.co/HuggingFaceM4/idefics-9b) (Multimodal) - [Idefics](https://huggingface.co/HuggingFaceM4/idefics-9b) (Multimodal)
- [Mllama](https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct) (Multimodal)
If the above list lacks the model you would like to serve, depending on the model's pipeline type, you can try to initialize and serve the model anyways to see how well it performs, but performance isn't guaranteed for non-optimized models: If the above list lacks the model you would like to serve, depending on the model's pipeline type, you can try to initialize and serve the model anyways to see how well it performs, but performance isn't guaranteed for non-optimized models:

View File

@ -324,7 +324,7 @@ class ModelType(enum.Enum):
MLLAMA = { MLLAMA = {
"type": "mllama", "type": "mllama",
"name": "Mllama", "name": "Mllama",
"url": "https://huggingface.co/xxx/xx", "url": "https://huggingface.co/meta-llama/Llama-3.2-11B-Vision-Instruct",
"multimodal": True, "multimodal": True,
} }