mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-10 20:04:52 +00:00
Update and rename custom_models.md to non_core_models.md
This commit is contained in:
parent
7dcd953969
commit
b33a66148c
@ -1,4 +1,4 @@
|
|||||||
# Custom Model Serving
|
# Non-core Model Serving
|
||||||
|
|
||||||
TGI supports various LLM architectures (see full list [here](https://github.com/huggingface/text-generation-inference#optimized-architectures)). If you wish to serve a model that is not one of the supported models, TGI will fallback to transformers implementation of that model. They can be loaded by:
|
TGI supports various LLM architectures (see full list [here](https://github.com/huggingface/text-generation-inference#optimized-architectures)). If you wish to serve a model that is not one of the supported models, TGI will fallback to transformers implementation of that model. They can be loaded by:
|
||||||
|
|
||||||
@ -18,4 +18,4 @@ You can serve these models using docker like below 👇
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker run --gpus all --shm-size 1g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:latest --model-id gpt2
|
docker run --gpus all --shm-size 1g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:latest --model-id gpt2
|
||||||
```
|
```
|
Loading…
Reference in New Issue
Block a user