mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-06-19 15:52:08 +00:00
fix: update docs for model addition
This commit is contained in:
parent
3e41ec28c7
commit
7372a0dc38
@ -272,7 +272,7 @@ Options:
|
|||||||
--huggingface-hub-cache <HUGGINGFACE_HUB_CACHE>
|
--huggingface-hub-cache <HUGGINGFACE_HUB_CACHE>
|
||||||
The location of the huggingface hub cache. Used to override the location if you want to provide a mounted disk for instance
|
The location of the huggingface hub cache. Used to override the location if you want to provide a mounted disk for instance
|
||||||
|
|
||||||
[env: HUGGINGFACE_HUB_CACHE=]
|
[env: HUGGINGFACE_HUB_CACHE=/nvme0n1/Models/]
|
||||||
|
|
||||||
```
|
```
|
||||||
## WEIGHTS_CACHE_OVERRIDE
|
## WEIGHTS_CACHE_OVERRIDE
|
||||||
@ -349,12 +349,6 @@ Options:
|
|||||||
--cors-allow-origin <CORS_ALLOW_ORIGIN>
|
--cors-allow-origin <CORS_ALLOW_ORIGIN>
|
||||||
[env: CORS_ALLOW_ORIGIN=]
|
[env: CORS_ALLOW_ORIGIN=]
|
||||||
|
|
||||||
```
|
|
||||||
## API_KEY
|
|
||||||
```shell
|
|
||||||
--api-key <API_KEY>
|
|
||||||
[env: API_KEY=]
|
|
||||||
|
|
||||||
```
|
```
|
||||||
## WATERMARK_GAMMA
|
## WATERMARK_GAMMA
|
||||||
```shell
|
```shell
|
||||||
@ -430,20 +424,6 @@ Options:
|
|||||||
|
|
||||||
[env: LORA_ADAPTERS=]
|
[env: LORA_ADAPTERS=]
|
||||||
|
|
||||||
```
|
|
||||||
## USAGE_STATS
|
|
||||||
```shell
|
|
||||||
--usage-stats <USAGE_STATS>
|
|
||||||
Control if anonymous usage stats are collected. Options are "on", "off" and "no-stack" Defaul is on
|
|
||||||
|
|
||||||
[env: USAGE_STATS=]
|
|
||||||
[default: on]
|
|
||||||
|
|
||||||
Possible values:
|
|
||||||
- on: Default option, usage statistics are collected anonymously
|
|
||||||
- off: Disables all collection of usage statistics
|
|
||||||
- no-stack: Doesn't send the error stack trace or error type, but allows sending a crash event
|
|
||||||
|
|
||||||
```
|
```
|
||||||
## HELP
|
## HELP
|
||||||
```shell
|
```shell
|
||||||
|
@ -32,6 +32,7 @@ Text Generation Inference enables serving optimized models on specific hardware
|
|||||||
- [Mpt](https://huggingface.co/mosaicml/mpt-7b-instruct)
|
- [Mpt](https://huggingface.co/mosaicml/mpt-7b-instruct)
|
||||||
- [Gpt2](https://huggingface.co/openai-community/gpt2)
|
- [Gpt2](https://huggingface.co/openai-community/gpt2)
|
||||||
- [Gpt Neox](https://huggingface.co/EleutherAI/gpt-neox-20b)
|
- [Gpt Neox](https://huggingface.co/EleutherAI/gpt-neox-20b)
|
||||||
|
- [Gptj](https://huggingface.co/EleutherAI/gpt-j-6b)
|
||||||
- [Idefics](https://huggingface.co/HuggingFaceM4/idefics-9b) (Multimodal)
|
- [Idefics](https://huggingface.co/HuggingFaceM4/idefics-9b) (Multimodal)
|
||||||
|
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user