Add docker run --help flag + add curl example in docker_launch

This commit is contained in:
osanseviero 2023-08-09 23:14:23 +02:00
parent 38652c9da5
commit 744d78d88e

View File

@ -17,8 +17,18 @@ To use GPUs, you need to install the [NVIDIA Container Toolkit](https://docs.nvi
</Tip> </Tip>
Once TGI is running, you can use the `generate` endpoint by doing requests. To learn more about how to query the endpoints, check the [Consuming TGI](./basic_tutorials/consuming_tgi) section.
```shell
curl 127.0.0.1:8080/generate -X POST -d '{"inputs":"What is Deep Learning?","parameters":{"max_new_tokens":20}}' -H 'Content-Type: application/json'
```
<Tip>
To see all possible flags and options, you can use the `--help` flag. It's possible to configure the number of shards, quantization, generation parameters, and more. To see all possible flags and options, you can use the `--help` flag. It's possible to configure the number of shards, quantization, generation parameters, and more.
``` ```shell
docker run ghcr.io/huggingface/text-generation-inference:1.0.0 --help docker run ghcr.io/huggingface/text-generation-inference:1.0.0 --help
``` ```
</Tip>