mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-21 23:12:07 +00:00
feat: update readme and container version
This commit is contained in:
parent
d4077f70db
commit
a6890cbea9
12
README.md
12
README.md
@ -68,6 +68,18 @@ Text Generation Inference (TGI) is a toolkit for deploying and serving Large Lan
|
||||
|
||||
## Get Started
|
||||
|
||||
### Quick Start ⚡️
|
||||
|
||||
The fastest way to get started is to use the quickstart script. This script simplifies the docker and nvidia container toolkit installation process. It also installs the latest version of the text-generation-inference container and runs it with a default model.
|
||||
|
||||
```bash
|
||||
curl --proto '=https' --tlsv1.2 -sSf \
|
||||
https://raw.githubusercontent.com/huggingface/text-generation-inference/quickstart.sh \
|
||||
| bash
|
||||
```
|
||||
 Always review the contents of a script before running it.
|
||||
|
||||
|
||||
### Docker
|
||||
|
||||
For a detailed starting guide, please see the [Quick Tour](https://huggingface.co/docs/text-generation-inference/quicktour). The easiest way of getting started is using the official Docker container:
|
||||
|
@ -35,4 +35,4 @@ model="HuggingFaceH4/zephyr-7b-beta"
|
||||
volume="$PWD/data"
|
||||
|
||||
# Run the Docker container in interactive mode to allow CTRL+C to stop the container
|
||||
docker run -it --gpus all --shm-size 1g -p 8080:80 -v "$volume:/data" ghcr.io/huggingface/text-generation-inference:2.0 --model-id "$model"
|
||||
docker run -it --gpus all --shm-size 1g -p 8080:80 -v "$volume:/data" ghcr.io/huggingface/text-generation-inference:latest --model-id "$model"
|
||||
|
Loading…
Reference in New Issue
Block a user