mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-26 12:32:10 +00:00
nix: experimental support for building a Docker image
Run using something like: ``` docker run \ --device nvidia.com/gpu=all \ -it --rm -p 8080:80 \ -v $PWD/data:/data \ -v $PWD/tmp:/tmp \ tgi-docker:latest \ --model-id <model_id> ```
This commit is contained in:
parent
5b6b74e21d
commit
77a36d45eb
34
flake.nix
34
flake.nix
@ -141,15 +141,31 @@
|
||||
};
|
||||
};
|
||||
|
||||
packages.default = pkgs.writeShellApplication {
|
||||
name = "text-generation-inference";
|
||||
runtimeInputs = [
|
||||
server
|
||||
router
|
||||
];
|
||||
text = ''
|
||||
${launcher}/bin/text-generation-launcher "$@"
|
||||
'';
|
||||
packages = rec {
|
||||
default = pkgs.writeShellApplication {
|
||||
name = "text-generation-inference";
|
||||
runtimeInputs = [
|
||||
server
|
||||
router
|
||||
];
|
||||
text = ''
|
||||
${launcher}/bin/text-generation-launcher "$@"
|
||||
'';
|
||||
};
|
||||
|
||||
dockerImage = pkgs.dockerTools.buildLayeredImage {
|
||||
name = "tgi-docker";
|
||||
tag = "latest";
|
||||
config = {
|
||||
EntryPoint = [ "${default}/bin/text-generation-inference" ];
|
||||
Env = [
|
||||
"HF_HOME=/data"
|
||||
"PORT=80"
|
||||
];
|
||||
|
||||
};
|
||||
contents = [ pkgs.cacert ];
|
||||
};
|
||||
};
|
||||
}
|
||||
);
|
||||
|
Loading…
Reference in New Issue
Block a user