diff --git a/Dockerfile_gaudi b/Dockerfile_gaudi index 06073fe4..2c77d483 100644 --- a/Dockerfile_gaudi +++ b/Dockerfile_gaudi @@ -96,6 +96,7 @@ RUN cd server && \ BUILD_CUDA_EXT=0 pip install git+https://github.com/AutoGPTQ/AutoGPTQ.git@097dd04e --no-build-isolation && \ pip install . --no-cache-dir RUN pip install git+https://github.com/sywangyi/vllm-hpu-extension.git +RUN pip install git+https://github.com/huggingface/transformers@29338949 # Install benchmarker COPY --from=builder /usr/src/target/release-opt/text-generation-benchmark /usr/local/bin/text-generation-benchmark # Install router diff --git a/backends/gaudi/server/requirements.txt b/backends/gaudi/server/requirements.txt index 0a6d2fbc..2431077c 100644 --- a/backends/gaudi/server/requirements.txt +++ b/backends/gaudi/server/requirements.txt @@ -90,7 +90,7 @@ threadpoolctl==3.6.0 ; python_version >= "3.9" and python_version < "3.13" tokenizers==0.21.1 ; python_version >= "3.9" and python_version < "3.13" torch==2.6.0 ; python_version >= "3.9" and python_version < "3.13" tqdm==4.67.1 ; python_version >= "3.9" and python_version < "3.13" -transformer @ git+https://github.com/huggingface/transformers@29338949 ; python_version >= "3.9" and python_version < "3.13" +transformers @ git+https://github.com/huggingface/transformers@29338949 ; python_version >= "3.9" and python_version < "3.13" triton==3.2.0 ; python_version >= "3.9" and python_version < "3.13" and platform_system == "Linux" and platform_machine == "x86_64" typer==0.15.2 ; python_version >= "3.9" and python_version < "3.13" typing-extensions==4.13.2 ; python_version >= "3.9" and python_version < "3.13"