mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-19 22:02:06 +00:00
* Putting back the NCCL forced upgrade. * . * ... * Ignoring conda. * Dropping conda from the buidl system + torch 2.6 * Cache min. * Rolling back torch version. * Reverting the EETQ modification. * Fix flash attention ? * Actually stay on flash v1. * Patching flash v1. * Torch 2.6, fork of rotary, eetq updated. * Put back nccl latest (override torch). * Slightly more reproducible build and not as scary.
14 lines
370 B
Plaintext
14 lines
370 B
Plaintext
eetq_commit := 465e9726bf7ae30803a2d0dd9e5d4315aef17491
|
|
|
|
eetq:
|
|
# Clone eetq
|
|
pip install packaging
|
|
git clone https://github.com/NetEase-FuXi/EETQ.git eetq
|
|
|
|
build-eetq: eetq
|
|
cd eetq && git fetch && git checkout $(eetq_commit) && git submodule update --init --recursive
|
|
cd eetq && python setup.py build
|
|
|
|
install-eetq: build-eetq
|
|
cd eetq && python setup.py install
|