Don't install flahs on the CPU tests.

This commit is contained in:
Nicolas Patry 2024-06-04 13:31:43 +00:00
parent d841a4900a
commit 76fef7b1d2
3 changed files with 8 additions and 2 deletions

View File

@ -68,7 +68,7 @@ jobs:
~/.cargo/git
- name: Install
run: |
make install
make install-cpu
- name: Run server tests
run: |
pip install pytest

View File

@ -1,6 +1,9 @@
install-server:
cd server && make install
install-server-cpu:
cd server && make install-server
install-router:
cd router && cargo install --path .
@ -12,6 +15,9 @@ install-benchmark:
install: install-server install-router install-launcher
install-cpu: install-server-cpu install-router install-launcher
server-dev:
cd server && make run-dev

View File

@ -23,7 +23,7 @@ install-server: gen-server
pip install -e ".[bnb, accelerate, quantize, peft, outlines]"
install: install-server install-cuda
install: install-cuda
echo "Installed server"
install-cuda: install-server install-flash-attention-v2-cuda install-vllm-cuda install-flash-attention