mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-21 23:12:07 +00:00
* Remove vLLM dependency for CUDA This change adds `attention-kernels` as a dependency for paged attention and cache reshaping. With that, we don't use vLLM anywhere for CUDA. Tested run (since we don't have paged attention in CI): ``` ❯ ATTENTION=paged python -m pytest integration-tests -k "llama and awq" --release [...] 5 snapshots passed. ``` * Fix clippy warning |
||
---|---|---|
.. | ||
__init__.py | ||
common.py | ||
cuda.py | ||
flash_attn_triton.py | ||
flashinfer.py | ||
ipex.py | ||
kv_cache.py | ||
rocm.py |