mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-11-18 23:15:59 +00:00
Custom masks are padded to the shape `[batch_size, max_len, max_len]`. However, flashinfer expects an unpadded mask of the shape `[sum(q_len[i] * k_len[i] for i in range(batch_size)]`. This change unpads the custom mask (currently only used by Gemma 3) to this shape (assuming q_len == k_len, since we only use the custom mask during prefill). |
||
|---|---|---|
| .. | ||
| custom_kernels | ||
| exllama_kernels | ||
| exllamav2_kernels | ||
| tests | ||
| text_generation_server | ||
| .gitignore | ||
| bounds-from-nix.py | ||
| kernels.lock | ||
| Makefile | ||
| Makefile-awq | ||
| Makefile-eetq | ||
| Makefile-exllamav2 | ||
| Makefile-flash-att | ||
| Makefile-flash-att-v2 | ||
| Makefile-flashinfer | ||
| Makefile-selective-scan | ||
| Makefile-vllm | ||
| pyproject.toml | ||
| README.md | ||
| req.txt | ||
| requirements_cuda.txt | ||
| requirements_gen.txt | ||
| requirements_intel.txt | ||
| requirements_rocm.txt | ||
| uv.lock | ||
Text Generation Inference Python gRPC Server
A Python gRPC server for Text Generation Inference
Install
make install
Run
make run-dev