mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 20:34:54 +00:00
Update kv_cache.py
This commit is contained in:
parent
2358c2bb54
commit
e03a7167ee
@ -26,8 +26,8 @@ class KVCache:
|
||||
|
||||
if (
|
||||
dtype == torch.float8_e5m2
|
||||
and ATTENTION != "flashinfer"
|
||||
and SYSTEM != "cuda"
|
||||
and (ATTENTION != "flashinfer"
|
||||
or SYSTEM != "cuda")
|
||||
):
|
||||
raise ValueError(
|
||||
"float8_e5m2 KV cache is currently only supported for flashinfer on CUDA"
|
||||
|
Loading…
Reference in New Issue
Block a user