Unverified Commit 0da4df4b authored by Florian Zimmermeister's avatar Florian Zimmermeister Committed by GitHub
Browse files

Fix FP8 KV-cache condition (#2611)

Update kv_cache.py
parent 2358c2bb
......@@ -26,8 +26,8 @@ class KVCache:
if (
dtype == torch.float8_e5m2
and ATTENTION != "flashinfer"
and SYSTEM != "cuda"
and (ATTENTION != "flashinfer"
or SYSTEM != "cuda")
):
raise ValueError(
"float8_e5m2 KV cache is currently only supported for flashinfer on CUDA"
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment