Unverified Commit 0da4df4b authored by Florian Zimmermeister's avatar Florian Zimmermeister Committed by GitHub
Browse files

Fix FP8 KV-cache condition (#2611)

Update kv_cache.py
parent 2358c2bb
...@@ -26,8 +26,8 @@ class KVCache: ...@@ -26,8 +26,8 @@ class KVCache:
if ( if (
dtype == torch.float8_e5m2 dtype == torch.float8_e5m2
and ATTENTION != "flashinfer" and (ATTENTION != "flashinfer"
and SYSTEM != "cuda" or SYSTEM != "cuda")
): ):
raise ValueError( raise ValueError(
"float8_e5m2 KV cache is currently only supported for flashinfer on CUDA" "float8_e5m2 KV cache is currently only supported for flashinfer on CUDA"
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment