Skip to content

Commit 7cd508d

Browse files
committed
Fix precommit
Signed-off-by: shuw <[email protected]>
1 parent 8d8ffbd commit 7cd508d

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

vllm/attention/backends/flashinfer.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -24,9 +24,6 @@
2424
BatchPrefillWithPagedKVCacheWrapper = None
2525
FLASHINFER_WORKSPACE_BUFFER_SIZE = 0
2626

27-
FLASHINFER_KV_CACHE_LAYOUT: str = os.getenv("FLASHINFER_KV_CACHE_LAYOUT",
28-
"NHD").upper()
29-
3027
import torch
3128

3229
import vllm.envs as envs
@@ -49,6 +46,9 @@
4946
from vllm.worker.model_runner import (ModelInputForGPUBuilder,
5047
ModelInputForGPUWithSamplingMetadata)
5148

49+
FLASHINFER_KV_CACHE_LAYOUT: str = os.getenv("FLASHINFER_KV_CACHE_LAYOUT",
50+
"NHD").upper()
51+
5252

5353
class FlashInferBackend(AttentionBackend):
5454

0 commit comments

Comments
 (0)