Skip to content

Commit 0f15b8d

Browse files
houseroadfacebook-github-bot
authored andcommitted
Enable Llama4 available in fbcode vllm (#16573)
Summary: Enable Llama4 in the fbcode vllm. Reviewed By: henryoier Differential Revision: D72945995
1 parent dc1b4a6 commit 0f15b8d

File tree

1 file changed

+6
-4
lines changed

1 file changed

+6
-4
lines changed

vllm/compilation/compiler_interface.py

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@
1111
import torch._inductor.compile_fx
1212
import torch.fx as fx
1313

14+
import vllm.envs as envs
1415
from vllm.config import VllmConfig
1516
from vllm.utils import is_torch_equal_or_newer
1617

@@ -296,10 +297,11 @@ def _get_shape_env() -> AlwaysHitShapeEnv:
296297
inner_compile=hijacked_compile_fx_inner,
297298
config_patches=current_config)
298299

299-
assert hash_str is not None, (
300-
"failed to get the hash of the compiled graph")
301-
assert file_path is not None, (
302-
"failed to get the file path of the compiled graph")
300+
if not envs.VLLM_DISABLE_COMPILE_CACHE:
301+
assert hash_str is not None, (
302+
"failed to get the hash of the compiled graph")
303+
assert file_path is not None, (
304+
"failed to get the file path of the compiled graph")
303305
return compiled_graph, (hash_str, file_path)
304306

305307
def load(self,

0 commit comments

Comments
 (0)