Skip to content

Commit a0d4d52

Browse files
committed
fix
Signed-off-by: junq <[email protected]>
1 parent 4fb2876 commit a0d4d52

File tree

1 file changed

+1
-3
lines changed

1 file changed

+1
-3
lines changed

tensorrt_llm/_torch/pyexecutor/model_engine.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -439,9 +439,7 @@ def __init__(
439439
self.kv_cache_manager_key = ResourceManagerType.KV_CACHE_MANAGER
440440
self.lora_model_config: Optional[LoraModelConfig] = None
441441
self.cuda_graph_dummy_request = None
442-
self.cuda_graph_model_engine = CUDAGraphModelEngine(
443-
self
444-
) if self._run_cuda_graphs or self._torch_compile_piecewise_cuda_graph else None
442+
self.cuda_graph_model_engine = CUDAGraphModelEngine(self)
445443

446444
# Setup the local cache indirection buffer only once and reuse it.
447445
# This way it can also be used for CUDA graphs.

0 commit comments

Comments
 (0)