Skip to content

Commit 71524a1

Browse files
authored
[https://nvbugs/5419066][fix] Use trt flow LLM (NVIDIA#6467)
Signed-off-by: Ivy Zhang <[email protected]>
1 parent 48768fd commit 71524a1

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

tests/integration/defs/examples/test_llama.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4069,7 +4069,8 @@ def test_llm_api_lookahead_decoding_1gpu(model_name, model_path):
40694069
"""
40704070
from defs.conftest import llm_models_root
40714071

4072-
from tensorrt_llm.llmapi import (LLM, BuildConfig, KvCacheConfig,
4072+
from tensorrt_llm._tensorrt_engine import LLM
4073+
from tensorrt_llm.llmapi import (BuildConfig, KvCacheConfig,
40734074
LookaheadDecodingConfig, SamplingParams)
40744075
build_config = BuildConfig(max_batch_size=128,
40754076
max_input_len=2048,

0 commit comments

Comments
 (0)