Skip to content

Commit 46dccb5

Browse files
authored
[None][chore] Prevent negative max_tokens passed into tllm request (#9037)
Signed-off-by: Junyi Xu <[email protected]>
1 parent b6bced8 commit 46dccb5

File tree

1 file changed

+9
-7
lines changed

1 file changed

+9
-7
lines changed

tensorrt_llm/executor/base_worker.py

Lines changed: 9 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -486,15 +486,13 @@ def _deduce_max_tokens(request: GenerationRequest,
486486
splited_prompt_len = int(len(prompt_token_ids) / cp_size)
487487
default_max_tokens = max_seq_len - splited_prompt_len - query_token_len
488488
if default_max_tokens <= 0:
489-
logger.warning(
490-
f"`default_max_tokens` ({default_max_tokens}) should be greater than 0, "
489+
# Raise error on `default_max_tokens` not enough, since max_tokens should be less than `default_max_tokens``
490+
raise ValueError(
491+
f"`default_max_tokens` ({default_max_tokens}) must be greater than 0, "
491492
f"`default_max_tokens` ({default_max_tokens}) = max_seq_len ({max_seq_len})"
492493
f" - `splited_prompt_len` ({splited_prompt_len}) - `query_token_len` ({query_token_len})"
493494
)
494-
if max_tokens is None:
495-
raise ValueError(
496-
"`max_tokens` must be set when `default_max_tokens` is illegal"
497-
)
495+
498496
# default_max_tokens is the biggest available value
499497
if max_tokens is None:
500498
return default_max_tokens
@@ -504,7 +502,11 @@ def _deduce_max_tokens(request: GenerationRequest,
504502
f"`default_max_tokens` ({default_max_tokens}), using default_max_tokens instead."
505503
)
506504
return default_max_tokens
507-
return max_tokens
505+
elif max_tokens <= 0:
506+
raise ValueError(
507+
f"`max_tokens` ({max_tokens}) must be greater than 0")
508+
else:
509+
return max_tokens
508510

509511
try:
510512
executor_request = tllm.Request(

0 commit comments

Comments
 (0)