File tree Expand file tree Collapse file tree 1 file changed +9
-7
lines changed Expand file tree Collapse file tree 1 file changed +9
-7
lines changed Original file line number Diff line number Diff line change @@ -486,15 +486,13 @@ def _deduce_max_tokens(request: GenerationRequest,
486486 splited_prompt_len = int (len (prompt_token_ids ) / cp_size )
487487 default_max_tokens = max_seq_len - splited_prompt_len - query_token_len
488488 if default_max_tokens <= 0 :
489- logger .warning (
490- f"`default_max_tokens` ({ default_max_tokens } ) should be greater than 0, "
489+ # Raise error on `default_max_tokens` not enough, since max_tokens should be less than `default_max_tokens``
490+ raise ValueError (
491+ f"`default_max_tokens` ({ default_max_tokens } ) must be greater than 0, "
491492 f"`default_max_tokens` ({ default_max_tokens } ) = max_seq_len ({ max_seq_len } )"
492493 f" - `splited_prompt_len` ({ splited_prompt_len } ) - `query_token_len` ({ query_token_len } )"
493494 )
494- if max_tokens is None :
495- raise ValueError (
496- "`max_tokens` must be set when `default_max_tokens` is illegal"
497- )
495+
498496 # default_max_tokens is the biggest available value
499497 if max_tokens is None :
500498 return default_max_tokens
@@ -504,7 +502,11 @@ def _deduce_max_tokens(request: GenerationRequest,
504502 f"`default_max_tokens` ({ default_max_tokens } ), using default_max_tokens instead."
505503 )
506504 return default_max_tokens
507- return max_tokens
505+ elif max_tokens <= 0 :
506+ raise ValueError (
507+ f"`max_tokens` ({ max_tokens } ) must be greater than 0" )
508+ else :
509+ return max_tokens
508510
509511 try :
510512 executor_request = tllm .Request (
You can’t perform that action at this time.
0 commit comments