Skip to content

Commit 7eabb38

Browse files
committed
remove hardcoding
1 parent 2b1c9dd commit 7eabb38

File tree

4 files changed

+5
-9
lines changed

4 files changed

+5
-9
lines changed

examples/llm-api/quickstart_advanced.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,6 @@ def add_llm_args(parser):
7070
choices=["auto", "TorchSampler", "TRTLLMSampler"])
7171
parser.add_argument('--tp_size', type=int, default=1)
7272
parser.add_argument('--pp_size', type=int, default=1)
73-
parser.add_argument('--cp_size', type=int, default=1)
7473
parser.add_argument('--moe_ep_size', type=int, default=-1)
7574
parser.add_argument('--moe_tp_size', type=int, default=-1)
7675
parser.add_argument('--moe_cluster_size', type=int, default=-1)
@@ -260,7 +259,6 @@ def setup_llm(args, **kwargs):
260259
attention_dp_config=attention_dp_config,
261260
tensor_parallel_size=args.tp_size,
262261
pipeline_parallel_size=args.pp_size,
263-
context_parallel_size=args.cp_size,
264262
moe_expert_parallel_size=args.moe_ep_size,
265263
moe_tensor_parallel_size=args.moe_tp_size,
266264
moe_cluster_parallel_size=args.moe_cluster_size,

tensorrt_llm/commands/serve.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -130,8 +130,6 @@ def get_llm_args(
130130
except KeyError:
131131
raise ValueError(f"Invalid cp_type: {cp_config['cp_type']}. " \
132132
f"Must be one of: {', '.join([t.name for t in CpType])}")
133-
if cp_config["cp_type"] == CpType.HELIX:
134-
cp_config['tokens_per_block'] = kv_cache_config.tokens_per_block
135133

136134
llm_args = {
137135
"model": model,

tensorrt_llm/mapping.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -469,11 +469,6 @@ def __init__(
469469
attn_cp_size=-1,
470470
enable_attention_dp=False,
471471
enable_lm_head_tp_in_adp=False):
472-
#################################################################
473-
# TODO: Remove this hardcoding and obtain cp_config from llm_args.
474-
if cp_size > 1:
475-
cp_config = {"cp_type": CpType.HELIX}
476-
#################################################################
477472
super().__init__(world_size=world_size,
478473
rank=rank,
479474
gpus_per_node=gpus_per_node,

tests/integration/defs/disaggregated/test_configs/disagg_config_ctxtp2_gentp1cp2_deepseek_v3_lite_bf16_tllm_gen.yaml

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@ context_servers:
1111
kv_cache_config:
1212
enable_block_reuse: False
1313
enable_partial_reuse: False
14+
tokens_per_block: 32
1415
tensor_parallel_size: 2
1516
pipeline_parallel_size: 1
1617
cache_transceiver_config:
@@ -23,9 +24,13 @@ generation_servers:
2324
pipeline_parallel_size: 1
2425
context_parallel_size: 2
2526
enable_chunked_prefill: False
27+
cp_config:
28+
cp_type: HELIX
29+
tokens_per_block: 32
2630
kv_cache_config:
2731
enable_block_reuse: False
2832
enable_partial_reuse: False
33+
tokens_per_block: 32
2934
cache_transceiver_config:
3035
backend: "UCX"
3136
urls:

0 commit comments

Comments
 (0)