We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4a8ea56 commit da4542aCopy full SHA for da4542a
tensorrt_llm/_torch/pyexecutor/sampler.py
@@ -1071,7 +1071,6 @@ def sample_async(
1071
model_outputs,
1072
new_tokens,
1073
num_context_logits_prefix_sum,
1074
- resource_manager=resource_manager,
1075
)
1076
1077
finish_reasons = self.store.finish_reasons
@@ -1655,8 +1654,6 @@ def _process_requests(
1655
1654
model_outputs: dict[str, torch.Tensor],
1656
new_tokens_cuda: torch.Tensor,
1657
num_context_logits_prefix_sum: list[int],
1658
- *,
1659
- resource_manager: Optional[ResourceManager] = None,
1660
) -> tuple[list[LlmRequest], torch.Tensor, torch.Tensor]:
1661
raw_logits_cuda = model_outputs["logits"]
1662
0 commit comments