Skip to content

[Usage]: the class member -- multimodalInput, multimodalEmbedding of request (at cpp/include/tensorrt_llm/executor/executor.h) seem unused in tensorrt-llm #9264

@july8023

Description

@july8023

System Info

when I reading the soure code , I noticed that the class member -- multimodalInput, multimodalEmbedding of request (at cpp/include/tensorrt_llm/executor/executor.h) seem unused in tensorrt-llm. Could you clarify where these two parameters come into play.

How would you like to use TensorRT-LLM

I want to run inference of a [specific model](put Hugging Face link here). I don't know how to integrate it with TensorRT-LLM or optimize it for my use case.

Specific questions:

  • Model:
  • Use case (e.g., chatbot, batch inference, real-time serving):
  • Expected throughput/latency requirements:
  • Multi-GPU setup needed:

Before submitting a new issue...

  • Make sure you already searched for relevant issues, and checked the documentation and examples for answers to frequently asked questions.

Metadata

Metadata

Assignees

Labels

MultimodalLabel for issues & PRs regarding Multimodal related objectsquestionFurther information is requested

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions