Skip to content

Limit token usage output parameter across all queried llm models #75

@converseKarl

Description

@converseKarl

I see hugging face implenentation asserts them but other models like gpt does not.

When the embeddings is past to say openai. Openai has a token output limiting control. Surely this should translate to these api's?

Metadata

Metadata

Assignees

Labels

enhancementNew feature or request

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions