Skip to content

How to set my value for the embedding size? #1200

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
aropb opened this issue Jun 5, 2025 · 7 comments
Closed

How to set my value for the embedding size? #1200

aropb opened this issue Jun 5, 2025 · 7 comments

Comments

@aropb
Copy link

aropb commented Jun 5, 2025

Description

The model supports setting the custom length of the embedded array. How do I set my value?

MRL Support indicates whether the embedding model supports custom dimensions for the final embedding.

@aropb aropb changed the title How to set your own value for the embedding size? How to set my value for the embedding size? Jun 5, 2025
@martindevans
Copy link
Member

This sounds like something we don't currently support (I've certainly never heard of it). Could you link some things in the main llama.cpp repo related to this? That'll help inform development for the next version.

@aropb
Copy link
Author

aropb commented Jun 6, 2025

I didn't find it.

Qwen3 embedding and rerank:
ggml-org/llama.cpp#14029

@CISC
Copy link

CISC commented Jun 6, 2025

You don't really set it, you just crop/ignore parts of the output, Matryoshka tuned models have the embedding weights sorted by importance.

@aropb
Copy link
Author

aropb commented Jun 6, 2025

I do not think so.
Then they would not have written about whether the model supports this mechanism (MRL Support) or not.

@CISC
Copy link

CISC commented Jun 6, 2025

I do not think so. Then they would not have written about whether the model supports this mechanism (MRL Support) or not.

I see, then it's probably possible to prompt the model to change the output, I see no further mention of it anywhere though.

@aropb
Copy link
Author

aropb commented Jun 6, 2025

me too, which is weird :)

@aropb
Copy link
Author

aropb commented Jun 6, 2025

@aropb aropb closed this as completed Jun 6, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants