Feat: Configure LLM parameters (temp, top_p) per model/agent profile #30173
Replies: 3 comments
-
Yes please! Inference settings make a huge difference, especially with smaller local LLMs. |
Beta Was this translation helpful? Give feedback.
-
👍 It would be great if we could simply pass our own configuration of those parameters. |
Beta Was this translation helpful? Give feedback.
-
We really need this! Otherwise for example the thinking process or reasoning models if included in the output without any formatting. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Description
Zed's language model configuration currently does not support specificying model parameters like
temperature
andtop_p
for inference providers. This limits users from fine-tuning model behavior for specific tasks (like code generation where lower temp is preferred vs documentation vs one purpose agents for like git mcp tool, etc...)Having the ability to compose our own agent capabilities (tools) in Zed is amazing, but model parameters are paramount to this paradigm.
Use case
Configuring model parameters per task or agent profile significantly enhances the quality and relevance of AI-generated content within Zed. Different scenarios benefit from distinct generation strategies:
Creative Content, Documentation: For drafting documentation, brainstorming, or generating explanatory text, users need to set higher
temperature
and suitabletop_p
values to foster more diverse and engaging outputs.Specialized Single-Purpose Agents: When configuring custom AI agents for specific tasks (e.g., a "Git Commit Message Pro," "Code Linter Rule Explainer"), fine-grained control over parameters like
temperature
,max_tokens
,repetition_penalty
andtop_k
for that agent's profile would ensure the agent consistently produces tailored, concise output adhering to specific constraints, making it a reliable and efficient specialized tool.Controlling Verbosity and Focus: For quick queries or summaries, users benefit from adjusting
max_tokens
to control output length orrepetition_penalty
to discourage rambling.Experimentation and Provider-Specific Optimizations: Advanced users or developers integrating new models need to specify provider-specific parameters. This allows full leverage of the chosen model/provider capabilities.
Without this granular control, users are subject to default parameters, leading to sub-optimal results for many common development, writing, and specialized agent workflows that Zed aims to support.
Expected Behavior:
Zed should probably:
Forward all configured model parameters to the API, allowing per-model customization of parameters like:
temperature
repetition_penalty
top_k
top_p
etc...
Allow for provider specific configuration objects, like:
Proposed Solution:
Beta Was this translation helpful? Give feedback.
All reactions