Skip to content

GPT-5 models configured with reasoning.summary: "auto" return condensed reasoning summaries that may omit important reasoning details.#55

Open
82deutschmark wants to merge 2 commits intoarcprize:mainfrom
82deutschmark:claude/reasoning-summary-detailed-014umjZv2UePfWU6zNxYHz9d
Open

GPT-5 models configured with reasoning.summary: "auto" return condensed reasoning summaries that may omit important reasoning details.#55
82deutschmark wants to merge 2 commits intoarcprize:mainfrom
82deutschmark:claude/reasoning-summary-detailed-014umjZv2UePfWU6zNxYHz9d

Conversation

@82deutschmark
Copy link
Copy Markdown

Problem

GPT-5 models configured with reasoning.summary: "auto" return condensed reasoning summaries that may omit important reasoning details. For a benchmarking harness, this means valuable interpretability data is lost even though reasoning tokens are being generated and paid for.

Real-world impact: Example from gpt-5-1-2025-11-13-thinking-high results shows ~60K-70K reasoning tokens charged but zero reasoning content returned because the text.verbosity parameter was not configured (addressed in separate PR) and reasoning.summary: "auto" provides only condensed summaries.

Solution

Changed reasoning.summary from "auto" to "detailed" for all GPT-5 model variants.

Difference between settings:

  • "auto" (old): Condensed summaries, model decides what to include
  • "detailed" (new): Comprehensive reasoning traces with full chain-of-thought

Related

This PR complements the code-level fix that ensures text.verbosity: "high" is set programmatically. Together, these changes ensure maximum reasoning content is captured from GPT-5 models.

Add _ensure_verbosity() helper that automatically sets text.verbosity
to "high" for Responses API calls unless explicitly configured. This
ensures GPT-5/o-series reasoning models return detailed reasoning logs.
Changes reasoning.summary from "auto" to "detailed" for all GPT-5
variants to ensure comprehensive reasoning traces are returned.

Rationale:
- "auto" returns condensed summaries that may omit reasoning details
- "detailed" provides full reasoning traces for better interpretability
- For benchmarking/research, complete reasoning logs are essential
- Affects 12 GPT-5 model configs (gpt-5, gpt-5-mini, gpt-5-nano variants)
@82deutschmark
Copy link
Copy Markdown
Author

@gkamradt is this the wrong fix for the issue? I've been curious about this for a while.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants