Introduction
GPT-5.5 raises the baseline for complex production workflows. It’s a strong fit for coding use cases, tool-heavy agents, grounded assistants, long-context retrieval, product-spec-to-plan workflows, and customer-facing workflows where execution quality and response polish are critical.
To get the most out of GPT-5.5, treat it as a new model family to tune for, not a drop-in replacement for gpt-5.2 or gpt-5.4. Begin migration with a fresh baseline instead of carrying over every instruction from an older prompt stack. Start with the smallest prompt that preserves the product contract, then tune reasoning effort, verbosity, tool descriptions, and output format against representative examples.
GPT-5.5 supports all API features that were already available with GPT-5.4, including prompt caching, hosted tools, tool search, compaction, and phase handling for manually replayed assistant items.
See the GPT-5.5 Prompting Guide for examples of successful prompting patterns.
What’s new
- More efficient reasoning: GPT-5.5 reaches strong results with fewer reasoning tokens than prior models, even at the same reasoning effort. This is especially useful in complex, tool-heavy, or multi-step workflows where token savings compound.
- Stronger task execution with outcome-first prompts: GPT-5.5 is better at working from a clear goal, preserving constraints, and turning product intent into concrete next steps. Describe the expected outcome, success criteria, allowed side effects, evidence rules, and output shape. Avoid step-by-step process guidance unless the exact path matters.
- Stronger and more precise tool use: GPT-5.5 is especially useful on large tool surfaces, multi-step service workflows, and long-running agent tasks. It tends to be more precise in tool selection and argument use.
- Tone is often more polished, but can be more direct: GPT-5.5 often produces warmer, more readable answers with less prompt scaffolding.
Behavioral changes
-
Reasoning effort now defaults to
medium: GPT-5.5 defaults tomediumreasoning effort. Treatmediumas the recommended balanced starting point for quality, reliability, latency, and cost. For latency-sensitive workflows, evaluatelowbeforenonewhen tool use, planning, search, or multi-step decision making still matters. Reservenonefor latency-critical tasks that don’t need reasoning or multi-chained tool calls, such as lightweight voice turns, fast information retrieval, and classification. Increase tohighorxhighonly when evals show a measurable quality gain that justifies the extra latency and cost. See the Reasoning models documentation for more details on recommended settings.Higher reasoning effort isn’t automatically better. If the task has conflicting instructions, weak stopping criteria, or open-ended tool access, higher effort can lead to overthinking, unnecessary searching, or output quality regressions. Increase effort only when evals show a measurable quality gain.
-
Image inputs preserve more visual detail by default: GPT-5.5 updates the default handling for image inputs to preserve more visual detail and improve computer use performance. When
image_detailis unset or set toauto, the model now usesoriginalbehavior, preserving images without resizing up to 10,240,000 pixels or a 6,000-pixel dimension limit. Forhigh, specify the value directly; it preserves images without resizing up to 2,500,000 pixels or a 2,048-pixel dimension limit.lownow focuses on context efficiency and resizes images above a 512-pixel dimension limit more aggressively than previous models. See the Images and vision documentation. -
Improved instruction following: GPT-5.5 interprets prompts in a literal and thorough manner, enabling specific, descriptive instructions when the product requires them. Define success criteria and stopping rules, especially for long-running, tool-heavy, or evidence-gathering workflows. See Write outcome-first prompts and Keep the right specificity.
-
Default style is more concise and direct: GPT-5.5 tends to be efficient, direct, and task-oriented by default. This is useful for many production workflows, but customer-facing or conversational experiences may need explicit personality, warmth, rationale, and formatting guidance. Use
text.verbosityintentionally:mediumis the default, andlowis often a better starting point for concise responses. See the GPT-5.5 prompting guide. -
Coding workflows need stronger orchestration: GPT-5.5 is better suited to complex coding tasks that require planning, tool use, codebase navigation, verification, and multi-step execution. For coding agents, be explicit about reuse, subagent delegation, test expectations, acceptance criteria, and when to continue versus ask for help.
Migration quickstart
Automated migration with Codex
Codex can apply the recommended changes in this guide with the OpenAI Docs Skill.
$openai-docs migrate this project to gpt-5.5To use this skill in other coding agents, download it from the OpenAI skills repository.
API and model parameters
- Update the model slug to
gpt-5.5. - Use the Responses API for any reasoning, tool-calling, or multi-turn use case.
- Tune
reasoning.effort. Uselowfor efficient reasoning,mediumfor a balanced point on the latency/performance curve,highfor complex agentic tasks that require hard reasoning and where latency matters less, andxhighfor the hardest asynchronous agentic tasks or evals that test the bounds of model intelligence. See the Reasoning models documentation. - To configure for more concise responses, set
text.verbositytolow. On GPT-5.5, this will result in proportionally more concise responses thanlowverbosity with GPT-5.4. - For tool-heavy or long-running workflows, verify that your application handles
phase, preambles, and assistant-item replay correctly. - Benchmark against other models on accuracy, token consumption, and end-to-end latency.
Prompting
- State the expected outcome and success criteria.
- Reduce or remove detailed step-by-step process guidance. Let GPT-5.5 choose the path unless the product requires that path.
- Remove output schema definitions from the prompt where possible. Use Structured Outputs instead.
- Optimize your prompt for caching: static parts first, dynamic parts last.
- Drop the current date. The model is already aware of the current UTC date.
- Review and optimize your prompts based on the guidance in Prompting GPT-5.5.
Using reasoning models
This guidance applies to GPT-5 series models and is worth revisiting whenever teams move workloads onto reasoning models. GPT-5.5 carries forward many capabilities that first appeared in earlier models, but they’re still worth reviewing if you are moving from an earlier GPT-5 model, GPT-4.1, or a reasoning model such as o3.
Teams can overlook these features because they sit partly in API configuration and orchestration rather than in the prompt itself. Used together, the Responses API, reasoning controls, verbosity, structured outputs, prompt caching, tool design, hosted tools, and state management help reasoning models deliver their best intelligence, reliability, latency, and cost profile.
- Responses API: GPT-5.5 works best in the Responses API. Use
previous_response_idfor multi-turn state handling. For stateless or Zero Data Retention flows, pass back the relevant returned output items each turn. See Passing context from the previous response for details. - Reasoning effort: Use
reasoning.effortto choose betweenlow,medium,high, orxhigh. The default ismedium, but many workloads will perform well withlow. Reservenonefor use cases where low latency is more important than intelligence. See Reasoning Models for detailed recommendations. - Verbosity: Use
text.verbosityto control output length. Treat final answer length as separate from reasoning quality; specify word budgets, section counts, table widths, or JSON-only output where needed. - Structured Outputs: Avoid describing the expected output schema in the prompt. Use Structured Outputs for automatic validation and increased accuracy.
- Prompt caching: Prompt caching works automatically for eligible long prompts and can reduce latency and input-token cost. To maximize cache hits, keep stable content at the beginning of the request. Put dynamic user-specific context near the end. For repeated traffic with common prefixes, use
prompt_cache_keyconsistently and trackusage.prompt_tokens_details.cached_tokens. - Tool calling: GPT-5.5 supports the same tool-calling patterns as GPT-5.4, including function tools and tool-heavy agent workflows. Put most tool-specific guidance in the tool descriptions themselves: what the tool does, when to use it, required inputs, side effects, retry safety, and common error modes. Add tool-specific context to system instructions only when it applies across tools or materially changes the agent’s operating policy.
- Hosted tools and tool search: Prefer OpenAI-hosted tools where they fit the workflow, such as web search, file search, code interpreter, image generation, and computer use. Hosted tools reduce custom orchestration burden and keep common tool patterns aligned with the Responses API and Agents SDK. Use custom function tools when you need to call your own systems, enforce domain-specific side effects, or expose internal business workflows. For large tool catalogs, consider using tool search to defer tool definitions and load only the relevant subset.
- Tool preambles: Preambles can improve chat UX because the user sees an initial, useful status update before the model generates the final response. They also make tool use easier to follow: the model can state what it’s about to check or do, then continue from that same assistant state after tool results arrive.
phasehandling: If your application manually manages Responses state by passing output items back each turn instead of usingprevious_response_id, preserve thephaseparameter on returned assistant output items and pass it back unchanged. This is especially important when using reasoning effort, preambles, or repeated tool calls. See Phase parameter.- Compaction: For long-running agents, use conversation/state compaction intentionally. Preserve completed actions, active assumptions, IDs, tool outcomes, unresolved blockers, and the next concrete goal.
- Agents SDK: For new agentic systems, use the latest Agents SDK patterns for tool orchestration, tracing, handoffs, and state management rather than rebuilding orchestration from scratch.
- Current date: GPT-5.5 is aware of the current date in UTC. You don’t need to add the current date to system instructions. Add explicit date or timezone context only when the application needs a business-specific timezone, policy-effective date, user-local date, or other non-UTC reference point.