The Big Picture: Three Model Generations in Three Weeks
Something rare happened in the last month: all three major AI providers shipped a new model generation within weeks of each other. Claude Opus 4.7 landed on April 16, GPT-5.5 became the ChatGPT default on May 5, and Google unveiled Gemini 3 at I/O on May 12.
Each model generation carries its own prompting philosophy. The unifying theme across all three is a shift from process-first to outcome-first prompting. These models handle reasoning complexity internally. They need practitioners to describe the destination precisely and get out of the way.
Here is what changed and what to do about it.
Claude Opus 4.7: Say Exactly What You Mean
Claude Opus 4.7 is Anthropic's most capable generally available model. It scored 87.6% on SWE-bench Verified (up from 80.8% for Opus 4.6), and its vision capabilities now support images up to 2,576 pixels on the long edge β more than three times the previous resolution.
The most important prompting change: Opus 4.7 interprets instructions more literally than 4.6. Prompts that relied on the model filling in gaps or inferring unstated constraints now underperform. If you want a specific format, length, tone, or structure, you must state it explicitly.
What to Start Doing
- Audit existing prompts for implicit assumptions. Any constraint you assumed the model would infer needs to be written out.
- Be explicit about output format (JSON, markdown, plain text), length (word or paragraph count), tone (formal, conversational, technical), and structure (sections, headers, bullets).
What to Stop Doing
- Remove scaffolding that older models needed. Instructions like "double-check the slide layout before returning" or forced interim status messages are unnecessary. Opus 4.7 self-verifies and emits progress updates natively. Stripping this scaffolding improves output quality by reducing noise in the instruction set.
New Capabilities
- xhigh effort tier. Anthropic introduced a fifth effort level between "high" and "max." It is the default in Claude Code for agentic coding tasks. Use xhigh for coding and multi-step agentic workflows, high for intelligence-sensitive tasks, medium for standard conversation, and low for classification and extraction.
- 3x vision resolution. Dense spreadsheets, complex diagrams, dashboard screenshots, and handwritten notes that were previously unreadable may now be processable with high fidelity.
Alongside Opus 4.7, Anthropic launched Claude Design, a research preview for creating live HTML-based visual outputs through conversational iteration, and Claude for Small Business, a package of connectors for QuickBooks, PayPal, HubSpot, and other platforms.
GPT-5.5: Outcome-First Prompting Becomes Official
GPT-5.5 is now the default model across ChatGPT and the OpenAI API. It supports a 1M token context window, image input, structured outputs, function calling, MCP, and web search.
OpenAI's prompting guidance makes the paradigm shift explicit: better prompts for GPT-5.5 are often shorter and simpler. The complexity that used to live in the prompt is now handled by the model.
Three Changes to Internalize
- Stop describing output schemas in prompts. Use Structured Outputs (the response_format API parameter) for automatic validation and increased accuracy. This is now the official recommendation.
- Put tool guidance in tool descriptions, not system prompts. What a tool does, when to use it, required inputs, side effects, retry safety, common error modes β all of this belongs in the tool description itself. System instructions should only contain cross-cutting policy that applies across tools.
- Default to plain paragraphs. OpenAI now recommends prose as the default format for conversation, explanations, reports, and documentation. Headers, bold text, bullets, and numbered lists should be used sparingly β only when the user requests them, when ranking or comparison is needed, or when prose would be harder to scan.
Formatting Default
This is worth emphasizing: GPT-5.5 produces better results when prompts do not force structured formatting everywhere. If you have been adding "respond with bullet points" or "use headers for each section" as a default, test removing it. The model's natural prose output is often higher quality.
Gemini 3: The Reasoning Leap and Agentic Vision
Google shipped Gemini 3 at I/O 2026 (May 12), representing its biggest model jump since the Gemini 1-to-2 transition. Gemini 3 Pro delivers significant improvements in reasoning, instruction following, tool use, and long-context capabilities compared to Gemini 2.5 Pro.
The Standout: Agentic Vision
Agentic Vision, available in Gemini 3 Flash, combines visual reasoning with code execution. The model can formulate plans to zoom in, inspect, and manipulate images step-by-step, grounding answers in visual evidence. This is fundamentally different from asking a model to describe an image β you can ask it to investigate one.
Prompting Changes
- Replace explicit chain-of-thought with the API parameter. If you were using complex CoT prompting to force Gemini 2.5 to reason, try Gemini 3 with
thinking_level: "high"and simplified prompts instead. The olderthinking_budgetparameter is deprecated. - Stop overriding temperature. Remove manual temperature settings and use the Gemini 3 default of 1.0. Setting it below 1.0 can cause looping or degraded performance.
- Use the right model variant. Gemini 3 Pro Preview was discontinued on March 26, 2026. Use
gemini-3.1-pro-previewfor current projects. If the model ignores custom tools, trygemini-3.1-pro-preview-customtools.
The Protocol Stack Solidifies
Since March, the MCP + A2A protocol stack has moved from emerging consensus to enterprise baseline.
MCP crossed 97 million monthly SDK downloads (Python + TypeScript combined) and has been adopted by every major AI provider: Anthropic, OpenAI, Google, Microsoft, and Amazon. Both MCP and A2A are now governed under the Linux Foundation's Agentic AI Foundation (AAIF), launched in December 2025 with six co-founders.
Google ADK 1.0 went GA at Cloud Next 2026 across Python, Go, Java, and TypeScript. The A2A protocol surpassed 150 organizations in production. The architecture is increasingly standardized: MCP for tool access (how agents interact with external systems) and A2A for agent-to-agent collaboration (how AI agents work together).
For prompt engineers, this means the job is expanding. You are no longer just writing system prompts β you are writing tool descriptions, defining agent role boundaries, and designing context assembly pipelines.
Context Engineering Goes Mainstream
In April, Gartner declared context engineering the breakout AI capability of 2026. The data now backs it up: according to the 2026 State of Context Management Report, 82% of IT and data leaders agree that prompt engineering alone is no longer sufficient to power AI at scale, and 95% of data teams plan to invest in context engineering training this year.
Agentic Context Engineering (ACE), introduced by researchers from Stanford University, SambaNova Systems, and UC Berkeley, treats context as an evolving playbook the agent can update over time, rather than a fixed prompt. This is the clearest articulation of the difference between prompt engineering (optimizing static instructions) and context engineering (designing dynamic information environments).
Two techniques are emerging as critical for agents working across extended time horizons:
- Compaction: Summarizing and discarding intermediate reasoning to prevent context pollution as conversation length grows.
- Structured note-taking: Teaching agents what to remember and what to forget, maintaining a curated working memory rather than an unbounded context window.
What to Do This Week
If you use Claude:
- Audit your prompts for implicit assumptions. Opus 4.7's literal instruction following means unstated constraints will be missed. Be explicit about format, length, tone, and structure.
- Try the xhigh effort level for agentic and coding tasks. Remove manual scaffolding (forced status updates, self-verification instructions) β the model handles these natively.
- If you work with images, re-test with Opus 4.7's higher-resolution vision. Documents and dashboards that were previously unreadable may now be processable.
If you use OpenAI:
- Move output schemas from prompts to Structured Outputs. This is now the official recommendation for GPT-5.5.
- Move tool guidance from system prompts to tool descriptions. System prompts should contain cross-cutting policy only.
- Test whether your prompts can be shortened. GPT-5.5 handles complexity that older models needed explicit step-by-step instructions for.
If you use Google:
- Migrate from Gemini 2.5 Pro to Gemini 3.1 Pro. Replace explicit chain-of-thought prompting with
thinking_level: "high". - Remove manual temperature settings and use the new default (1.0).
- Explore Agentic Vision for any workflows that involve image analysis or visual reasoning.
For all platforms:
- Invest in context engineering infrastructure (retrieval pipelines, tool schemas, structured memory) over longer system prompts.
- Write tool descriptions as carefully as you write system prompts β this is where prompting effort increasingly delivers the highest ROI.