AI ToolsOpenAIGPT-5.5AutomationLLMs

GPT-5.5 Is Here: OpenAI's Push Toward Agentic Computing

Tariq OsmaniTariq Osmani7 min read
GPT-5.5 Is Here: OpenAI's Push Toward Agentic Computing

OpenAI dropped GPT-5.5 on April 23, 2026 — just six weeks after GPT-5.4 hit the market. The company is calling it its "smartest and most intuitive to use model" yet, and the positioning tells you where OpenAI is heading: away from single-turn chat and toward agentic computing — models that handle multi-step workflows with minimal hand-holding. If you're running AI-powered automations or thinking about where to place the next bet in your stack, here's a clear breakdown of what changed and what it actually means for B2B operations.


What OpenAI Actually Shipped

GPT-5.5 (internal codename "Spud," per Axios) is a meaningful step up from GPT-5.4 on several dimensions, but the headline isn't a benchmark number — it's the shift in what the model is designed to do. Greg Brockman framed it as a "faster, sharper thinker for fewer tokens," which is the polite way of saying the model gets further on harder problems without burning the token budget.

According to OpenAI's announcement, GPT-5.5 is notably better at:

  • Writing and debugging code
  • Researching online and pulling together sources
  • Analyzing data, creating documents and spreadsheets
  • Operating software and moving across tools until a task is finished

That last bullet is the interesting one. It's not incremental — it's a repositioning.


GPT-5.4 vs GPT-5.5 vs GPT-5.5 Pro

Here's how the lineup breaks down across what OpenAI has actually confirmed:

CapabilityGPT-5.4GPT-5.5GPT-5.5 Pro
Context windowStandard1M tokens1M tokens
API input price$5 / 1M tokens
API output price$30 / 1M tokens
AvailabilityPlus, Pro, Business, EnterprisePlus, Pro, Business, EnterprisePro, Business, Enterprise only
Codex supportYesYesYes
Best forGeneral chat, codingAgentic workflows, long-context tasksDeepest reasoning, research workflows

OpenAI hasn't published specific benchmark percentages for GPT-5.5 alongside the release, but TechCrunch reports it scores higher across benchmarks than prior OpenAI models, Google's Gemini 3.1 Pro, and Anthropic's Claude Opus 4.5. Take that directionally — the company chose not to lead with numbers this time, which is itself a signal.


The 1M Context Window at $5 / $30 per Million

The pricing structure matters more than it looks at first glance. At $5 per million input tokens and $30 per million output tokens, running full-document or full-codebase context through GPT-5.5 is economically viable for production automations — not just demos.

For reference, a 500-page PDF runs around 200,000 tokens. That's a single input for roughly $1. Whole Notion workspaces, complete CRM histories, multi-repo codebases — the entire workflow category of "give the model everything and let it figure out what matters" becomes a real option rather than a chunk-and-stitch engineering problem.

A visualization of distributed AI workflows processing data in parallel


Agentic Workflows: The Real Story

Bloomberg's headline captured the positioning well: GPT-5.5 is built to "field tasks with limited instructions." This is OpenAI leaning hard into the agentic direction — models that take a goal, break it down, and work through multi-step processes without needing a human to approve every sub-step.

OpenAI specifically called out that GPT-5.5 "handles multi-step workflows more autonomously with less user input." In practice, this looks like:

  • Research tasks that span multiple tools and sources without needing a new prompt at each step
  • Code tasks that touch multiple files and test outputs before reporting back
  • Data workflows where the model pulls, transforms, and writes to a destination as a single unit of work

This is the same direction Anthropic is pushing with task budgets on Claude Opus 4.7. The industry is converging on agentic loops as the primary unit of value — and GPT-5.5 is OpenAI's clearest statement yet that they see the same future.


Scientific and Technical Research Gains

OpenAI highlighted "meaningful gains on scientific and technical research workflows" as a specific area of improvement. Mark Chen, OpenAI's Chief Research Officer, said GPT-5.5 could "help expert scientists make progress" in research workflows — not replace them, but accelerate the grind of literature review, hypothesis generation, and data analysis.

Jakub Pachocki, OpenAI's Chief Scientist, added an interesting caveat: "The last two years have been surprisingly slow" in terms of improvement pace. That's a notable admission from OpenAI leadership, and it reframes GPT-5.5 as part of a renewed push rather than a steady march.


Availability and Rollout

GPT-5.5 is rolling out now across:

  • ChatGPT Plus, Pro, Business, and Enterprise — all tiers
  • GPT-5.5 Pro — Pro, Business, and Enterprise only
  • Codex — integrated for coding workflows
  • API — at the pricing above, with a 1M-token context window

The rollout is also framed as part of OpenAI's "super app" strategy — ChatGPT, Codex, and the recently announced AI browser converging into a single surface area for agentic work.


Safety and Red-Teaming

OpenAI ran GPT-5.5 through its full safety and preparedness framework evaluation, including internal and external red-teamers and nearly 200 trusted early-access partners before public release. For enterprise buyers, this is the table-stakes reassurance — but the six-week gap from GPT-5.4 is fast by any historical standard, so the pre-release cohort doing real-world stress testing matters.


What This Means for B2B Automation

The 1M-token context window plus genuinely better multi-step task handling is the combination that's directly relevant to business automation. Most of the automation workloads I see are bottlenecked by one of two things: context that doesn't fit, or a model that can't hold a multi-step goal without a human driving each sub-step.

GPT-5.5 moves the needle on both. Concretely:

  • Document-heavy workflows — contract review, RFP response generation, compliance auditing — can now ingest full context without chunking, which reduces the error surface dramatically
  • Multi-tool agentic runs — a workflow that searches, summarizes, writes, and updates a CRM can stay inside a single model invocation instead of being stitched together with orchestration code
  • Cost-predictable automations — $5 / $30 per million tokens is within range for per-task budgets on medium-value operations ($10–50 per run), which opens up a category of workflows that were previously too expensive at lower context limits

A modern office workspace showing automated business processes running on multiple screens


How Smart AI Workspace Approaches This

I build AI automation workflows for businesses end-to-end — and when a model like GPT-5.5 lands, my job is to figure out which existing client pipelines benefit immediately and which should wait for a proven track record.

For GPT-5.5 specifically, the early fit is clear in a few places:

Long-context document workflows — if a client's current pipeline is chunking and re-stitching large documents, GPT-5.5's 1M context often lets me collapse that into a single call with less failure surface.

Agentic research and reporting — the "take a goal, produce a deliverable" category (competitive research briefs, investment memos, compliance summaries) benefits from stronger multi-step handling. Less orchestration code, fewer brittle handoffs between steps.

Dual-model routing — in practice I rarely pick one model for everything. The pattern that works is routing each step of a workflow to the model that wins on that specific sub-task — GPT-5.5 for long-context synthesis and tool use, Claude Opus 4.7 for precise coding and structured output. GPT-5.5 expands the surface area where OpenAI is the right call.

The discipline is the same as always: only move production workloads after the model has proven out on real client data, not benchmark demos.


Ready to Put GPT-5.5 to Work?

If you're running an automation pipeline that's bottlenecked by context limits or brittle multi-step handoffs, GPT-5.5 might be the unlock — or the routing pattern might shift now that it exists. Either way, the right move is to map your actual workflows against what each model does well, rather than picking a favorite and forcing it.

If you want to walk through your automation stack and figure out where GPT-5.5 fits (and where it doesn't), I can help.

Talk to me about your automation needs →


Sources: OpenAI · TechCrunch · CNBC · Bloomberg · Fortune