GPT-5.4 Mini and Nano just dropped
Everyone in the AI space is talking about the new small models from OpenAI.
Most people are reading the marketing.
I read the benchmarks.
Here is what actually changed:
Mini went from 42% to 72.1% on OSWorld-Verified. Terminal-Bench jumped from 38.2% to 60%. SWE-Bench Pro moved from 45.7% to 54.4%.
That is not a minor upgrade. That changes where you draw the line between your main model and your worker model.
Now here is the part most people will miss.
The 400k context window is a trap.
Mini scores 33.6% on the 128K to 256K long-context needle test. Stuff 200,000 tokens of logs into it and it fails roughly 66% of the time.
Context window and context skill are not the same thing.
The routing decision is actually straightforward once you see it clearly:
Full model handles planning, ambiguity, and final judgment.
Mini runs parallel subagents, tool calls, and screenshot-heavy workflows.
Nano handles classification, extraction, ranking, and tight JSON tasks only.
One hard rule: do not put nano anywhere near UI navigation or computer use. It scores 39% on OSWorld while mini scores 72.1%.
The builders who win here will not just swap models.
They will redesign routing.
Would love to hear how others in this community are thinking about model tiering in their agentic stacks. Drop your current setup below.
1
1 comment
Karthik R
5
GPT-5.4 Mini and Nano just dropped
AI Automation Society
skool.com/ai-automation-society
A community built to master no-code AI automations. Join to learn, discuss, and build the systems that will shape the future of work.
Leaderboard (30-day)
Powered by