GPT-5.4 Review: Features, Benchmarks & Honest Verdict (March 2026) — Recensione Completa

GPT-5.4 Review: Features, Benchmarks & Honest Verdict

OpenAI released GPT-5.4 on March 5, 2026 — and for once the hype is largely justified. This is the most significant capability jump in the GPT-5 series since the original launch in August 2025, and it changes the calculus on which AI model to use for professional work. Here is a complete breakdown based on independent benchmarks, real-world testing, and a head-to-head comparison with Claude Opus 4.6 and Gemini 3.1 Pro.

What Is GPT-5.4?

GPT-5.4 unifies two threads that previously existed as separate products: GPT-5.2’s reasoning capabilities and GPT-5.3-Codex’s coding performance. The result is a single model that handles both. It is available in three tiers: GPT-5.4 Thinking (included with ChatGPT Plus at $20/month), GPT-5.4 Pro (requires the $200/month Pro plan), and direct API access at $2.50 per million input tokens and $15.00 per million output tokens.

The Headline Feature: Computer Use That Actually Works

GPT-5.4 is the first general-purpose AI model to score above human performance on a real-world desktop task benchmark. On OSWorld-Verified — which tests whether an AI can navigate operating systems and complete actual desktop workflows — GPT-5.4 scored 75.0%. The human baseline is 72.4%. This is not a parlor trick. The model can receive screenshots, move a cursor, click elements, type text, and chain multiple actions together to complete workflows like filling out forms, configuring software settings, or running test suites through a GUI — all without human intervention at each step.

The Computer Use API operates through a loop: the model sees a screenshot, decides on an action, executes it, receives a new screenshot, and repeats. GPT-5.4 also introduces “Tool Search” for agent workflows: instead of loading every available tool definition into the prompt (burning tokens), the model intelligently searches and selects tools on demand, reducing token usage by 47% on the MCP Atlas benchmark.

Context Window: 1 Million Tokens, With a Catch

The API supports up to 1.05 million tokens of context — enough for entire legal document sets, large codebases, or multi-year financial reports in a single call. The catch: 1M context is an opt-in feature, and input token pricing doubles from $2.50 to $5.00 per million tokens once you cross 272K tokens. For most users, the standard 272K window is more than sufficient and significantly cheaper.

Benchmark Results: Honest Numbers

On GDPval — OpenAI’s benchmark spanning 44 professions including law, finance, and medicine — GPT-5.4 scored 83%, matching or exceeding human professionals in 83% of task comparisons. On the BigLaw Bench specifically, the score was 91%, making it genuinely useful for legal document analysis. On the investment banking benchmark for financial modeling, GPT-5.4 Thinking scored 87.3% — up from 43.7% with the original GPT-5. The model is 33% less likely to make errors on individual factual claims compared to GPT-5.2, and overall responses contain 18% fewer errors.

But there are limits. On SWE-bench (software engineering tasks), GPT-5.4 scores 57.7% — significantly behind Claude Opus 4.6 at 80.8% and Gemini 3.1 Pro at 80.6%. A widely shared blind test by evaluator Nate B Jones exposed a commonsense reasoning gap: when asked about the fastest route to a car wash, GPT-5.4 suggested walking (for environmental reasons), missing the obvious constraint that you need the car at the car wash. Claude and Gemini both gave the correct answer immediately.

Steerable Thinking Plans: The Best New UX Feature

In ChatGPT, GPT-5.4 Thinking now shows its reasoning plan upfront before generating the full response. You can review the plan and redirect the model mid-response. For anyone who has waited 30 minutes for a long AI output only to receive something off-target, this is a material quality-of-life improvement that reduces wasted iterations on complex tasks.

GPT-5.4 vs Claude Opus 4.6 vs Gemini 3.1 Pro

The March 2026 AI landscape has three serious contenders at the frontier, and none dominates every category. GPT-5.4 wins on knowledge work, computer use, and professional document tasks — slide decks, financial models, legal analysis. Claude Opus 4.6 wins on coding quality (80.8% SWE-bench vs 57.7%), multi-file refactoring, and complex instruction following. Gemini 3.1 Pro wins on value, multimodal capabilities (native audio and video), and PhD-level science reasoning — and it is the cheapest of the three at $2/$12 per million tokens. The practical recommendation is to use GPT-5.4 for professional document work and computer automation, Claude for production code, and Gemini for high-volume or multimodal workloads.

Should You Upgrade?

For ChatGPT Plus subscribers, yes. GPT-5.4 Thinking is included in your $20/month plan and replaces GPT-5.2 as the default. The improvements in professional task performance, computer use, and the thinking plan feature are all immediately useful for regular users. For developers, the API upgrade depends on your use case: coding-heavy workflows may still prefer Claude Opus 4.6 or Gemini 3.1 Pro, but agentic workflows involving computer automation or large document processing have a clear new leader.


GPT-5.4: Recensione Completa, Benchmark e Verdetto Onesto (Marzo 2026)

OpenAI ha rilasciato GPT-5.4 il 5 marzo 2026 — e per una volta il clamore è ampiamente giustificato. Questo è il salto di capacità più significativo nella serie GPT-5 dal lancio originale nell’agosto 2025. Ecco una panoramica completa basata su benchmark indipendenti, test reali e un confronto diretto con Claude Opus 4.6 e Gemini 3.1 Pro.

La Funzione Principale: Computer Use che Funziona Davvero

GPT-5.4 è il primo modello AI di uso generale a superare le prestazioni umane su un benchmark reale di attività desktop. Su OSWorld-Verified — che testa se un’AI può navigare sistemi operativi e completare flussi di lavoro desktop reali — GPT-5.4 ha ottenuto il 75,0%. La baseline umana è 72,4%. Il modello può ricevere screenshot, muovere un cursore, cliccare elementi, digitare testo e concatenare più azioni per completare workflow — tutto senza intervento umano ad ogni passaggio.

Finestra di Contesto: 1 Milione di Token, Con una Precisazione

L’API supporta fino a 1,05 milioni di token — sufficiente per interi set di documenti legali, grandi codebase o report finanziari pluriennali in una singola chiamata. La precisazione: il contesto da 1M è una funzione opt-in, e il prezzo raddoppia oltre i 272K token. Per la maggior parte degli utenti, la finestra standard da 272K è più che sufficiente.

Benchmark: I Numeri Onesti

Su GDPval — il benchmark di OpenAI su 44 professioni — GPT-5.4 ha ottenuto l’83%, eguagliando o superando i professionisti umani nell’83% dei confronti. Sul BigLaw Bench il punteggio è stato del 91%. Tuttavia su SWE-bench (ingegneria del software) GPT-5.4 ottiene il 57,7% — significativamente indietro rispetto a Claude Opus 4.6 all’80,8% e Gemini 3.1 Pro all’80,6%.

GPT-5.4 vs Claude Opus 4.6 vs Gemini 3.1 Pro

GPT-5.4 vince su lavoro professionale, computer use e documenti. Claude Opus 4.6 vince su qualità del codice e seguire istruzioni complesse. Gemini 3.1 Pro vince su valore, capacità multimodali e ragionamento scientifico a livello PhD — ed è il più economico dei tre. La raccomandazione pratica: usa GPT-5.4 per documenti professionali e automazione, Claude per codice in produzione, Gemini per carichi di lavoro ad alto volume o multimodali.

Leave a Comment