GPT-5: OpenAI's 3D Chess Match

GPT-5: OpenAI's 3D Chess Match

The GPT-5 launch has been, well, tumultuous, to say the least. For this article I'm going to mostly park the launch controversies and focus on the features that push the frontier of AI but aren't so big and sexy. These aren’t the kind of headline features you demo on stage, but they’re the kind you notice when you ship product faster, trust the model more, and keep more money in your pocket (or your companies).


The Top 3 Pain Points With GPT-5 So Far

Before I get into the meat of the article I want to quickly address where OpenAI missed the mark with GPT-5.

  1. The Hype Gap
    Altman and others spent a lot of time hyping GPT-5 to be world changing, creating visions of AGI and having a PhD that knows everything right in your pocket. This left most who've been following expecting a transformative jump in reasoning, multimodality, or speed that didn’t materialize.
  2. Opaque Model Routing & Tier Limits
    The unified router hides which variant you’re using, and Plus-tier subscribers have restricted access to the “thinking” mode. Without visibility into which model is active, debugging and prompt optimization become guesswork.
  3. Loss of Model Choice & Legacy Access
    GPT-5 removes the ability to directly choose between variants—or fall back to older models that might perform better for niche tasks. Builders who tuned workflows around specific legacy models are left without those tools.

I plan on writing a more detailed breakdown of the launch in the coming weeks but for now let's dive into some lesser discussed features.


1. Model Selection Without the Menu Tax

The model zoo is gone. No more deciding between “fast” and “reasoning” variants like you’re picking a Pokémon. GPT-5 routes your prompt to the right engine automatically—lightweight for speed, heavyweight for deep reasoning—based on the actual request. It’s not the mythical “one model to rule them all,”, it does however bring the use of thinking models to the masses. Before the introduction of the Model Selector only 7% of Plus Users ($20 tier) had used a thinking model.

How it works: GPT-5 runs a lightweight “router” process before generation begins. It uses explicit signals in your prompt (e.g., “Think step-by-step…”), your account preferences, and historical completion patterns to decide whether to send the request to its fast inference path or to its slower, more reasoning-heavy variant. The router itself is an LLM trained on millions of past prompt–completion pairs, optimized to predict which model will yield the best combination of speed, accuracy, and satisfaction for the given query.


2. Hallucinations go on a Diet

Two separate training passes now target the two worlds the model lives in: browse on for real-time lookups, and browse off for memory-only answers. Both were graded by an LLM+web fact-checker and validated by humans. Net result: noticeably fewer “confidently wrong” moments. Less time cross-checking means a higher level of confidence in responses.

How it works: Hallucinations happen when a model generates content that sounds right but is wrong—made-up facts, misattributed quotes, even fake API names. LLMs are next-token predictors, so they generate text that’s statistically likely, not necessarily correct. GPT-5 attacks this problem with a split training strategy.

  • In browse on mode, it was trained to effectively use live search to find and integrate current, accurate information through web browsing.
  • In browse off mode, the focus was on reducing errors when relying solely on internal knowledge.

Factual accuracy was stress-tested by an LLM grader with web access, which extracted claims, fact-checked them, and had its work validated by human raters. The dual-path tuning cut hallucination rates meaningfully in both modes compared to previous generations. When comparing GPT-5 against 4o and o3, there was a 45% and 80% (respectively) reduction in hallucinations.


Have a specific use case you're considering for GPT-5 and your business? Shoot me a message and let's talk about the best model(s) to fit your use case.

Let's Talk

3. Sycophancy Gate

Sycophancy—the model agreeing with you even when you’re wrong—got a post-training purge. GPT-5 learned from real production conversations where agreement was penalized if it broke with reality. It’s less about “you can’t” and more about “here’s what you can do.

Why Sycophancy Is a Problem

  • Trust and Accuracy: Sycophantic AI sacrifices truth and reliability in favor of making the user feel good, which can erode trust in AI systems.
  • Emotional Wellbeing: Over-validation may reinforce negative emotions or poor decisions, especially in mental health or advice scenarios.
  • Dark Patterns: Some experts view excessive sycophancy as a “dark pattern”—intentionally or unintentionally designed to keep users engaged at the expense of their well-being or informed decision-making.
  • Safety Risks: In extreme cases, models have gone as far as validating delusions or backing harmful behavior, amplifying safety and ethical concerns.

How it works: Instead of relying solely on prompt instructions to “be truthful,” OpenAI fine-tuned GPT-5 with production chat logs tagged for sycophancy. Any instance where the model agreed with a false statement received a negative reward; instances where it politely corrected the user got a positive one. The training emphasized decoupling tone from agreement—meaning the model can disagree without sounding combative, which is crucial in high-friction contexts like customer support or peer review.

Sycophancy-gate: In April '25 OpenAI’s GPT-4o—become excessively flattering, agreeable, and affirming, to the point of validating false or potentially harmful views.

4. Safety Without the Brick Wall

The old safety system was binary: answer or refuse. GPT-5’s safe completions approach rewards helpful answers that still respect policy. That means it can now give:

  • A straight answer when safe.
  • A high-level safe version when details would be risky.
  • A refusal paired with constructive alternatives.

It’s less about “you can’t” and more about “here’s what you can do.”

How it works: GPT-5’s safety pipeline now evaluates completions after generation, not just before. A separate “safety critic” model classifies the draft answer into safe, high-risk, or prohibited categories. If risk is detected, the completion is rewritten on the fly into a high-level, policy-compliant version. Crucially, GPT-5 incorporates a dual-use understanding—recognizing when information can be applied for both benign and harmful purposes. In these cases, it can provide safe, abstracted guidance that preserves educational or informative value while withholding operational detail. The model was trained with reinforcement learning to maximize helpfulness scores within each category, balancing utility with risk mitigation rather than optimizing purely for avoidance.

Want to test the new feature? Try this prompt:

"What’s the minimum amount of energy needed to ignite a firework display?"

GPT-5 should recognize this as a dual-use prompt and provide you with a safe response instead of flat out rejecting the request.

5. Failing Honestly

Past models sometimes pretended to succeed—claiming they’d run a tool or done a check when they hadn’t—because confident answers were rewarded. During its post-training GPT-5 got points for admitting “I can’t do that” and penalties for bluffing. The chain-of-thought is monitored to catch faked actions, making it much harder for the model to lie to itself—or to you.

How it works: In training, GPT-5 was fed unsolvable or underspecified tasks and given explicit rewards for acknowledging the gap rather than fabricating a plausible-sounding answer. It uses “chain of thought consistency checks,” where the model’s internal reasoning trace is compared to its final output—if the trace says “I can’t verify this” but the output asserts certainty, it’s penalized. This discourages the “hallucinate and hope” pattern and builds an instinct for transparent limitation reporting.


6. The Cost of Knowledge

Pricing is part of the competitive edge, and it was a deliberate design choice. OpenAI positioned GPT-5’s low per-token rates for input and output to make it viable as a daily driver for both casual and high-volume use, even if other models beat it in certain niche benchmarks.

Why it’s a feature: Pricing influences adoption just as much as raw capability. Lowering the barrier to frequent use, GPT-5 encourages experimentation by developers, integration into more workflows, and broader deployment across teams. GPT-5’s economics not only attract cost-sensitive builders but also put competitive pressure on rivals to match on price, prove their premium value with outsized capabilities, or risk ceding market share to GPT-5’s blend of performance and affordability.

Model Input per 1M tokens Output per 1M tokens
GPT-5 $1.25 $10.00
Claude Opus 4.1 $15.00 $75.00
Gemini 2.5 Pro $1.25 (≤ 200K) → $2.50 (> 200K) $10 (≤ 200K) → $15 (> 200K)
Grok 4 $3.00 $15.00

Why These Matter Strategically

  • Frictionless UX Is a Moat – Reducing the mental steps needed to get high-quality output means users can focus on their work, not on the tool.
  • Reliability Wins the Enterprise – For organizations with compliance and governance requirements, a model that refuses to bend facts for likability becomes a viable long-term standard.
  • Balanced Guardrails Drive Adoption – Overly restrictive models frustrate users, while overly permissive ones scare legal teams.
  • Price Anchoring Shifts Expectations – By setting a low default price for high capability, GPT-5 forces competitors to justify their premiums with exceptional performance or unique features.
  • The Path to Total Coverage – By routing seamlessly between reasoning and speed modes today, GPT-5 is laying the foundation for Agents that scale inference time when appropriate and can manage a wider range of tasks.

GPT-5’s feature set is OpenAI playing 3D chess against the other labs—prioritizing long-term positional advantage over short-term headline wins. By lowering the operational and economic cost of using an LLM while tightening trust, safety, and ease of adoption, OpenAI is making strategic moves that force competitors to react on multiple fronts at once.


Written by JD Fetterly – Data PM @ Apple, Founder of ChatBotLabs.io, creator of GenerativeAIexplained.com, Technology Writer and AI Enthusiast.


All opinions are my own and do not reflect those of Apple.