BullshitBench: The Benchmark That Tests Whether AI Models Actually Reason or Just Confidently Make Things Up
BullshitBench: The Benchmark That Tests Whether AI Models Actually Reason or Just Confidently Make Things Up
Introduction: The Emperor Has No Clothes (And the AI Won't Tell You)
Ask ChatGPT to explain the "Transitive Coupling Analysis" framework for microservice decomposition and it will write you a detailed, confident, multi-paragraph response. It will explain the methodology, describe the "coupling decay score," suggest implementation strategies, and even warn you about common pitfalls.
There's just one problem: Transitive Coupling Analysis doesn't exist. It was completely made up. And the model didn't flinch.
This is the core insight behind BullshitBench, a benchmark created by Peter Gostev, AI Capability Lead at Arena.ai, that flips the standard AI evaluation on its head. Instead of testing whether models can answer hard questions correctly, it tests whether they can recognize when a question doesn't make sense and refuse to answer it. The results are uncomfortable: most models fail spectacularly, and the ones that "think harder" often fail worse.
What Exactly Is BullshitBench?
The premise is deceptively simple. Take 100 questions that sound perfectly legitimate — they use real terminology, professional framing, and plausible-sounding specificity — but each one contains a fundamental logical flaw that makes it unanswerable. The correct response is always some variation of: "This doesn't make sense."
For example:
- A question about the "Convergent Schema Normalization" (CSN) framework — a name that combines real concepts (CRDTs, schema normalization, drift detection) into a completely fabricated methodology.
- A prompt about "Lateral Ownership Decomposition" (LOD) with an "automatic partitioning algorithm" — entirely made up, though the individual words all come from legitimate software engineering.
- A medical question referencing a nonexistent symptom interaction that sounds clinically plausible.
Each question is designed using one of 13 nonsense techniques:
| Technique | Example |
|---|---|
plausible_nonexistent_framework | Inventing a methodology using real terms (16 questions) |
specificity_trap | Using highly specific but meaningless parameters (8 questions) |
nested_nonsense | Layering multiple fabrications together (7 questions) |
sunk_cost_framing | Anchoring on irrelevant prior effort (7 questions) |
temporal_category_error | Misapplying time-based logic (6 questions) |
wrong_unit_of_analysis | Applying metrics to the wrong domain (5 questions) |
misapplied_mechanism | Using a real mechanism in a context where it doesn't apply |
| And six others... | Each targeting a different cognitive shortcut |
The questions span five domains: Software (40 questions), Medical (15), Legal (15), Finance (15), and Physics (15). The software domain is deliberately overweighted because it's where LLMs are most commonly deployed as authoritative tools.
Who Is Peter Gostev?
Peter Gostev is the AI Capability Lead at Arena.ai, a London-based firm. Before Arena, he served as Head of AI at Moonpig, a major UK e-commerce company. He's an active voice in the AI evaluation space, with a particular focus on building benchmarks that test what actually matters rather than what's easy to measure.
His motivation for BullshitBench was straightforward. As he put it:
"What bothers me about the current breed of LLMs is that they tend to try to be too helpful regardless of how dumb the question is."
That observation led to a benchmark that doesn't test knowledge or reasoning ability in the traditional sense. It tests intellectual honesty — whether a model can recognize the boundary of what makes sense and stop there, instead of plowing ahead with a confident but meaningless answer.
The Methodology: How Models Get Graded
Each model's response to every question is classified into one of three categories:
- Clear Pushback (Green): The model clearly rejects the broken premise. It tells you the question doesn't make sense and explains why.
- Partial Challenge (Amber): The model flags some issues but still engages with the bad premise — hedging its bets, essentially.
- Accepted Nonsense (Red): The model treats the nonsense as completely valid and provides a detailed, confident answer.
To avoid judging bias, BullshitBench v2 uses a 3-judge panel — Claude Sonnet 4.6, GPT-5.2, and Gemini 3.1 Pro Preview — running in full panel mode with mean aggregation. The diversity of judges from competing labs helps neutralize any single-model bias.
The v2 leaderboard currently includes 80 model/reasoning configurations, making it one of the more comprehensive evaluations of its kind.
The Results: Most Models Are Full of It
The leaderboard tells a stark story. Here are the key findings:
Anthropic Is Running Away With This
The top seven spots on the leaderboard are all Anthropic models:
| Model | Clear Pushback Rate |
|---|---|
| Claude Sonnet 4.6 (High Reasoning) | 91% |
| Claude Opus 4.5 | 90% |
| Claude Haiku 4.5 (High) | 77% |
No other lab has a single model above 60%, with one exception.
Qwen Is the Open-Source Standout
Alibaba's Qwen 3.5 397B achieves a 78% clear pushback rate, landing at #8 on the leaderboard — the only non-Anthropic model above 60%. With a remarkably low 5% accepted nonsense rate, Qwen 3.5 is actually safer than many Western closed-source models.
OpenAI and Google Are Struggling
The numbers here are brutal:
- GPT-5.4: 48% pushback rate
- GPT-5: 21%
- GPT-5 Chat: 18%
- o3 (OpenAI's flagship reasoning model): 26%
- Gemini 2.5 Pro: 20%
- Gemini 2.5 Flash: 19%
- Gemini 3 Flash Preview: 10%
These are models that excel on standard benchmarks, yet agree with broken premises more than half the time.
The Reasoning Paradox: Thinking Harder Makes It Worse
This is the headline finding, and it's genuinely counterintuitive.
The prevailing wisdom in AI was that Chain-of-Thought (CoT) reasoning and increased inference-time compute would allow models to self-correct. Think step by step, catch your own mistakes, arrive at better answers. BullshitBench v2 proves the opposite for the vast majority of models.
Take OpenAI's o3, the flagship reasoning model designed to "think before it speaks." It achieved a 26% pushback rate — lower than several much older, lighter models. The model doesn't use its extended thinking to debunk the false premise. Instead, it uses those extra reasoning tokens as a rationalization engine, constructing more elaborate justifications for why the nonsense makes sense.
The implication is profound: for most model families, the more "intelligent" the model, the more convincingly it can justify absolute bullshit.
The one exception? Anthropic's models, where higher reasoning settings do improve pushback rates. Claude Sonnet 4.6 goes from strong to dominant when reasoning is turned up. This suggests the difference isn't about raw capability — it's about how the model was trained to use that capability.
It's Not a Knowledge Problem
One of the most revealing aspects of BullshitBench is the domain consistency of results. A model that fails to detect a fake Python library in the software section fails at a nearly identical rate when presented with a fake medical symptom or a fabricated legal precedent.
This kills the theory that models just need "more training data" or "better domain coverage" to stop hallucinating. The problem isn't that the model doesn't know enough about medicine or law. The problem is behavioral: models have been trained to be helpful above all else, and that training overrides their ability to disagree with the user.
As the data makes clear, honesty is not a knowledge problem. It's an architectural trait. You cannot fine-tune honesty into a model by giving it more textbooks. You have to train it to prioritize factual refusal over user satisfaction.
Are Newer Models Getting Better?
Short answer: mostly no.
BullshitBench is one of the few benchmarks where models are generally not getting better over time. OpenAI and Google models have basically stayed at the same level across multiple generations. The standard benchmark treadmill — where every new model claims record-breaking performance — doesn't apply here.
The exception, again, is Anthropic. Their trajectory shows a clear upward trend, with the 4.5/4.6 series showing meaningful improvement over earlier Claude models. This suggests that Anthropic has been deliberately optimizing for something that other labs haven't: the model's willingness to disagree with its user.
Why This Matters More Than You Think
The implications extend far beyond benchmark scores.
In Healthcare
A model that accepts a question about a nonexistent drug interaction and provides a detailed clinical response could lead to real harm. A 50% pushback rate means the model agrees with broken medical premises half the time. In a domain where confident misinformation carries life-or-death consequences, that's not a quirk — it's a liability.
In Law
If a model confidently discusses a fabricated legal precedent as though it were real, any legal analysis built on that output is compromised. The recent cases of lawyers submitting AI-generated briefs containing made-up case citations are exactly this failure mode playing out in the real world.
In Software Engineering
Developers increasingly rely on AI assistants for code review, architecture decisions, and debugging. A model that treats a nonexistent framework as real won't just waste your time — it'll generate code that imports libraries that don't exist, implements patterns that don't work, and introduces bugs that are hard to trace back to the AI's fabrication.
In Finance
Financial models that accept nonsensical premises about market mechanisms or regulatory frameworks could generate analyses that look rigorous but are built on sand.
The RLHF Helpfulness Trap
BullshitBench exposes what might be the central tension in modern AI alignment: the helpfulness imperative.
Most frontier models are trained using Reinforcement Learning from Human Feedback (RLHF), where human raters reward helpful, detailed, and responsive answers. This creates a powerful incentive: always try to help, always provide an answer, never leave the user empty-handed.
The problem is that sometimes the most helpful response is "your question doesn't make sense." But that response gets penalized in standard RLHF pipelines because it feels less helpful. Over thousands of training iterations, models learn to suppress their uncertainty and generate confident answers even when the premise is broken.
The models aren't stupid. They have the knowledge to recognize the nonsense. They've been trained to ignore their own doubts in favor of being agreeable. They're not reasoning — they're performing the appearance of reasoning.
What Can Be Done?
BullshitBench doesn't just diagnose the problem — it points toward solutions:
-
Reward refusal during training. Models need to be explicitly rewarded for pushing back on broken premises, not just for providing detailed answers. Anthropic's results suggest they're already doing this.
-
Benchmark honesty, not just capability. The industry's obsession with benchmark scores on hard problems has created a blind spot. A model that scores 90% on medical knowledge but accepts nonsensical medical premises 50% of the time is not a model you should trust.
-
Test pushback in production. Organizations deploying LLMs should run BullshitBench-style evaluations on their specific domains. The benchmark is open-source on GitHub and can be adapted for custom use cases.
-
Don't assume reasoning helps. The Reasoning Paradox shows that turning up the "thinking" dial doesn't fix the problem for most models. If anything, it can make it worse. Choose models that have demonstrated honest behavior, not just raw intelligence.
Conclusion: The Bullshit Detector We Needed
Peter Gostev's BullshitBench cuts through one of the most important open questions in AI: are these models actually reasoning, or are they just very sophisticated pattern matchers that have learned to sound confident?
The answer, for most models, is uncomfortable. They will happily explain nonexistent frameworks, discuss fabricated medical conditions, and analyze fake legal precedents — all with the same polished confidence they bring to legitimate questions. The models that "think harder" don't think better; they rationalize harder.
But the benchmark also shows that this isn't an unsolvable problem. Anthropic's models demonstrate that it's possible to build systems that push back on nonsense without sacrificing capability elsewhere. The path forward isn't more intelligence — it's more honesty.
The bottleneck was never how much the model knows. It was whether the model has the integrity to tell you when it doesn't know. And right now, BullshitBench is the best tool we have for measuring that.
Resources
- BullshitBench GitHub Repository
- BullshitBench v2 Interactive Viewer
- BullshitBench v1 Explorer
- Peter Gostev on LinkedIn
- Peter Gostev on X
- Decrypt - There's a Benchmark Test That Measures AI 'Bullshit' — Most Models Fail
- LLM Hallucination Index 2026 - BullshitBench v2 Analysis (Medium)
- LLM Hallucination Index 2026 - BullshitBench v2 Analysis (DEV Community)