← Back to Library

"The AI con" con

Bentham's Bulldog delivers a scathing, evidence-driven takedown of a prominent critique claiming artificial intelligence is a hollow hype cycle. The piece is notable not just for its rebuttal, but for exposing how a specific ideological framework can blind even intelligent observers to rapid technological reality. In an era where policy and investment hinge on understanding what machines can actually do, this commentary forces a reckoning with the gap between academic theory and functional capability.

The Stagnant Thesis

The core of the argument rests on the observation that the critique being dismantled—Emily Bender and Alex Hanna's The AI Con—reads like a relic from 2020. Bentham's Bulldog writes, "Reading Emily Bender and Alex Hanna's The AI Con is like time-travelling back to 2020." The author points out that while the original text accurately described the limitations of early large language models, it has failed to account for the exponential leap in capability that has occurred since. This is a crucial distinction; it suggests the critique is not merely wrong, but obsolete, relying on a snapshot of technology that no longer exists.

"The AI con" con

The original authors argue that AI is fundamentally a marketing trick, a "stochastic parrot" that merely mimics human speech without understanding. Bentham's Bulldog counters this by highlighting the sheer utility of modern systems. "The Bender and Hanna thesis is that AI is massively overhyped. That it's a mostly useless technology with little upside and massive downside," the commentator notes. Yet, the evidence presented by Bentham's Bulldog suggests the opposite: AI is already inventing novel math proofs, automating complex coding tasks, and outperforming humans in specific cognitive domains. The author argues that maintaining the position that AI will replace jobs without improving productivity has "grown less plausible since, having had a head-on collision with the facts."

The AI Con is what you get when a thesis you've been stochastically parroting for years is decisively disproven by the evidence: it's a desperate and error-filled attempt to rescue a deeply implausible thesis.

This framing is powerful because it shifts the debate from philosophy to empirical performance. However, critics might note that dismissing the original authors' concerns about environmental cost and inequality entirely risks ignoring valid, long-term structural risks that aren't immediately visible in a chatbot's output. Bentham's Bulldog acknowledges the book is "very well written" and engaging, even if the logic is flawed, which adds a layer of nuance to the critique.

The Definition Trap

A significant portion of the commentary targets the original authors' refusal to define intelligence or consciousness in a way that allows for machine cognition. Bentham's Bulldog highlights a particularly telling exchange where the original authors dismiss the possibility of machine sentience by claiming they "don't have conversations with people who don't posit my humanity as an axiom of the conversation." The commentator finds this response baffling, noting, "What a weird response."

The piece draws a sharp parallel to the history of philosophy to strengthen its point. Bentham's Bulldog reminds readers that John Searle's famous "Chinese room" argument, often cited by skeptics, was described by the original authors as "an extremely othering way of making the argument." This historical context is vital; it shows how the original authors are using philosophical tools not to clarify, but to exclude. Bentham's Bulldog writes, "They suggest that a better term for AI would be 'stochastic parrots,' or 'a racist pile of linear algebra.' Okay." The commentator then dismantles this by pointing out that human neurons also work through pattern association, citing the principle that "Neurons that fire together, wire together."

The argument here is that if we accept human intelligence as a result of physical, deterministic processes, we cannot logically deny the same potential to machines that exhibit similar outputs. Bentham's Bulldog argues that the original authors' insistence on a unique, undefinable human essence is a rhetorical shield. "The fact that it's hard to define some property doesn't mean it can't be possessed," the commentator asserts, comparing it to the difficulty of defining "explosion" or "knowledge" without denying their existence.

The Cost of Ideology

Perhaps the most contentious part of the commentary addresses the original authors' tendency to conflate technical critiques with political identity. Bentham's Bulldog points out that when the original authors are challenged on the similarity between human and machine cognition, they retreat to claims of racism. "But while AI boosters have spent time devaluing what it means to be human, the sharpest and clearest critiques have come from Black, brown, poor, queer, and disabled scholars and activists," the original text is quoted. Bentham's Bulldog finds this move unpersuasive, arguing that the fact that AI systems sometimes fail to recognize darker skin tones does not prove that the underlying mechanism of intelligence is fundamentally different from human cognition.

The commentary also tackles the original authors' dismissal of economic viability and scientific progress. Bentham's Bulldog notes the irony in claiming AI is useless while simultaneously admitting it is replacing writers and filling gaps in the workforce. "It cannot be both that AI can nicely replace writers and that it's useless," the commentator writes. This contradiction is central to the piece's thesis: the original authors are so committed to a narrative of AI failure that they ignore the very successes they describe.

If AI produces useless slop, then how is it replacing writers?

The piece also addresses the original authors' rejection of AI safety research, or "alignment," by claiming that AI development is not inevitable. Bentham's Bulldog finds this logic flawed, comparing it to arguing that bioweapons research isn't inevitable, so we shouldn't plan for the consequences if they are developed. The commentator argues that "You don't have to think that, say, development is bioweapons is inevitable to think that it's worth having a plan for what happens if we develop bioweapons."

Bottom Line

Bentham's Bulldog's strongest asset is its relentless focus on empirical reality over philosophical abstraction, effectively dismantling the claim that AI is merely a "stochastic parrot" by pointing to its ability to generate novel proofs and automate complex tasks. However, the piece's biggest vulnerability is its tendency to dismiss the socio-economic and environmental concerns raised by the original authors as mere ideological posturing, potentially underestimating the real-world costs of rapid scaling. Readers should watch for how the gap between current AI capabilities and the original authors' predictions continues to widen, forcing a necessary update to the policy and ethical frameworks surrounding the technology.

Deep Dives

Explore these related deep dives:

  • Chinese room

    The excerpt discusses Searle's famous Chinese room argument as a key concept from the book

  • John Searle (philosopher)

    The philosopher whose Chinese room argument is specifically named and discussed in the text

Sources

"The AI con" con

by Bentham's Bulldog · · Read full article

Reading Emily Bender and Alex Hanna’s The AI Con is like time-travelling back to 2020. Many passages are a stark reminder of the world of peak woke—they described Searle’s famous Chinese room argument as “an extremely othering way of making the argument.” More importantly, the picture they paint of AI capabilities would have been an accurate enough description of the AIs of 2020, but seemingly has not advanced since then.

Perhaps it is no surprise that the authors admit to not using the technology they discuss.

The Bender and Hanna (B&H) thesis is that AI is massively overhyped. That it’s a mostly useless technology with little upside and massive downside. It will devastate the environment, use up enormous quantities of water, exacerbate inequality, and so on, all without having any real benefit. It will not assist seriously with scientific advancements, nor will it be economically viable. The authors seem to curiously maintain the position that it will replace a number of jobs but won’t improve the productivity of jobs.

This position was not plausible in 2020, and has grown less plausible since, having had a head-on collision with the facts. AI can invent novel math proofs that impress the best mathematician in the world. It can automate away big chunks of coding. They can assist writing far better than a secretary. In seconds, by asking LLMs, you can get a detailed and well-researched answer with sources. They can almost instantly complete tasks that take people hours.

The AI Con is what you get when a thesis you’ve been stochastically parroting for years is decisively disproven by the evidence: it’s a desperate and error-filled attempt to rescue a deeply implausible thesis. Nearly everything in the book is poorly reasoned or misleading. It is also written with the confidence of someone expounding trivialities, rather than arguing for a contentious thesis. The authors rarely saw the need to respond to objections.

The best thing I can say about the book is it’s very well written. B&H are engaging writers. Some bits are quite funny (e.g. “Meta’s LeCun beclowned himself.”) I rarely found myself bored when reading (though this is perhaps because it is hard to be bored when encountering outrageous falsehoods on nearly every page).

The case for AI being a big deal is pretty straightforward. AI is advancing rapidly. It can already write better than most undergraduates, and possesses a great degree ...