← Back to Library

AI companies are pivoting from creating gods to building products. Good

Arvind Narayanan cuts through the trillion-dollar hype to deliver a sobering truth: the AI industry's greatest failure isn't a lack of capability, but a fundamental misunderstanding of what makes software actually useful. While the market fixates on the race to build "gods," Narayanan argues the real story is a painful, necessary pivot toward boring, reliable product-market fit. This is essential listening for anyone trying to separate the noise of speculation from the signal of actual utility.

The Pivot from Gods to Gadgets

Narayanan diagnoses the current chaos as a result of two equally flawed strategies. On one side, research labs like OpenAI and Anthropic treated their models as ends in themselves, neglecting the user experience. On the other, giants like Microsoft and Google panicked, shoving AI into every product without asking if it belonged there. "The generality of LLMs allowed developers to fool themselves into thinking that they were exempt from the need to find a product-market fit," Narayanan writes. This observation is sharp because it exposes a dangerous arrogance: the belief that a smart model can replace careful design.

AI companies are pivoting from creating gods to building products. Good

The consequences of this arrogance are visible in the public perception of the technology. Early adopters were disproportionately "bad actors" because they were the only ones willing to wrestle with raw, unpolished tools, while everyday users were left with "occasionally useful and more often annoying" features. Narayanan notes that "OpenAI seems to be transitioning from a research lab focused on a speculative future to something resembling a regular product company." This shift is framed not as a moral awakening, but as a survival strategy. The author suggests that if you strip away the boardroom drama, the core story is simply about "creating gods to building products."

Critics might argue that this pivot stifles the very innovation that makes AI revolutionary, but Narayanan counters that without solving basic usability, there is no revolution to have. The market is forcing a correction that the industry refused to make voluntarily.

If you take all the human-interest elements out of the OpenAI boardroom drama, it was fundamentally about the company's shift from creating gods to building products.

The Five Barriers to Reality

Even as companies pivot, Narayanan outlines five specific hurdles that stand between current AI and commercial success. The first is cost. While prices have dropped, the author warns that "cost improvements directly translate to accuracy improvements" because cheaper models allow for more retries to overcome randomness. This reframes the "too cheap to meter" narrative, suggesting that the cheapest model might actually be the most expensive if it requires a thousand attempts to get a single right answer.

The second barrier, reliability, is perhaps the most critical for consumer trust. Narayanan draws a sharp distinction between capability and reliability: "If an AI system performs a task correctly 90% of the time, we can say that it is capable of performing the task but it cannot do so reliably." This is a crucial insight for business leaders who might mistake a 90% success rate for a viable product. In traditional software, a 10% failure rate is unacceptable; in AI, it's currently the norm. The author argues that "companies will have to adapt AI to user expectations instead, and make AI behave like traditional software."

Privacy and security present the next set of challenges. Narayanan points out that while training data was once public, the new wave of assistants requires deep access to private user data to be useful. He highlights the controversy around Microsoft's attempt to take screenshots of user PCs, noting that "there was an outcry and the company backtracked." This suggests that technical feasibility does not equal social acceptance. Furthermore, the author warns that while accidental failures are fixable, security vulnerabilities like "AI worms" are a looming threat that companies are currently ignoring.

Finally, the user interface remains a bottleneck. Narayanan envisions a future where AI is invisible, perhaps integrated into glasses, but notes that "the constrained user interface leaves very little room for incorrect or unexpected behavior." This creates a paradox: the more seamless the experience, the less room there is for the model to hallucinate.

If your AI travel agent books vacations to the correct destination only 90% of the time, it won't be successful.

The Long Road Ahead

Narayanan concludes by pushing back against the "trend extrapolation" that promises massive societal shifts within a year or two. He argues that the challenges are "sociotechnical and not purely technical," meaning they involve human behavior, trust, and workflow integration, not just better code. "We should expect this to happen on a timescale of a decade or more rather than a year or two," he writes. This is a necessary reality check for an industry addicted to quarterly hype cycles.

The author's skepticism is grounded in the observation that even if the technology improves, organizations must still "train people to use it productively while avoiding its pitfalls." This human element is often the missing variable in AI forecasts. While some might argue that rapid iteration will solve these issues faster than predicted, Narayanan's experience suggests that the friction between stochastic AI and deterministic human needs will slow progress significantly.

Bottom Line

Narayanan's strongest contribution is reframing the AI bubble not as a financial failure, but as a product design crisis that is only now being addressed. His biggest vulnerability is the assumption that the industry will successfully execute this pivot without causing significant collateral damage to user trust along the way. Readers should watch for whether the shift toward reliability actually translates to products that people can depend on, or if the race for scale will continue to outpace the need for safety. The decade-long timeline he proposes is a sobering metric for anyone expecting an immediate revolution.

Sources

AI companies are pivoting from creating gods to building products. Good

by Arvind Narayanan · AI Snake Oil · Read full article

AI companies are collectively planning to spend a trillion dollars on hardware and data centers, but there’s been relatively little to show for it so far. This has led to a chorus of concerns that generative AI is a bubble. We won’t offer any predictions on what’s about to happen. But we think we have a solid diagnosis of how things got to this point in the first place.

In this post, we explain the mistakes that AI companies have made and how they have been trying to correct them. Then we will talk about five barriers they still have to overcome in order to make generative AI commercially successful enough to justify the investment.

Product-market fit.

When ChatGPT launched, people found a thousand unexpected uses for it. This got AI developers overexcited. They completely misunderstood the market, underestimating the huge gap between proofs of concept and reliable products. This misunderstanding led to two opposing but equally flawed approaches to commercializing LLMs. 

OpenAI and Anthropic focused on building models and not worrying about products. For example, it took 6 months for OpenAI to bother to release a ChatGPT iOS app and 8 months for an Android app!

Google and Microsoft shoved AI into everything in a panicked race, without thinking about which products would actually benefit from AI and how they should be integrated.

Both groups of companies forgot the “make something people want” mantra. The generality of LLMs allowed developers to fool themselves into thinking that they were exempt from the need to find a product-market fit, as if prompting a model to perform a task is a replacement for carefully designed products or features.

OpenAI and Anthropic’s DIY approach meant that early adopters of LLMs disproportionately tended to be bad actors, since they are more invested in figuring out how to adapt new technologies for their purposes, whereas everyday users want easy-to-use products. This has contributed to a poor public perception of the technology.1

Meanwhile the AI-in-your-face approach by Microsoft and Google has led to features that are occasionally useful and more often annoying. It also led to many unforced errors due to inadequate testing like Microsoft's early Sydney chatbot and Google's Gemini image generator. This has also caused a backlash.

But companies are changing their ways. OpenAI seems to be transitioning from a research lab focused on a speculative future to something resembling a regular product company. ...