← Back to Library

The z.ai playbook

This piece cuts through the noise of "China vs. US" AI narratives to reveal a startling reality: Chinese developers are not just catching up, they are actively reverse-engineering Silicon Valley's cultural capital to build superior products. Jordan Schneider's interview with Zixuan Li exposes a recursive feedback loop where Chinese engineers obsessively track American influencers to refine their own models, a strategy that has propelled Z.ai's GLM 4.5 to the top of global benchmarks. For the busy professional, this is not just a tech update; it is a masterclass in how a constrained ecosystem can outmaneuver a dominant one through agility and open-source leverage.

The Architecture of Ambition

Schneider frames the conversation around a specific pivot: the shift from generic chatbots to specialized, agentic tools. The core argument is that Z.ai's success isn't accidental but the result of a deliberate architectural choice to distill multiple specialized models into a single, unified system. Li explains that the team first built three separate "teaching models" before merging them, a process that required unprecedented internal alignment. "Everyone is working on a single target," Li states, noting that even the heads of pre-training and fine-tuning teams sit next to each other to ensure a unified vision for reasoning and coding capabilities.

The z.ai playbook

This collaborative intensity stands in stark contrast to the siloed structures often found in larger, more bureaucratic organizations. The result, as Schneider highlights, is a model that doesn't just chat but acts. "We are no longer putting simple chat at the top of our priorities. Instead, we are exploring more on the coding side and the agent side," Li admits, signaling a strategic move toward economic utility over conversational novelty. This focus on "agentic tool use" is what allowed GLM 4.5 to punch above its weight, ranking in the top tier of the Chatbot Arena despite the company's relative obscurity in the West.

"You need to feel the trend yourself. You need to combine the results from experiments and the trends — what's going on within your competitors' teams — to feel the move yourself."

The commentary here is sharp: Li argues that leadership cannot be detached from the code. In an era where models evolve weekly, a CEO or CTO who delegates training entirely is already obsolete. This mirrors the "knowledge distillation" concepts explored in related deep dives, where the transfer of expertise from larger, complex systems into smaller, efficient ones is key. Here, however, the distillation is cultural as much as technical; the team is distilling the intent of global trends into their local architecture. Critics might argue that this "founder-led" intensity is unsustainable at scale, but the current output suggests it is the only viable path for a startup competing against giants.

The Open Source Gambit

Perhaps the most provocative section of Schneider's coverage is the revelation that open-sourcing is a survival strategy for Chinese firms, not just an ideological stance. The logic is pragmatic: without open weights, Western enterprises will not adopt the technology. "It's not easy to get famous in the United States because people just don't accept your API. They need to be stored in the US," Li explains. By releasing models like GLM to the public, Z.ai bypasses geopolitical friction and gains a foothold in the global developer ecosystem.

This strategy was refined by observing the rise of DeepSeek. Li notes, "We also learned from DeepSeek because our flagship model was closed source back in 2024. But when DeepSeek R1 launched, we realized that you can be really famous for open sourcing your model while getting some business return through API or other collaborations." The goal is to "expand the cake first and then take a bite of it." This reframes the open-source movement from a purely altruistic academic exercise into a sophisticated market-entry tactic.

"If we don't open source our models, we'll never have an opportunity to join this conversation."

The implication is profound: the global AI conversation is currently gated by code availability. By withholding models, Chinese companies risk irrelevance in the very markets they hope to influence. Schneider effectively highlights how this creates a paradox where the most "closed" nation in terms of internet policy is producing some of the most "open" AI models to gain legitimacy. This aligns with the broader history of open-source software, where community adoption often drives faster innovation than proprietary walled gardens, a dynamic that Tsinghua University researchers have long championed.

The Recursive Echo Chamber

Schneider astutely identifies a unique cultural phenomenon: the recursive nature of the Chinese AI discourse. Chinese media and developers are not just reacting to American trends; they are curating them. Li points out that the team monitors American "Key Opinion Leaders" (KOLs) like Andrej Karpathy and Sam Altman to gauge which models are gaining traction. "All the social media will try to grasp their core ideas immediately," Li says, describing a feedback loop where Chinese engineers build what the West says is valuable, which then gets reported back to the West, creating a self-fulfilling prophecy of global relevance.

This dynamic creates a strange inversion where a model's success in the US is often a prerequisite for its success in China. "Chinese enterprises are still paying attention to the global brand and your global performance," Li notes, suggesting that domestic buyers trust a model more if Silicon Valley validates it. This challenges the narrative of a bifurcated tech world. Instead, there is a single, albeit tense, global market where reputation is the primary currency.

"We need to combine with these top products to gain fame."

The commentary here is essential: it suggests that the "China AI" label is no longer a barrier but a variable in a global equation. The fear of job loss and AI pessimism, which Li mentions as prevalent in China, is being managed by focusing on tools that demonstrably increase efficiency, such as coding assistants. This pragmatic approach avoids the philosophical debates that often stall Western adoption, focusing instead on the immediate utility of the technology.

Bottom Line

Schneider's coverage succeeds by stripping away the geopolitical theater to reveal the raw mechanics of competition: agility, open-source leverage, and a relentless focus on utility. The strongest part of the argument is the revelation that open-sourcing is a calculated business move to bypass Western skepticism, a strategy that is rapidly reshaping the global AI landscape. The biggest vulnerability remains the dependency on Western validation; if the global community turns away, the recursive loop breaks. The reader should watch whether Z.ai can maintain its momentum as the "open" phase transitions into the "monetization" phase, a test that will determine if this playbook is a one-off success or a new standard for the industry.

Deep Dives

Explore these related deep dives:

  • Knowledge distillation

    The article describes Z.ai's technique of building three separate 'teaching models' and distilling them into GLM 4.5. Knowledge distillation is a foundational machine learning technique that readers would benefit from understanding to grasp how modern AI labs efficiently create unified models.

  • Tsinghua University

    Zhipu AI was founded by researchers from Tsinghua University in 2019, and the article discusses the close relationship between Chinese AI companies and academic institutions. Understanding Tsinghua's role as China's premier technical university provides context for the talent pipeline feeding Chinese AI development.

  • Open-source software

    The article explicitly discusses 'the role of open source in the Chinese AI ecosystem' as a key topic and mentions Z.ai's relationship with open-source coding tools. Understanding the philosophy and economics of open source helps readers grasp the strategic decisions Chinese AI companies make about model release.

Sources

The z.ai playbook

by Jordan Schneider · ChinaTalk · Read full article

Zixuan Li is Director of Product and genAI Strategy at Z.ai (also known as Zhipu 智谱 AI). The release of their benchmark-topping flagship model, GLM 4.5, was akin to “another DeepSeek moment,” in the words of Nathan Lambert.

Our conversation today covers…

What sets Z.ai apart from other Chinese models, including coding, role-playing capabilities, and translations of cryptic Chinese internet content,

Why Chinese AI companies chase recognition from Silicon Valley thought leaders,

The role of open source in the Chinese AI ecosystem,

Fears of job loss and the prevalence of AI pessimism in China,

How Z.ai trains its models, and what capabilities the company is targeting next.

Co-hosting today are Irene Zhang, long-time ChinaTalk analyst, as well asNathan Lambert of the Interconnects Substack.

Listen now on your favorite podcast app.

The Z.ai Model and Chinese Open Source.

Jordan Schneider: Zixuan, could you introduce yourself?

Zixuan Li: Hi everyone, I’m Zixuan Li from Z.ai. I manage a lot of things, like global partnerships, Z.ai chat model evaluation, and our API services. If you’ve heard of the GLM Coding Plan, I’m actually in charge of that, too. I studied AI for science and AI safety at MIT, where I did research on AI applications and AI alignment.

Jordan Schneider: Let’s do a little bit of Zhipu AI’s backstory. When was it founded? How would you place it within the broader landscape of teams developing models in China?

Zixuan Li: Zhipu AI and Z.ai were founded in 2019, and we were chasing AGI at that time, but not with LLMs, but with some graphic network or graphic compute. We did something similar to Google Scholar called AMiner. We used that type of thing to connect all the data resources from journals and research papers into a database. People could easily search and map these scholars and their contributions. It was very popular at that time.

However, we shifted to the exploration of large language models in 2020. We launched our paper, GLM, in 2021. I believe that was about one year ahead of the launch of GPT-3.5, so it was a very, very early stage. We were one of the first companies to explore large language models. After that, we continuously improved the performance of our models and tried a new architecture. GLM is a new architecture, actually, but we’re going to explore more in the future.

I believe we became famous with ...