← Back to Library

️‍♂️ AI agents demystified

Johnny Chang doesn't just define a buzzword; he maps the architectural shift from passive chatbots to active workers, arguing that 2025 is the inflection point where AI stops talking and starts doing. While the industry chatter focuses on novelty, Chang grounds the discussion in the tangible mechanics of how these "agents" will actually execute tasks, offering a rare clarity on the difference between a model that guesses and a system that plans.

From Chatbot to Colleague

The piece opens with a bold prediction from OpenAI's leadership, but Chang quickly pivots to the practical definition that matters for implementation. He cites TechCrunch to establish the baseline: AI agents are "AI-fueled software that does a series of jobs for you that a human customer service agent, HR person, or IT help desk employee might have done in the past." This framing is crucial because it moves the conversation away from abstract intelligence toward specific labor substitution.

️‍♂️ AI agents demystified

Chang then leverages a particularly vivid analogy from Google's white paper to explain the cognitive leap required for true agency. "Just as a chef creates dishes through an iterative process of gathering information about their available ingredients and customer preferences, reasoning through their recipes, and taking action to cook the dish, agents are designed to achieve their goals by processing information, making informed decisions, and refining actions based on feedback or previous outputs." This comparison effectively demystifies the "black box" of AI, replacing it with a logical workflow that professionals can visualize.

The distinction Chang draws between standard language models and agents is the article's analytical core. He notes that while models like ChatGPT generate responses using existing training data, agents are "more dynamic and can connect to external tools and APIs to access live data." This capability allows them to "plan and execute tasks autonomously," a feature that makes them uniquely suited for complex, real-world scenarios where static knowledge is insufficient. Chang's breakdown of the four-step process—Perceive, Reason, Act, and Learn—provides a necessary structural framework for understanding how these systems operate without human intervention at every step.

"The model also uses reasoning and logic frameworks such as ReAct, Chain-of-Thought, or Tree-of-Thoughts to reason and plan systematically."

This technical detail is vital, yet Chang wisely contextualizes it by explaining the "data flywheel" effect. He describes how agents "get better over time by learning from each interaction," creating a feedback loop that improves efficiency and accuracy. Critics might argue that this optimism overlooks the significant risks of autonomous systems making errors in critical domains, but Chang acknowledges the necessity of "guardrails" to limit independent action, suggesting a balanced view of the technology's potential and its constraints.

The Corporate Reality Check

Moving from theory to practice, Chang surveys how major corporations are already integrating these tools, providing concrete evidence that this is not merely a future projection. He highlights Johnson & Johnson using agents in chemical synthesis to aid drug discovery, and Moody's deploying a network of 35 interconnected agents for financial analysis. These examples serve to validate the "chef" analogy, showing that the iterative process of gathering data and refining outputs is already driving value in high-stakes industries.

The coverage of eBay and Deutsche Telekom further illustrates the breadth of application, from coding and marketing to internal policy queries. Chang notes that eBay is "planning on implementing them in the future to help customers find items and sellers list their products," signaling a shift from internal efficiency to customer-facing transformation. This section effectively demonstrates that the "workforce" mentioned in the opening hook is already being assembled, albeit in pilot programs and specific departments.

Education at the Crossroads

The most compelling portion of Chang's analysis focuses on the educational sector, where the stakes involve not just efficiency but the fundamental nature of learning. He outlines three distinct use cases: a personalized study coach, a laboratory safety supervisor, and a historical reenactment agent. The vision of an agent that "analyzes a student's learning patterns, test results, and homework completion rates to create customized study schedules" promises a level of personalization that human teachers simply cannot scale.

However, Chang does not shy away from the complexities. He references a study from Peking University and Tsinghua University regarding "Learning-By-Teaching with ChatGPT," which found that while interacting with a teachable agent improved students' coding skills, the tool's tendency to produce "error-free outputs limited opportunities for learners to develop debugging skills." This is a critical nuance often missed in the rush to adopt AI. Chang synthesizes student opinion on this trade-off, noting that "intuitive coding skills have become far more crucial" because debugging can be supported by tools, yet the loss of struggle in the learning process remains a valid concern.

The piece also highlights the student-led push for policy, mentioning the Princeton SPIA students who formed an AI Policy Task Force to address cybersecurity and misinformation. This inclusion grounds the technological discussion in the human element, reminding readers that the next generation is already demanding a say in how these systems are governed.

"Students may submit their 400-word reflection to the short essay contest here. The deadline has been extended to January 18th."

While Chang celebrates the potential for "interactive" and "personalized" learning, a counterargument worth considering is the data privacy implications he briefly mentions. The ability of an agent to monitor a student's every move in real-time to provide safety alerts or personalized feedback creates a surveillance infrastructure within the classroom that warrants deeper scrutiny than the article allows.

Multimodality and Global Access

The commentary concludes by touching on the rapid expansion of AI capabilities beyond text, specifically OpenAI's Sora video generation tool. Chang notes the tool's ability to create "high-definition video clips from text prompts," positioning it as a major leap in multimodality. However, he also includes the friction this has caused, citing protests over "insufficient transparency and support for the arts" and allegations of unpaid labor. This balanced reporting adds necessary weight to the technological triumph, acknowledging the human cost of training data.

Finally, Chang broadens the scope to global equity, highlighting Google's partnership with UNESCO to launch CogLabs in Africa. This program allows students to "build AI models" using 3D-printed parts and smartphones without internet access. By emphasizing "sustainable development, gender equality, and quality education," Chang frames AI not just as a tool for corporate profit or academic efficiency, but as a potential lever for closing the global digital divide.

Bottom Line

Johnny Chang's piece succeeds by stripping away the hype to reveal the mechanical reality of AI agents: they are not magic, but structured systems of perception, reasoning, and action that are already reshaping industries. The strongest element is the nuanced treatment of education, where the promise of personalization is weighed against the risk of eroding essential problem-solving skills. The biggest vulnerability remains the unaddressed depth of data privacy concerns, which will likely become the primary battleground as these agents move from pilot programs to widespread adoption.

Sources

️‍♂️ AI agents demystified

by Johnny Chang · AI x Education · Read full article

Happy New Year! As we enter 2025, many are calling it the year of AI agents. OpenAI’s CEO Sam Altman's latest blog post states, "We believe that, in 2025, we may see the first AI agents 'join the workforce' and materially change the output of companies." But what exactly are these AI agents, and how will they transform our world? In this edition, we will explore the concept of AI agents, answer your questions along the way, and share insights into their potential impact.

Here is an overview of today’s newsletter:

Understand what AI Agents are and their potential impact on education

Learn how students are influencing AI policy recommendations

Explore ChatGPT’s potential as a teachable agent in programming education

Discover OpenAI’s latest video generation tool called Sora

Practical AI Usage and Policies.

What exactly are AI agents?

TechCrunch shared a simple definition for AI agents as “AI-fueled software that does a series of jobs for you that a human customer service agent, HR person, or IT help desk employee might have done in the past, although it could ultimately involve any task. You ask it to do things, and it does them for you.“

In Google's white paper titled Agents, they compare an AI agent to a chef. Just as a chef creates dishes through an iterative process of gathering information about their available ingredients and customer preferences, reasoning through their recipes, and taking action to cook the dish, agents are designed to achieve their goals by processing information, making informed decisions, and refining actions based on feedback or previous outputs.

Check out this short 6-minute video to learn the basics of AI agents:

How are AI agents different from standard language models like ChatGPT?

While models like ChatGPT generate responses using their existing training data, agents are more dynamic and can connect to external tools and APIs to access live data. They can even plan and execute tasks autonomously. This makes them better for handling complex, real-world scenarios when they need to adapt to different situations and use up-to-date information to make decisions.

How do AI agents work?

Broadly speaking, AI agents use a four-step process for problem-solving as described in this blog by Nvidia:

Perceive: The AI agent collects data from multiple sources, such as sensors and databases, and analyzes this information to identify important patterns and relevant details.

Reason: A large language model works behind the scenes ...