← Back to Library

Import AI 428: Jupyter agents; palisade's usb cable hacker; distributed training tools from exo

This week's landscape of artificial intelligence is defined less by headline-grabbing breakthroughs and more by the gritty, unglamorous work of making systems actually function in the real world. Jack Clark's latest analysis cuts through the hype to reveal a critical truth: the bottleneck for AI is no longer just raw computing power, but the ability of these systems to navigate physical chaos, understand scientific nuance, and be weaponized by adversaries with minimal resources. The evidence suggests we are entering an era where the most dangerous and transformative AI developments will happen in the margins of agricultural fields, USB ports, and physics textbooks.

The Reality Check on Physical Intelligence

Clark begins by dismantling the assumption that AI is ready for the physical world. He highlights a new dataset from Argentinian researchers involving a robot tasked with weeding soybean fields. Despite the task sounding simple, the data reveals a stark failure of current technology. "Real world robotics continues to be the most challenging thing for AI," Clark notes, pointing out that even basic localization and mapping systems "fail to accurately predict the correct locations, often by breaking down during the course of a run."

Import AI 428: Jupyter agents; palisade's usb cable hacker; distributed training tools from exo

The argument here is that the "sim-to-real" gap remains a massive chasm. While digital models can outperform humans at chess or Go, a robot in a soybean field struggles with the unstructured, messy variables of nature. Clark writes, "Papers like this highlight how even simple-seeming tasks, like getting a robot in a soybean field to accurately figure out where it is and map its environment, is more challenging than people might suspect." This is a crucial correction to the narrative that general-purpose robots are just around the corner. The complexity of the physical world demands a level of sensory integration and adaptability that current models simply lack.

Critics might argue that focusing on agricultural failures ignores the rapid progress in controlled environments like warehouses. However, Clark's point stands: until AI can handle the unpredictability of a farm, it cannot be trusted in the broader, unstructured human environment.

Basic inputs for useful robots are still missing; the real world is messier than any simulation.

Bridging the Gap in Scientific Workflows

Moving from physical to digital labor, Clark examines a new initiative from Hugging Face designed to make AI systems better at reading and executing code within Jupyter notebooks. This is not just about writing code; it is about understanding the context of scientific experimentation. The dataset, comprising over 50,000 synthetic notebooks, is built to train agents to answer questions like "How many total trainable parameters does the LSTM model have?" by actually running the code.

Clark argues that this approach addresses a fundamental flaw in how we measure AI intelligence: "Often, one of the problems in understanding AI capabilities is seeing how well they do when given the same tools and workflows as people." By forcing AI to interact with the actual tools scientists use, we move beyond static testing to dynamic problem-solving. The researchers behind the dataset note that "the resulting examples include natural questions about a dataset/notebook, verified answers, and step-by-step execution traces suitable for agent training."

This framing is effective because it shifts the goalpost from "can the AI guess the answer?" to "can the AI do the work?" It suggests a future where AI acts as a true research assistant rather than just a text generator. However, a counterargument worth considering is whether synthetic data can truly capture the nuance of human error and creative problem-solving found in real-world research logs.

The Optimizer Stagnation

In a move that will surprise many chasing the "next big thing" in training algorithms, Clark highlights research from the Stanford-backed organization Marin, which suggests that the industry's obsession with new optimizers may be misplaced. The study rigorously tested ten different optimizers across various model scales and found that the workhorse of the industry, Adam, remains superior.

"No optimizer achieved the 2× step-wise speedup from prior claims; the best was ≈ 1.4× over AdamW," Clark reports, quoting the researchers. This is a sobering reality check for a field often driven by marketing and incremental claims of efficiency. Clark suggests that this kind of "unglamorous but important" empirical work is exactly what the field needs to ground its expectations.

The implication is clear: the low-hanging fruit of training efficiency has been picked, and the next leaps will require fundamental architectural changes rather than just tweaking the math of the optimizer. While some might argue that 1.4x speedups are still significant at scale, the lack of a doubling in performance challenges the narrative of exponential progress in training efficiency.

The Democratization of Cyber Threats

Perhaps the most alarming segment of the analysis comes from Palisade Research, which has demonstrated an autonomous AI hacker hidden inside a USB cable. This is not a theoretical risk; it is a functional proof-of-concept where a small device, once plugged in, downloads an AI agent that interacts with a Large Language Model to guide its hacking attempts.

Clark describes the agent as sitting "between" a human and a script: "It's faster than a human but slower than a traditional script, and is similarly less adaptable than a human but more adaptable than a script." The cost is shockingly low—roughly $200 for hardware and less than a dollar per run for the AI processing. "Smart, pint-sized hacking agents are coming," Clark warns, framing this as a preview of a future where digital skills can be cloned and deployed on cheap hardware.

This development forces a re-evaluation of physical security. The barrier to entry for sophisticated cyberattacks is collapsing. While the current system is "far too discoverable and dumb," as Clark admits, the trajectory is clear. If these agents become smaller and faster, the threat landscape shifts from targeted, high-cost operations to ubiquitous, low-cost harassment and intrusion.

Systems like this have numerous limitations that mean they can't (yet) be used in the world - but if we wind the clock forward, we can imagine a future where hackers can digitally clone their relevant skills into a small model.

Lowering the Barrier to Distributed Training

On the defensive side of the equation, Clark points to EXO Gym, a new software tool that allows researchers to simulate distributed training on a single laptop. Distributed training, which involves coordinating many computers to train a model, is usually a resource-intensive process reserved for well-funded labs. EXO Gym aims to change that by making it easy to test different algorithms without the hardware overhead.

"If exo gym brings the time to try out a new distributed algo from a week down to half a day, then I hope that more people will be able to contribute to research in this field," writes developer Matt Beton, a sentiment Clark endorses. The argument is that democratizing the tools of distributed training will lead to more innovation and better de-risking of algorithms.

This is a double-edged sword. While it accelerates legitimate research, it also lowers the barrier for bad actors to experiment with large-scale model training. However, Clark's focus remains on the positive potential for the scientific community to iterate faster and more cheaply.

The New Frontier of Scientific Evaluation

Finally, Clark introduces a new benchmark, CMPhysBench, which tests AI on graduate-level condensed matter physics. The results are sobering: even the best models score below 30%. "The best ones were, in order, Grok 4, OpenAI o3, and Gemini 2.5 Pro, scoring 28.8%, 25.5%, and 23.46%," Clark notes.

The significance here is not just the low score, but the shift in evaluation standards. Clark reflects on how far the field has come: "About five years ago... the closest you get to science/math are some components of superglue... Five years on, we're evaluating frontier AI systems by testing out how well they do at condensed matter physics - we've come so, so far in such a short period of time."

The authors of the benchmark suggest that to improve, models need "physics-aware verification" and coupling with symbolic tools. This highlights a critical gap: current models are good at pattern matching but struggle with the rigorous logic and verification required in hard sciences. The lack of a clear gap between "reasoning" and "non-reasoning" models suggests that current "reasoning" techniques are not yet sufficient for deep scientific inquiry.

Bottom Line

Clark's commentary effectively grounds the AI conversation in the tangible realities of physical limitations, scientific rigor, and emerging security threats. The strongest part of the argument is the emphasis on empirical testing over hype, particularly regarding optimizers and scientific benchmarks. However, the piece slightly underplays the geopolitical implications of tools like EXO Gym, which could accelerate the development of frontier models by non-state actors. The reader should watch for how quickly the "USB cable hacker" concept evolves from a proof-of-concept into a tangible threat to physical infrastructure.

Sources

Import AI 428: Jupyter agents; palisade's usb cable hacker; distributed training tools from exo

by Jack Clark · Import AI · Read full article

Welcome to Import AI, a newsletter about AI research. Import AI runs on lattes, ramen, and feedback from readers. If you’d like to support this, please subscribe.

Soybean situational awareness:…Real world robotics continues to be the most challenging thing for AI…Argentinian researchers have released a multi-modal dataset recorded by a weed removing robot working in a soybean agricultural field. The dataset is captured by an RGB camera, stereo IR camera, a 6-Axis IMU, three 9-Axis IMU, and three GNSS receivers and wheel encoders. The dataset was gathered by a four-wheeled robot platform which is designed to automate the weeding of large crop fields.All of the gathered data was made through having the robot doing six varied runs over a soybean field, and all the data is synchronized and appropriately time-stamped. In tests, the researchers show that contemporary simultaneous localization and mapping (SLAM) systems fail to accurately predict the correct locations, often by breaking down during the course of a run.Why this matters - basic inputs for useful robots: As a rule, whenever you go into the real world, you tend to run into issues. Papers like this highlight how even simple-seeming tasks, like getting a robot in a soybean field to accurately figure out where it is and map its environment, is more challenging than people might suspect. Read more: The Rosario Dataset v2: Multimodal Dataset for Agricultural Robotics (arXiv). Get the dataset here: The Rosario Dataset v2 (GitHub).***Hugging Face makes it easier for AI systems to learn to use Jupyter notebooks:…Expect AI for science systems to get better as a consequence…Hugging Face has produced a dataset of synthetic data based on real Kaggle Jupyter notebooks, along with a test to see if AI systems can correctly answer questions about the contents of the notebooks (e.g., "How many total trainable parameters does the LSTM model have?", or "What percentage of customers with only 1 banking product eventually churned?"). This dataset can be used to train AI systems to be able to easily parse the contents of Jupyter notebooks and execute Python code to answer questions within them. This is a useful skill as Jupyter notebooks are commonly used by researchers in a wide variety of scientific and business disciplines to conduct experiments, so making AI systems better at understanding them will ultimately make AI systems more effective at accelerating the work of human scientists.Contents: The dataset contains 51,389 synthetic notebooks ...