Toggle light / dark theme

Brett Adcock: Humanoids Run on Neural Net, Autonomous Manufacturing, and $50 Trillion Market #229

Humanoid robots with full-body autonomy are rapidly advancing and are expected to create a $50 trillion market, transforming industries, economy, and daily life ## ## Questions to inspire discussion.

Neural Network Architecture & Control.

đŸ€– Q: How does Figure 3’s neural network control differ from traditional robotics? A: Figure 3 uses end-to-end neural networks for full-body control, manipulation, and room-scale planning, replacing the previous C++-based control stack entirely, with System Zero being a fully learned reinforcement learning controller running with no code on the robot.

🎯 Q: What enables Figure 3’s high-frequency motor control for complex tasks? A: Palm cameras and onboard inference enable high-frequency torque control of 40+ motors for complex bimanual tasks, replanning, and error recovery in dynamic environments, representing a significant improvement over previous models.

🔄 Q: How does Figure’s data-driven approach create competitive advantage? A: Data accumulation and neural net retraining provides competitive advantage over traditional C++ code, allowing rapid iteration and improvement, with positive transfer observed as diverse knowledge enables emergent generalization with larger pre-training datasets.

🧠 Q: Where is the robot’s compute located and why? A: The brain-like compute unit is in the head for sensors and heat dissipation, while the torso contains the majority of onboard computation, with potential for latex or silicone face for human-like interaction.

Rethinking Memory Mechanisms of Foundation Agents in the Second Half: A Survey

The research of artificial intelligence is undergoing a paradigm shift from prioritizing model innovations over benchmark scores towards emphasizing problem definition and rigorous real-world evaluation. As the field enters the “second half,” the central challenge becomes real utility in long-horizon, dynamic, and user-dependent environments, where agents face context explosion and must continuously accumulate, manage, and selectively reuse large volumes of information across extended interactions. Memory, with hundreds of papers released this year, therefore emerges as the critical solution to fill the utility gap. In this survey, we provide a unified view of foundation agent memory along three dimensions: memory substrate (internal and external), cognitive mechanism (episodic, semantic, sensory, working, and procedural), and memory subject (agent- and user-centric). We then analyze how memory is instantiated and operated under different agent topologies and highlight learning policies over memory operations. Finally, we review evaluation benchmarks and metrics for assessing memory utility, and outline various open challenges and future directions.

Using synthetic biology and AI to address global antimicrobial resistance threat

James J. Collins, the Termeer Professor of Medical Engineering and Science at MIT and faculty co-lead of the Abdul Latif Jameel Clinic for Machine Learning in Health, is embarking on a multidisciplinary research project that applies synthetic biology and generative artificial intelligence to the growing global threat of antimicrobial resistance (AMR).

The research project is sponsored by Jameel Research, part of the Abdul Latif Jameel International network. The initial three-year, $3 million research project in MIT’s Department of Biological Engineering and Institute of Medical Engineering and Science focuses on developing and validating programmable antibacterials against key pathogens.

AMR — driven by the overuse and misuse of antibiotics — has accelerated the rise of drug-resistant infections, while the development of new antibacterial tools has slowed. The impact is felt worldwide, especially in low-and middle-income countries, where limited diagnostic infrastructure causes delays or ineffective treatment.

Digital twin reveals how eye cells lose their organization in leading cause of vision loss

National Institutes of Health (NIH) researchers have developed a digital replica of crucial eye cells, providing a new tool for studying how the cells organize themselves when they are healthy and affected by diseases. The platform opens a new door for therapeutic discovery for blinding diseases such as age-related macular degeneration (AMD), a leading cause of vision loss in people over 50. The study is published in the journal npj Artificial Intelligence.

“This work represents the first-ever subcellular resolution digital twin of a differentiated human primary cell, demonstrating how the eye is an ideal proving ground for developing methods that could be used more generally in biomedical research,” Kapil Bharti, Ph.D., scientific director at the NIH’s National Eye Institute (NEI).

The researchers created a highly detailed, 3D data-driven digital twin of retinal pigment epithelial (RPE) cells, which perform vital recycling and supportive roles to light-sensing photoreceptors in the retina. In diseases such as AMD, RPE cells die, which eventually leads to the death of photoreceptor cells, causing loss of vision.

Bio-inspired chip helps robots and self-driving cars react faster to movement

Robots and self-driving cars could soon benefit from a new kind of brain-inspired hardware that can allegedly detect movement and react faster than a human. A new study published in the journal Nature Communications details how an international team built their neuromorphic temporal-attention hardware system to speed up automated driving decisions.

The problem with current robotic vision and self-driving vehicles is a significant delay in processing what they see. While today’s top AI programs can recognize objects accurately, the calculations are so complex that they can take up to half a second to complete. That may not sound like a lot, but at highway speeds, even a one-second delay means a car travels 27 meters before it even begins to react. That is too long and too slow a reaction time.

To solve this problem, the team worked on a hardware solution rather than tinkering with software, modeling it on how human vision works. When we view a situation, our visual system doesn’t analyze every detail at once. It first detects changes in brightness and movement, then processes the more complex details later.

Ammonia leaks can be spotted in under two seconds using new alveoli-inspired droplet sensor

Researchers from Guangxi University, China have developed a new gas sensor that detects ammonia with a record speed of 1.4 seconds. The sensor’s design mimics the structure of alveoli—the tiny air sacs in human lungs—while relying on a triboelectric nanogenerator (TENG) that converts mechanical energy into electrical energy. The sensor uses a process that is driven by A-droplets, which are tiny water droplets containing a trapped air bubble. These droplets exploit ammonia’s affinity for water to rapidly capture NH₃ when it is present.

When an ammonia-laden droplet falls onto the sensor, its mechanical impact completes an electrical circuit, generating signals that are converted into accurate gas measurements at a speed that exceeds existing ammonia gas sensors.

To take detection precision a step further, the team integrated an AI model that analyzes electrical signals and converts them into time-frequency images. After training on these images, the system classified ammonia into five concentration levels (0–200 ppm), achieving up to 98.4% detection accuracy.

AI method accelerates liquid simulations by learning fundamental physical relationships

Researchers at the University of Bayreuth have developed a method using artificial intelligence that can significantly speed up the calculation of liquid properties. The AI approach predicts the chemical potential—an indispensable quantity for describing liquids in thermodynamic equilibrium. The researchers present their findings in a new study published in Physical Review Letters.

Many common AI methods are based on the principle of supervised machine learning: a model—for instance, a neural network—is specifically trained to predict a particular target quantity directly. One example that illustrates this approach is image recognition, where the AI system is shown numerous images in which it is known whether or not a cat is depicted. On this basis, the system learns to identify cats in new, previously unseen images.

“However, such a direct approach is difficult in the case of the chemical potential, because determining it usually requires computationally expensive algorithms,” says Prof. Dr. Matthias Schmidt, Chair of Theoretical Physics II at the University of Bayreuth. He and his research associate Dr. Florian SammĂŒller address this challenge with their newly developed AI method. It is based on a neural network that incorporates the theoretical structure of liquids—and more generally, of soft matter—allowing it to predict their properties with great accuracy.

/* */