Thomas Berrueta (California Institute of Technology)- Robot learning on the edge: Online learning in hardware
Abstract: Deploying learning algorithms on safety-critical hardware, such as autonomous race cars, remains a significant challenge. While sim-to-real transfer is a popular paradigm, persistent model mismatch and unmodeled dynamics can lead to suboptimal or unsafe behavior on high-performance hardware. This talk argues for an alternative approach: learning directly on the physical system. However, online learning in hardware introduces critical challenges—the need for sample efficiency, operational safety requirements during deployment, and the difficulties of deploying new models in real-time. First, I address the problem of sample efficiency by examining how temporal correlations can hinder online reinforcement learning. Second, I address the challenge of integrating learning modules within sensitive control stacks by designing contractive interfaces between modules. Lastly, I present a framework for safety-critical envelope expansion that allows continual deployment of learned models in real-time. I demonstrate these approaches on diverse hardware platforms, including an autonomous IndyCar capable of reaching speeds exceeding 170mph.
Speakers
Thomas Berrueta
Thomas Berrueta is an interdisciplinary roboticist and Postdoctoral Scholar at the California Institute of Technology’s Computing and Mathematical Sciences Department. As part of his research at Caltech, he leads the technical development of high-performance hardware platforms, including an autonomous racecar capable of speeds exceeding 170 mph. Thomas earned his Ph.D. at Northwestern University, where he was awarded the Presidential Fellowship and named a Microsoft Future Leader in Robotics and AI for his work on the statistical mechanics of robot learning. His work has been featured in Ars Technica, Popular Mechanics, Scientific American, Gizmodo, and Science. His research focuses on real-time learning in safety-critical systems, leveraging techniques from reinforcement learning, optimal control, and statistical physics to guarantee operational safety without sacrificing performance.