Projects
Our work is organized into three areas: Learning for Control, Control for Learning, and Applications to Robotics.
Control for Learning
Short blurb: Many learning algorithms hinge on gradient-based methods—regression and classification, backpropagation, and policy gradients in reinforcement learning all rely on them. In data-driven settings, however, gradients are difficult to estimate accurately due to measurement noise, numerical error, and inexact subroutines. We address this by bringing control-theoretic ideas to learning, ensuring that gradient methods remain robust to these imperfections.
Goal: Treat gradient-based algorithms as dynamical systems and connect the Polyak–Łojasiewicz (PL) inequalities to the input-to-state stability (ISS) of gradient-based methods. See references J1, J2, J3, and J4 for details.
Learning for Control
Short blurb: Classical optimal control often assumes accurate system models, which are hard to obtain—and even when available, solving for the optimal controller can be intractable. We integrate reinforcement learning with optimal control to develop data-driven, adaptive methods that learn near-optimal controllers directly from interaction data, without requiring explicit model knowledge.
Goal: Time-delay and output-feedback systems arise widely in connected and autonomous vehicles, communication networks, and robotics. We develop reinforcement learning algorithms for these two classes that guarantee algorithmic convergence and provide closed-loop stability and near-optimal performance. See references J5, J6, and C1 for details.
Applications to Robotics
Short blurb: We apply tools from two complementary fields—reinforcement learning and control—to wheeled-bipedal robots J6, autonomous vehicles J7, and autonomous bicycles C2.