Discussion about this post

User's avatar
Avik De's avatar

This course sounds great, and I'm also very much looking forward to the syllabus and other info! I'm also a huge fan of L4DC; thanks for your work on that.

One thing I'm maybe not understanding is the distinction you're making between optimal control and feedback (possibly just to do with definitions). E.g. LQR is an optimal controller for an LTI sytem, and outputs a linear feedback controller. If you consider a PID controller as feedback, its behavior is tied to a local Lypaunov / LaSalle function, just as an optimal controller's behavior is tied to a value function. MPC will locally estimate (and re-estimate online) the value function, and would also be considered to be a "feedback" controller. Trajectory optimization (I agree) has no feedback, but typically in robotics this trajectory will still be stabilized using a feedback controller generated using MPC or LQR. Similarly, for a learned policy generating actions from observations, domain randomization during training necessitates some amount stochastic robustness in the feedback control, which appears to be key for all the robotics behaviors being developed using RL in simulation these days.

Josh's avatar

Is there any way I can access the notes or recording or tune in live?

Doing my PhD in AI and human control processes and your article captures a lot of the concepts I'm looking into!

8 more comments...

No posts

Ready for more?