16-745: Dynamic Optimization
Instructor: Chris Atkeson, cga at cmu
TA: Akshara Rai, arai at andrew
MW 4:30-6 NSH 3002
The TA will hold office hourse from 3:30-4:30 Monday and Wednesday in NSH 4223.
Events of Interest
Last year's course
Jan 11: Introduction to the course.
Goal: Introduce course.
This years emphasis is TO BE DETERMINED
What do you want? ...
Jan 13: Function Optimization Example
Goal: Introduce you to a useful tool, MATLAB
and its optimization subroutines, and show you how to use them on an example.
Robotics: redundant inverse kinematics.
Using Matlab's fminsearch and fminunc.
Using Matlab's fminsearch and fminunc, with
Using Matlab's fmincon.
Relationship of Jacobian approach to gradient descent.
Jan 18: No Class
Jan 20: Handling 3D Orientation
Goal: Enable you to do 3D robotics using optimization (and do the inverse kinematics assignment).
Euler angles, and
Metrics for how close two orientations are:
Metrics for 3D Rotations: Comparison and Analysis,
Rigid-Body Attitude Control: Using Rotation Matrices for Continuous, Singularity-Free Control Laws,
Closed-Loop Manipulator Control Using Quaternion Feedback
Rotation matrix for small rotations
Function optimization using
order gradient methods
Goal: Review gradient descent approaches.
A nice chapter on function optimization techniques:
Numerical Recipes in C, chapter 10
(2nd or 3rd edition, 2nd edition is electronically available for free
under Obsolete Versions):
Minimization or Maximization of Functions,
This material from any other numerical methods book is also fine.
software list 1,
conjugate gradient v2,
quasi-Newton/variable metric methods, and
A Biased History of Artificial Neural Networks
Goal: Make gradient descent and the chain rule more interesting.
Rectifier units (ReLU),
Siyuan Feng: Optimization for Robot Control (and the DARPA Robotics Challenge)
Goal: Case study in the use of optimization for robot control.
Siyuan's web page with videos/papers,
Chris's web page with videos/papers,
Feb 1: Non-gradient ("derivative-free") function optimization methods:
Goal: Review non-gradient approaches.
local unimodal sampling,
Nelder Mead/Simplex/Amoeba method,
fit surfaces (for example
Response Surface Methodology (RSM),
Memory-based Stochastic Optimization, and
Derivative-free optimization: A review of algorithms and comparison of software implementations by Luis Miguel Rios and Nikolaos V. Sahinidis,
Book: Introduction to Derivative-Free Optimization
Covariance Matrix Adaptation Evolution Strategy.
Goal: Understand currently popular state of the art method.
See also Hansen web page.
Feb 3: Constraints.
Goal: Understand how to best handle constraints.
Soft/hard constraints, penalty functions,
Augmented Lagrangian method,
Interior point methods vs. Simplex methods vs. soft constraint methods,
Quadratic Programming and
Sequential quadratic programming,
Goal: Understand QP components used in state of the art robot control.
Feb 3: Automatic differentiation
Goal: Learn how taking derivatives is much easier than you thought.
Feb 8: Dynamics and Numerical Integration
Goal: Review "mental simulation".
Continous time, discrete time. Euler integration, Forward and inverse dynamics. Linearization.
Feb 8: Formulating trajectory optimization as function optimization.
Goal: Use the tools we have so far to do trajectory optimization.
Examples of formulating a trajectory optimization problem
as a function optimization problem:
Case Studies In Trajectory Optimization: Trains, Planes, And Other
Robert J. Vanderbei
Example use of AMPL
A free trial version of AMPL is available from here.
AMPL is also available for remote use through the Neos Server.
Click on SNOPT/[AMPL Input] under Nonlinearly Constrained Optimization.
Example use of Matlab: pend1-x-u,
Spacetime Optimization: Witkin paper text
Witkin paper figures
Use of splines in trajectory optimization.
Goal: Force smooth solutions.
Cubic Hermite spline.
Quintic Hermite interpolation.
Policy optimization I: Use function optimization.
Goal: Optimize feedback.
What is a policy?
Known in machine learning/reinforcement learning as policy search or refinement, ...
See examples in CMA-ES section for policy optimization.
Feb: Ways to robustify function optimization:
Goal: Tricks of the trade.
Problems: How choose method?, more of an art than a science, local minima, bad answers, discontinuities, redundant/rank deficient constraints,
bad scaling, no formulas for derivatives, you are lazy, computational cost.
Techniques: Levenberg Marquardt,
scaling and preconditioning, regularize parameters, soft constraints,
Paper on continuation methods,
Hand of God, allow constraint violations, add extra constraints,
Goal: This is what makes dynamic optimization special.
Linear Quadratic Regulator,
Goal: An important special case.
Differential Dynamic Programming
Feb: Ways to reduce the curse of dimensionality
Feb: Policy Optimization II: Optimization using model-based gradients
Robustness to random disturbances, varying initial conditions, parametric
model error, structural modeling error such as
high frequency unmodelled dynamics,
and model jumps (touchdown and liftoff during walking, for example).
Monte Carlo trajectory/policy optimization.
Feb: Receding Horizon Control (a.k.a. Model Predictive Control (MPC)).
Feb: Robustness using Linear Matrix Inequalities
Robustness to parametric uncertainty in the linear(ized) model.
I can't find a good reference on robustness using linear matrix inequalities,
but here is a tutorial on LMIs
Feb: Robustness: Policy Optimization with Multiple Models.
Monte-Carlo, DP, and DDP approaches to Multiple Models.
Gaussian Propagation (like Kalman Filter),
Unscented (like Unscented Filter), Second Order Kalman Filter (See Kendrick below).
Review of Gaussians slides
State estimation slides
Matlab Kalman filter example
minimum jerk trajectory subroutine.
Example mobile robot Kalman filter slides
March 7-11: No Class
March: Robustness and state estimation:
Linear-quadratic-Gaussian control (LQG),
Separation principle, Certainty equivalence,
Example of bad interactions, Loop Transfer Recovery (LTR),
A paper on the topic,
Policy optimization approaches.
Information state DP.
March: Local Approaches to Dual Control/Stochastic DDP
Information state trajectory optimization.
Stochastic Control for Economic Models,
David Kendrick, Second Edition 2002.
March: A*-like algorithms: R*
March: Avoiding obstacles using Sampling based methods: RRT,
Random Sampling DP
March: Avoiding obstacles using gradient methods: CHOMP
April 6: Handling contact:
Posa Talk 1pm April 8 NSH ???,,
Trajectory Optimization for Full-Body Movements with Complex Contacts
April: Learning From Demonstration
April: Reinforcement Learning: Model free policy optimization.
Kober, J.; Peters, J. (2011). Policy Search for Motor Primitives in Robotics, Machine Learning, 84, 1-2, pp.171-203
April: Comparison of various RL methods: CMA-ES, CEM, PI2.
Freek Stulp and Olivier Sigaud. Path Integral Policy Improvement with Covariance Matrix Adaptation. In Proceedings of the 29th International Conference on Machine Learning (ICML), 2012.
Trajectory optimization based on integrating the dynamics:
calculus of variations,
Discrete time Pontryagin's minimum principle,
Pontryagin's minimum principle,
multiple shooting methods,
Learning during optimization
Apr. 25: Project presentations
Apr. 27: Project presentations
May ?: Project Writeups Due
Assignment 0 (Due Jan. 17): Send CGA email:
Who are you?
Why are you here?
What research do you do?
Describe any optimization you have done (point me to papers or
web pages if they exist).
Any project ideas?
What topics would you especially like the course to cover?
Be sure your name is obvious in the email, and you mention the course
name or number. I teach more than one course, and a random email from
firstname.lastname@example.org is hard for me to process.
Assignment 1 (Due Feb. 7): Using Optimization
to do Inverse Kinematics
Assignment 2 (Due April 30): Using LQR and