Locally weighted Learning (also known as memorybased learning, instancebased learning, lazylearning, and closely related to kernel density estimation, similarity searching and casebased reasoning).
Locally weighted learning is simple but appealing, both intuitively and statistically. And it has been around since the turn of the century. When you want to predict what is going to happen in the future, you simply reach into a database of all your previous experiences, grab some similar experiences, combine them (perhaps by a weighted average that weights more similar experiences more strongly) and use the combination to make a prediction, do a regression, or many other more sophisticated operations. We like this approach to learning, especially for learning process dynamics or robot dynamics, because it is very flexible (low bias) so provided we have plenty of data we will eventually get an accurate model.
Simple easy to use locally weighted regression implementation
GMBLMAIN and GGMBLMAIN: Efficient locally weighted learning packages.
Vizier: a commercial locally weighted learning system
Locally weighted learning applied to control of processes and dynamic systems.
We have also developed new nonlinear controls algorithms for exploiting instancebased models that can predict in realtime. We are interested in high accuracy in very finegrained and nonlinear control tasks such as robot billiards (better than 0.1% accuracy needed) that linearized control methods would be unable to attain.
Our instancebased learning systems have also been applied to robot juggling and many industrial dynamics problems such as "how are viscosity, moisture, temperature, and two dozen other variables related during the stages of my cooking process?”
A paper that overviews how locally weighted learning can be applied to process control.
Autonomous parameter tuning and model selection with crossvalidation
Aggressive model selection is important to our stated goal of autonomous statistics, and we have pushed it hard. This has led to practical cascaded approaches to crossvalidation in which searches for good models occur over increasingly large "onionpeel"style spaces of algorithms. Within each layer, models are selected by leaveoneout crossvalidation. Which is the best layer? That is judged by an outer level of crossvalidation on further independent testsets.
We have pushed model selection further with "racing" algorithms that evaluate many models in parallel and use blocked comparisons to, in the early stages, prune away unpromising models that, with very high confidence, will not eventually evaluate as the best.
A early paper that overviews the practical issues of intense crossvalidation (new research, algorithms and analysis due shortly...)
GMBLMAIN, GGMBLMAIN and Vizier: software systems that tune themselves with intense crossvalidation.
A short paper on "Racing": the use of blocking techniques to accelerate intense model selection.


Decision and Reinforcement Learning 