| Members | Meetings | Robots | The Robot Learning Laboratory | Research | Publications | Resources |

Robot Learning Talks @ CMU

We have informal meetings almost every Thursday at 5pm in Wean Hall room 7220.

The purpose of the meetings is to make people of the robot learning community of CMU and visitors aware of interesting papers and results that have come to the attention of the speakers, and also for the informal exchange of ideas.

Food is provided :)

If you have any questions concerning our meetings, please contact Geoff Gordon at "ggordon AT cs DOT cmu DOT edu". If you'd like to suggest a change to this web page, please contact Joelle Pineau or Max Likhachev at "maxim+ AT cs DOT cmu DOT edu".



 
Wed
10/15/03
12.00pm
NSH
1505
Geoff Gordon Solving extensive-form games and playing one-card poker

 

Spring 2003

Here is a tentative schedule of upcoming presentations.
 
Thu 07/31/03
5.00pm
Wean 7220
Drew Bagnell Covariant Policy Search (IJCAI'03)

Most standard policy search reinforcement learning algorithms exhibit non-covariant behavior during the learning process. That is, the "steepest descent" rule depends on the exact parameterization of a policy class we are considering, and not simply on the policy class itself. One may analyze these algorithms in information-geometric terms and discover that there is indeed a version of "steepest descent" that is natural and covariant. Although the approach we describe here is motivated by purely theoretical considerations, the resulting algorithm obeys Vapnik's maxim that "Nothing is more practical than a good theory": in practice, the resulting covariant algorithm has lead to the best known performance in some benchmarks RL applications and often outperforms canonical gradient methods by orders of magnitude.  

Thu 07/24/03
5.00pm
Wean 7220
Brendan McMahan Planning in the Presence of Cost Functions Controlled by an Adversary (ICML'03)

We investigate methods for planning in a Markov Decision Process where the cost function is chosen by an adversary after we fix our policy. As a running example, we consider a robot path planning problem where costs are influenced by sensors that an adversary places in the environment. We formulate the problem as a zero-sum matrix game where rows correspond to deterministic policies for the planning player and columns correspond to cost vectors the adversary can select. For a fixed cost vector, fast algorithms (such as value iteration) are available for solving MDPs. We develop efficient algorithms for matrix games where such best response oracles exist. We show that for our path planning problem these algorithms are at least an order of magnitude faster than direct solution of the linear programming formulation.  

Thu 07/10/03
5.00pm
Wean 7220
Geoff Gordon Tutorial on Support Vector Machines

 

Thu 07/03/03
5.00pm
Wean 4623
Vandi Verma Variable Resolution Particle Filter

Particle filters are used extensively for tracking the state of nonlinear dynamic systems. This paper presents a new particle filter that maintains samples in the state space at dynamically varying resolution for computational efficiency. Resolution within statespace varies by region, depending on the belief that the true state lies within each region. Where belief is strong, resolution is fine. Where belief is low, resolution is coarse, abstracting multiple similar states together. The resolution of the statespace is dynamically updated as the belief changes. The proposed algorithm makes an explicit bias-variance tradeoff to select between maintaining samples in a biased generalization of a region of state space versus in a high variance specialization at fine resolution. Samples are maintained at a coarser resolution when the bias introduced by the generalization to a coarse resolution is outweighed by the gain in terms of reduction in variance, and at a finer resolution when it is not. Maintaining samples in abstraction prevents potential hypotheses from being eliminated prematurely for lack of a sufficient number of particles. Empirical results show that our variable resolution particle filter requires significantly lower computation for performance comparable to a classical particle filter.  

Thu 06/19/03
5.00pm
Wean 4623
Mike Montemerlo Thesis practice talk

 

Thu 05/08/03
5.00pm
Wean 7220
Vandi Verma Efficient Monitoring for Planetary Rovers

Planetary rovers operate in environments where human intervention is expensive, slow, unreliable, or impossible. It is therefore essential to monitor the behavior of these robots so that contingencies may be addressed before they result in catastrophic failures. This monitoring needs to be efficient since there is limited computational power available on rovers. We propose an efficient particle filter for monitoring faults that combines the Unscented Kalman Filter (UKF) [7] and the Variable Resolution Particle Filter (VRPF) [16]. We begin by using the UKF to obtain an improved proposal distribution for a particle filter which tracks discrete fault variables as part of its state space. This requires computing an unscented transform for every particle and every possible discrete transition to a fault or nominal state at each instant in time. Since there are potentially a large number of faults that may occur at any instant, this approach does not scale well. We use the VRPF to address this concern. The VRPF tracks abstract states that may represent single states or sets of states. There are many fewer transitions between states when they are represented in abstraction. We show that the VRPF in conjunction with a UKF proposal improves performance and may potentially be used in large state spaces. Experimental results show a significant improvement in efficiency.  

Thu 05/01/03
5.30pm
Wean 7220
Dieter Fox People Tracking and Multi-Robot Exploration

This talk consists of two parts. In the first part I will present some recent work we did on particle filter-based people tracking. In the context of tracking multiple people using a combination of anonymous and id sensors installed throughout the environment, particles represent assignments between objects and observations (similar to multi hypothesis tracking). To estimate the location of people using GPS (outdoors) or very noisy, sparse indoor data, we project particles onto graph structures. Such a representation has two key advantages. First, less samples are needed to estimate a location on the graph. Second, the graph provides a discretization of continuous trajectories, thereby making the learning of long term behaviour patterns much easier. In the second part I will describe a hierarchical Bayesian approach used for multi robot exploration. The technique allows multiple robots to merge their maps even under complete uncertainty about their relative start locations.  

Thu 03/20/03
5.00pm
Wean 7220
Trey Smith Brenda Ng, Leonid Peshkin, Avi Pfeffer, "Factored Particles for Scalable Monitoring," UAI-02

This paper presents a new family of approximate monitoring algorithms that combines the best qualities of the particle filtering and Boyen-Koller methods. Our algorithms maintain an approximate representation the belief state in the form of sets of factored particles, that correspond to samples of clusters of state variables. Empirical results show that the factored particle method outperforms ordinary particle filtering.  

Thu 03/13/03
5.00pm
Wean 7220
Brendan McMahan I'll discuss an algorithm for online convex programming, and try to motivate it with some interesting problems from a robotics perspective, including an online shortest path problem. This is work being done here at CMU by Martin Zinkevich:

Martin Zinkevich, "Online Convex Programming and Generalized Infinitesimal Gradient Ascent," CMU Technical Report CMU-CS-03-110.

Convex programming involves a convex set F and a convex functionc:F->R. The goal of convex programming is to find a point in F which minimizes c. In this paper, we introduce online convex programming. In online convex programming, the convex set is known in advance, but in each step of some repeated optimization problem, one must select a point in F before seeing the cost function for that step. This can be used to model factory production, farm production, and many other industrial optimization problems where one is unaware of the value of the items produced until they have already been constructed. We introduce an algorithm for this domain, apply it to repeated games, and show that it is really a generalization of infinitesimal gradient ascent, and the results here imply that generalized infinitesimal gradient ascent (GIGA) is universally consistent.  

Thu 03/06/03
5.00pm
Wean 7220
Joelle Pineau and Geoff Gordon Belief compression in POMDPs; Poisson PCA.

 

Thu 2/6/03
5.00pm
Wean 7220
Drew Bagnell H-infinity control

Thu 1/30/03
5.00pm
Wean 7220
Meeting cancelled.

 

Thu 1/23/03
5.00pm
Wean 7220
Allison Bruce Atkeson and Morimoto, "Nonparametric Representation of Policies and Value Functions: A Trajectory-Based Approach", NIPS 2002. (Cont'd.)

Second half of last week's talk.

Thu 1/16/03
5.00pm
Wean 7220
Allison Bruce Atkeson and Morimoto, "Nonparametric Representation of Policies and Value Functions: A Trajectory-Based Approach", NIPS 2002

This paper describes a method for reinforcement learning in which policies and value functions are represented along selected trajectories. Local linear models are used to estimate the dynamics of the system. Given a trajectory, the first and second derivatives of the value function and the first derivatives of the policy can be updated in a backwards sweep that is derived from the discrete time Riccati equations. In addition to explaining this approach and describing its application to two simple tasks (pendulum swinging and hopping), I'll attempt to relate it to some of the control theory that we learned last semester.

Additional reading:

Atkeson, "Using Local Trajectory Optimizers to Speed Up Global Optimization in Dynamic Programming", NIPS 94.

Stengel, R. F., "Optimal Control and Estimation", pp.270-284.

Fall 2002
Thu 12/05/02
5.00pm
Wean 7220
Geoff Gordon Designing a simple controller using the tricks of feedback linearization and proportional feedback.

 

Thu 11/28/02
5.00pm
Wean 7220
No meeting - Thanksgiving.

 

Thu 11/21/02
5.00pm
Wean 7220
Meeting cancelled.

 

Thu 11/14/02
5.00pm
Wean 7220
Rosemary Emery-Montemerlo Optimality Conditions for control of linear problems.

Stengel, R. F., "Optimal Control and Estimation", Section 3.4.

Thu 11/7/02
5.00pm
Wean 7220
Andrew Stein and Dave Ferguson How to derive linear controllers and Matlab demo of adaptive robot arm. [Slides&pictures(.tgz)]

 

Thu 10/31/02
5.00pm
Wean 7220
Mike Montemerlo More fun with control: Bode plots, root locus, frequency response and block diagrams.

Stengel, R. F., "Optimal Control and Estimation", Chapter 2.

Thu 10/24/02
5.00pm
Wean 7220
Matt Rosencrantz How to find a deterministic optimal trajectory.

Stengel, R. F., "Optimal Control and Estimation", Sections 3.1-3.4.

Thu 10/17/02
5.00pm
Wean 7220
Brendan McMahan Behavior of linear ODEs and how to solve them using Laplace transform methods. [SLIDES (.pdf)]

Stengel, R. F., "Optimal Control and Estimation", Sections 2.3, 2.5, 2.6. Examples 2.6.1.

Thu 10/10/02
5.00pm
Wean 7220
No meeting.

 

Thu 10/3/02
5.00pm
Wean 7220
Drago Anguelov Learning Object Maps from Laser Range Data

In this talk I will motivate the need for learning symbolic maps that represent the world in terms of objects. I will describe an algorithm that learns a map in terms of high-level objects -- walls and doors -- rather than in terms of occupancy grids. The algorithm uses a probabilistic generative model, that describes the shape and motion properties of the objects in the environment. It also includes a realistic sensor model, which addresses the phenomenon of geometric occlusion and specifies how the objects in the environment generate the robot's laser range readings. I will show how the algorithm uses EM to construct a map of an office environment containing both static walls and dynamic doors.

Thu 9/26/02
5.00pm
Wean 7220
Rahul Biswas A Hierarchical Non-Stationary Object Mapping Algorithm for Mobile Robots

I will present some of our recent work in mapping environments where objects change location over time. Using a set of static maps collected by a mobile robot, we learn high-fidelity shape models of both objects and classes of objects. I will discuss our techniques for leveraging multiple observations to improve models, solving the data association problem between observations and obejcts, and clustering objects into classes. I will also show a technique for determining how many objects and classes exist even when only a subset of objects is present in any map.

Spring 2002
Thu 05/16/02
5.00pm
Wean 7220
Rosemary Emery and Mike Montemerlo ICRA redux

 

Thu 05/10/02
5.00pm
Wean 7220
Rosemary Emery TBA

 

Thu 05/03/02
5.00pm
Wean 7220
No meeting.

 

Thu 04/25/02
5.00pm
Wean 7220
Mike Montemerlo Michael Littman. "PROVERB: the Probabilistic Cruciverbalist"

 

Thu 04/09/02
5.00pm
Wean 7220
Allison Bruce Tim Oates, Laura Firoiu, and Paul R. Cohen. Clustering time series with hidden markov models and dynamic time warping. In working notes of the IJCAI-99 workshop on Neural, Symbolic and Reinforcement Learning Methods for Sequence Learning, pages 17-21, 1999.

This paper presents a method for automatically determining K, the number of generating HMMs, and for learning the parameters of those HMMs. An initial estimate of K is obtained by unsupervised clustering of the time series using dynamic time warping (DTW) to assess similarity. In addition to producing an estimate of K, this process yields an initial partitioning of the data. As later sections will explain, DTW is related to HMM training algorithms but is weaker in several respects. Therefore, the clusters based on DTW are likely to contain mistakes. These initial clusters serve as input to a process that trains one HMM on each cluster and iteratively moves time series between clusters based on their likelihoods given the various HMMs. Ultimately the process converges to a final clustering of the data and a generative model (the HMMs) for each of the clusters.  

Thu 02/28/02
5.00pm
Wean 7220
Sungwoo Park In this talk, I'll show you several works on extending traditional framework of programming languages to incorporate probabilistic expressions, and introduce my work in this direction, whose main idea is to use sampling functions to represent probabilistic quantities.

Traditional framework of programming languages assumes that sufficient information is available at the time of actual computation. However, it fails to model systems in which uncertainty is inherent. In order to model such systems, various probabilistic modeling languages have been developed. The problem with all these approaches is that they are primarily built upon discrete computational framework only. As a consequence, it is hard to encode continuous probabilistic quantities, which frequently arise in the robotics applications, under this discrete computation framework. In this talk, I will introduce a different approach to express probabilistic quantities. The main idea is simple: instead of probability measures, we exploit sampling functions, that is, measurable mappings from Borel sets to sample domains. This approach brings both discrete and continuous probability distributions (and even weird distributions which do not belong to either class) under a unified framework of computation. I will also discuss an application of this approach to a real robotics problem, and its drawbacks (and hopefully strengths).  

Thu 02/21/02
5.00pm
Wean 7220
Tom Minka Most informative projections

Visualization allows quick determination of an appropriate data model, and projection is an intuitive way to visualize high-dimensional data. I will present new projection techniques for classification and regression data. In contrast to unsupervised methods like PCA, these projections deliberately try to preserve the dependence or `mutual information' between the input and output variables, thereby revealing what sort of classifier or regression function you should be using. For classification, they are typically much more informative than Fisher projection, which assumes the classes are Gaussian with equal covariance. Furthermore, with the right choice of objective function, the optimal projections can be computed quickly.  

Thu 01/31/02
5.00pm
Wean 7220
Nick Roy "Algorithms for Inverse Reinforcement Learning" Andrew Ng and Stuart Russell

This paper addresses the problem of inverse reinforcement learning IRL in Markov decision processes, that is, the problem of extracting a reward function given observed, optimal behavior. IRL may be useful for apprenticeship learning to acquire skilled behavior, and for ascertaining the reward function being optimized by a natural system. We first characterize the set of all reward functions for which a given policy is optimal. We then derive three algorithms for IRL. The first two deal with the case where the entire policy is known; we handle tabulated reward functions on a finite state space and linear functional approximation of the reward function over a potentially infinite state space. The third algorithm deals with the more realistic case in which the policy is known only through a finite set of observed trajectories. In all cases, a key issue is degeneracy -- the existence of a large set of reward functions for which the observed policy is optimal. To remove degeneracy, we suggest some natural heuristics that attempt to pick a reward function that maximally differentiates the observed policy from other, sub-optimal policies. This results in an efficiently solvable linear programming formulation of the IRL problem. We demonstrate our algorithms on simple discrete/finite and continuous/infinite state problems.  

Thu 01/24/02
5.00pm
Wean 7220
David Hershberger Randomized Kinodynamic Planning (1999) Steven M. LaValle, James J. Kuffner, Jr.

Abstract: This paper presents a state-space perspective on the kinodynamic planning problem, and introduces a randomized path planning technique that computes collision-free kinodynamic trajectories for high degreeof -freedom problems. By using a state space formulation, the kinodynamic planning problem is treated as a 2n-dimensional nonholonomic planning problem, derived from an n-dimensional configuration space. The state space serves the same role as the configuration space for basic path planning; however, standard randomized path planning techniques do not directly apply to planning trajectories in the state space. We have developed a randomized planning approach that is particularly tailored to kinodynamic problems in state spaces, although it also applies to standard nonholonomic and holonomic planning problems. The basis for this approach is the construction of a tree that attempts to rapidly and uniformly explore the state space, offering benefits that are similar to those obtained by successful randomized planning methods, but applies to a much broader class of problems. Some preliminary results are discussed for an implementation that determines kinodynamic trajectories for hovercrafts and satellites in cluttered environments, resulting in state spaces of up to 12 dimensions.  

Thu 01/17/02
5.00pm
Wean 7220
Chuck Rosenberg Partially Labeled Classification with Markov Random Walks (2001) Martin Szummer and Tommi Jaakkola

To classify a large number of unlabeled examples we combine a limited number of labeled examples with a Markov random walk representation over the unlabeled examples. The random walk representation exploits any low dimensional structure in the data in a robust, probabilistic manner. We develop and compare several estimation criteria/algorithms suited to this representation. This includes in particular multi-way classification with an average margin criterion which permits a closed form solution. The time scale of the random walk regularizes the representation and can be set through a margin-based criterion favoring unambiguous classification. We also extend this basic regularization by adapting time scales for individual examples. We demonstrate the approach on synthetic examples and on text classification problems.  

Fall 2001

We have a draft schedule of upcoming presentations.
 
Thu 11/29/01
5.00pm
Wean 7220
Vandi Verma Risk-sensitive particle filters We propose a particle filter that incorporates a model of cost when generating particles. The approach is motivated by the observation that the cost of accidentally not tracking hypotheses might be significant in some areas of state space, and irrelevant in others. By incorporating a cost model into particle filtering, states that are more critical to the system performance are more likely to be tracked. Automatic calculation of the cost model is implemented using an MDP value function calculation that estimates the value of tracking a particular state. Experiments in two mobile robot domains illustrate the appropriateness of the approach.  

Thu 11/15/01
5.00pm
Wean 7220
No meeting  

Thu 11/8/01
5.00pm
Wean 7220
No meeting  

Thu 11/1/01
5.00pm
Wean 7220
John Langford True Error bounds for Neural Networks I will describe a technique for bounding the true error of a stochastic neural network using only training examples. This technique has been succesfully tested resulting in quantitatively meaningful true error bounds with just 100 learning examples. This is joint work with Rich Caruana.  

Thu 10/25/01
5.00pm
Wean 7220
Mike Montemerlo Rao-Blackwellised Particle Filtering for Dynamic Bayesian Networks  

Thu 10/18/01
5.00pm
Wean 7220
Allison Bruce How to make robots more likeable I'll be presenting the results of an experiment that I did this fall that relates a robot's use of expressiveness and motion to its success at performing a social task. I'll also talk a little bit about some work I'm doing now to try to use HMMs to recognize simple behaviors (such as approach and avoidance) so a robot can predict how a person will respond to it and behave appropriately.  

Thu 10/11/01
5.00pm
Wean 7220
Dimitris Margaritis A Bayesian Multiresolution Independence Test for Continuous Variables In this paper I present a method of computing the posterior probability of conditional independence of two or more continuous variables from data, examined at several resolutions. The motivation for this kind of test is in learning the structure of Bayesian networks by algorithms that use independence tests to infer the network structure, in domain s that contain any mix of continuous, ordinal and categorical variables.  

Thu 10/4/01
5.00pm
Wean 7220
No meeting  

Thu 9/27/01
5.00pm
Wean 7220
Joelle Pineau Policy invariance under reward transformations: Theory and application to reward shaping Paper by: Andrew Y. Ng, Daishi Harada, Stuart Russell.  

Thu 9/20/01
5.00pm
Wean 7220
Geoff Gordon Brainstorming about Naive Bayes
Thu 9/4/01
5.00pm
Wean 7220
Doug Aberdeen Gradient Ascent on POMDP Policy Graphs Two new gradient descent algorithms for learning to use finite-state controllers as memory in POMDPs.  Comparisons to older methods.  Heaven, Hell, and robot navigation.  

Thu 8/30/01
5.00pm
Wean 7220
Geoff Gordon Notes on conditioning I'll talk about a paper from last NIPS which compares gradient descent to Kalman filtering as a model of different types of conditioning in animals.  Conclusion: Kalman filtering predicts actual behavior better in backward blocking experiments.  

Spring 2001
 
Thu 4/26/01
5.00pm
Wean 7220
Jaimie Schulte Multiagent planning under uncertainty
Thu 4/19/01
5.00pm
Wean 7220
Eric Klavins Decentralized Algorithms for Cyclic Robotic Systems
Thu 4/12/01
5.00pm
Wean 7220
Mike MCMC
Thu 4/5/01
5.00pm
Wean 7220
John Langford Decision Theoretic Particle Filters
Thu 3/29/01
5.00pm
Wean 7220
Dimitris Margaritis UAI
Thu 3/22/01
5.00pm
Wean 7220
Sungwoo Park Frob
Thu 3/15/01
5.00pm
Wean 7220
Greg Armstrong How robot's work
Thu 3/8/01
5.00pm
Wean 7220
No Meeting
Thu 3/1/01
5.00pm
Wean 7220
No Meeting
Thu 2/22/01
5.00pm
Wean 7220
No meeting
Thu 2/15/01
5.00pm
Wean 7220
Vandi Verma Bayesian Fault Detection and Diagnosis in Dynamic Systems
Thu 2/8/01
5.00pm
Wean 7220
Nick Roy Active Learning
Thu 2/1/01
5.00pm
Wean 7220
Frank Dellaert RANSAC I will describe RANSAC, a stochastic search algorithm that is used a lot in computer vision to find a good set of correspondences in the context of motion estimation.
Thu 1/25/01
5.00pm
Wean 7220
Tom Minka ``Mixture Kalman filters''  

and Rao-Blackwellization from ``Sequential Monte Carlo methods for Dynamic Systems'' and ``On Sequential Simulation-Based Methods for Bayesian Filtering''  

Fall 2000
 
Thu 12/7/00
5.00pm
Wean 7220
Dimitris Margaritis ``Approximate Bayes Net Inference''  

I'll present a paper of Daphne Koller, Lerner and Angelov from UAI-99: "A General Algorithm for Approximate Inference and its Applications to Hybrid Bayes Nets".

Thu 11/30/00   NIPS*2000 (no meeting)
Thu 11/23/00   Thanksgiving break (no meeting)
Thu 11/16/00
4.30pm
Wean 7220
Joelle Pineau ``Hierarchical Policy Constraint for POMDP Planning and Execution''  

POMDPs provide a useful framework for decision-making in the presence of uncertainty, however finding an optimal policy is computationally infeasible for large problems. I will present a hierarchical approach to POMDPs which takes advantage of structure in the domain to find policies for complex tasks. The idea is to divide the action set along domain-specific task decompositions. A POMDP problem is thereby broken-down into a hierachically related set of smaller problems (i.e. subtasks). A value function is learned locally (i.e. related to the local set of actions) for each subtask, thereby yielding local policies.

This extends the work on hierarchical reinforcement learning (Dietterich, NIPS'99) to the POMDP framework; the task decomposition and state abstraction are similar, however I propose significantly different planning and execution algorithms to ensure consistent belief states between subtasks, and to accomodate partial observability of subtask completion. This work is applied to the problem of dialogue management, for which I will present results showing that planning time was significantly reduced compared to a conventional POMDP, while the policy performed much better than when using a greedy heuristic approximation.

Thu 11/9/00
5.00pm
Wean 7220
Alexander Gray ``A Tutorial on Gaussian Processes''  

I'll explain what Gaussian processes are, and why they are interesting. I may also talk about Gaussian process networks (which won the ICML-2000 best paper award), and how to overcome the computational problem of learning Gaussian processes using exciting and fabulous new CMU algorithms!

Thu 11/2/00
5.00pm
Wean 7220
John Langford ``A Tutorial on Wavelets in Statistics''  

Thu 10/26/00
5.00pm
Wean 7220
Chuck Rosenberg ``Semi-supervised Pattern Recognition Approaches in Vision''  

An overview of two papers:

Recognition of Planar Object Classes  
M.C. Burl and P. Perona  
CVPR 96, San Francisco, CA, June 1996  
ftp://vision.caltech.edu/pub/tech-reports-vision/CVPR96-recog.ps.gz
Unsupervised Learning of Models for Recognition  
M. Weber, M. Welling and P. Perona  
ECCV 2000, Dublin, Ireland, June 2000
ftp://vision.caltech.edu/pub/tech-reports/ECCV00-recog.ps.gz
Thu 10/19/00
5.00pm
Wean 7220
Daniel Nikovski ``PEGASUS Reinforcement Learning''  

An overview of the ICML-2000 paper by Ng and Jordan.

Spring 2000
 
Thu 5/11/00
5.00pm
Wean 7220
Liam Pedersen ``Bayes Networks on Ice: The Robotic Search for Antarctic Meteorites''  

In January 2000 the Nomad robot was deployed to the ice sheets of Antarctica and made the first ever autonomous field identification of a meteorite by a robot. Nomad's rock/meteorite classifier is based on a Bayes network which addresses several issues unique to classifying rocks from a mobile robotic vehicle: the ambiguity of rock classes, the non-trivial costs of deploying sensors, noisy data, limited training examples and the need to incorporate human expert opinions.

I'll also describe the Robotic Antarctic Meteorite Search project and the latest results from the recent expedition 2000 to the Elephant Moraine in Antarctica.

Thu 5/4/00
5.00pm
Wean 7220
John Langford ``Making Sample Complexity Bounds Quantitatively Useful''  

Thu 4/27/00
5.00pm
Wean 7220
Nick Roy ``Finding Approximate POMDP Solutions through Belief Compression''  

Human beings take for granted their ability to move around and interact with each other in a wide range of environments. These same environments are tremendous sources of uncertainty for mobile robots. Not only does navigation introduce positional uncertainty, but systems that try to interact with human beings are faced with the tremendously noisy and ambiguous behaviours that humans exhibit.

The Partially Observable Markov Decision Process (POMDP) constitutes a particular planning methodology that attempts to model uncertainty in the way that a system acts and senses the world. POMDPs are unfortunately useless for most real world problems due to the overwhelming computational complexity involved with planning in belief spaces. I propose an method for finding an approximation of the optimal POMDP policy by compressing the full belief state into statistics that represent the belief space compactly, and demonstrate the utility of this technique in the domains of mobile robot navigation and speech dialogue management.

Thu 4/20/00
5.00pm
Wean 7220
Eric Bauer ``Boosted Mixture Models''  

Thu 4/13/00
5.00pm
Wean 7220
Daniel Nikovski ``Pattern discovery via entropy minimization, parameter extinction, and deterministic annealing''  

I will present three papers by Matthew Brand on several novel techniques for learning probabilistic models and pattern discovery. Entropic priors are derived from the basic premise that the world is predictable, and actually make sense, unlike most other priors people have been using. Parameter extinction is a related technique that produces sparse probability tables and might even make the learned models interpretable by humans. Finally, deterministic annealing overcomes shallow local maxima in likelihood and produces quasi-global maxima, which improves the quality of the models. The combination of these techniques results in a method for learning the structure of probabilistic models, which is much better and faster than the usual test-and-discard approach.

The three papers are available online, if you want to check them out in advance: http://www.merl.com/projects/structure/index.html

Thu 4/6/00
5.00pm
Wean 7220
Dieter Fox ``Multi-Hypothesis Tracking''  

I will present some of the work done in Stuart Russell's group on traffic surveillance. One of the key problems in this context is to detect that cars observed at different cameras are actually the same car (object identification). They apply techniques similar to those used in the data association community to solve this problem.

Suggested Paper:

Hanna Pasula, Stuart Russell, Michael Ostland, and Ya'acov Ritov,
Tracking many objects with many sensors. In Proc. IJCAI-99, Stockholm, 1999.

Thu 3/23/00
5.00pm
Wean 7220
John Langford ``Improvements in Multi-Robot Localization''  

Thu 3/16/00
5.00pm
Wean 4625
Mike Gross ``PERSES: A Vision-Based Interactive Mobile Shopping Assistant''  

Mobile robot with omnicam vision and gesture-based interaction.

Thu 3/9/00
5.00pm
Wean 7220
Frank Dellaert ``Structure from Motion with EM''  

Thu 3/2/00
5.00pm
Wean 7220
Sebastian Thrun ``SVM Tutorial''  

Thu 2/10/00
5.00pm
Wean 7220
Joelle Pineau ``Introduction to Hierarchical Reinforcement Learning''  

Suggested Papers:

Dietterich, T. G., 1998, The MAXQ method for hierarchical reinforcement learning, ICML 1998.

Dayan, P. and Hinton, G. E., 1993, Feudal Reinforcement Learning, NIPS 5.

Parr, R. and Russell, S., 1997 Reinforcement Learning with Hierarchies of Machines, NIPS 1997.

In this talk I will present an introduction to hierarchical reinforcement learning. Hierarchical decomposition of reinforcement learning (RL) problems has been used to accelerate learning in large-scale MDP problems by exploiting structure in the problem domain. The components of the hierarchy typically consist of related subtasks with varying degrees of specialization, which can independently learn optimal policies. The papers listed above each present a different approach to hierarchical RL. I will explain the general motivation, characteristics, advantages and limitations of hierarchical RL, and will present a few examples drawn from the Dietterich(1998) and Dayan&Hinton(1993) papers.
Thu 2/3/00
5.00pm
Wean 7220
Daniel Nikovski ``An Introduction to Variational Methods for Graphical Models''  

Paper by M. I. Jordan, Z. Ghahramani, T.S. Jaakkola and L.K. Saul

In M.I. Jordan (editor) Learning in Graphical Models, p. 105, Dordrecht: Kluwer Academic Publishers (1998).

This paper presents a tutorial introduction to the use of variational methods for inference and learning in graphical models. We present a number of examples of graphical models, including QMR-DT database, the sigmoid belief network, the Boltzmann machine, and several variants of hidden Markov models, in which it is infeasible to run exact inference algorithms. We then introduct variational methods, showing how upper and lower bounds can be found for local probabilities, and discussing methods for extending these bounds to bounds on global probabilities of interest. Finally we return to the examples and demonstrate how variational algorithms can be formulated in each case.


Talks from previous semesters are linked below:

Retreats!

ALL

1999 Spring , Fall
1998 Spring , Fall
1997 Fall


Dimitris Margaritis (last update April 8, 2002)