As we work to move robots out of factories and into human environments, we must empower robots to interact freely in unstructured, cluttered spaces. Humans do this easily, using diverse, whole-arm, nonprehensile actions such as pushing or pulling in everyday tasks. These interaction strategies make difficult tasks easier and impossible tasks possible.
In this thesis, we aim to enable robots with similar capabilities. In particular, we formulate methods for planning robust open-loop trajectories that solve the rearrangement planning problem using nonprehensile interactions. In these problems, a robot must plan in a cluttered environment, reasoning about moving multiple objects in order to achieve a goal.
The problem is difficult because we must plan in continuous, high-dimensional state and action spaces. Additionally, during planning we must respect the physical constraints induced by the nonprehensile interaction between the robot and the objects in the scene.
Our key insight is that by embedding physics models directly into our planners we can naturally produce solutions that use nonprehensile interactions such as pushing. This also allows us to easily generate plans that exhibit full arm manipulation and simultaneous object interaction without the need for programmer defined high-level primitives that specifically encode this interaction. We show that by generating these diverse actions, we are able to find solutions for motion planning problems in highly cluttered, unstructured environments.
In the first part of this thesis we formulate the rearrangement planning problem as a classical motion planning problem. We show that we can embed physics simulators into randomized planners. We propose methods for reducing the search space and speeding planning time in order to make the planners useful in real-world scenarios.
The second part of the thesis tackles the imperfect and imprecise worlds that reflect the true reality for robots working in human environments. We pose the rearrangement planning under uncertainty problem as an instance of conformant probabilistic planning and offer methods for solving the problem. We demonstrate the effectiveness of our algorithms on two platforms: the home care robot HERB and the NASA rover K-Rex.
We demonstrate expanded autonomous capability on HERB, allowing him to work better in high clutter, completing previously infeasible tasks and speeding feasible task execution. In addition, we show these planners increase autonomy for the NASA rover K-Rex by allowing the rover to actively interact with the environment.
Siddhartha S. Srinivasa (Chair)
Matthew T. Mason
David Hsu (National University of Singapore)
Terrence W. Fong (NASA Ames Research Center)
Data driven approaches to modeling time-series are important in a variety of applications from market prediction in economics to the simulation of robotic systems. However, traditional supervised machine learning techniques designed for i.i.d. data often perform poorly on these sequential problems. This thesis proposes that time series and sequential prediction, whether for forecasting, filtering, or reinforcement learning, can be effectively achieved by directly training recurrent prediction procedures rather then building generative probabilistic models.
To this end, we introduce a new training algorithm for learned time-series models, Data as Demonstrator (DaD), that theoretically and empirically improves multi-step prediction performance on model classes such as recurrent neural networks, kernel regressors, and random forests. Additionally, experimental results indicate that DaD can accelerate model-based reinforcement learning. We next show that latent-state time-series models, where a sufficient state parametrization may be unknown, can be learned effectively in a supervised way. Our approach, Predictive State Inference Machines (PSIMs), directly optimizes – through a DaD-style training procedure – the inference performance without local optima by identifying the recurrent hidden state as a predictive belief over statistics of future observations. Fundamental to our learning framework is that the prediction of observable quantities is a lingua franca for building AI systems. We propose three extensions that leverage this general idea and adapt it to a variety of problems. The first aims to improve the training time and performance of more sophisticated recurrent neural networks. The second extends the PSIM framework to controlled dynamical systems. The third looks to train recurrent architectures for reinforcement learning problems.
J. Andrew Bagnell (Co-chair)
Martial Hebert (Co-chair)
Byron Boots (Georgia Institute of Technology)
It is a paradox that often the more severe a person's motor impairment, the more challenging it is for them to operate the very assistive machines which might enhance their quality of life. A primary aim of my lab is to address this confound by incorporating robotics autonomy and intelligence into assistive machines---to offload some of the control burden from the user. Robots already synthetically sense, act in and reason about the world, and these technologies can be leveraged to help bridge the gap left by sensory, motor or cognitive impairments in the users of assistive machines. However, here the human-robot team is a very particular one: the robot is physically supporting or attached to the human, replacing or enhancing lost or diminished function. In this case getting the allocation of control between the human and robot right is absolutely essential, and will be critical for the adoption of physically assistive robots within larger society. This talk will overview some of the ongoing projects and studies in my lab, whose research lies at the intersection of artificial intelligence, rehabilitation robotics and machine learning. We are working with a range of hardware platforms, including smart wheelchairs and assistive robotic arms. A distinguishing theme present within many of our projects is that the machine automation is customizable---to a user's unique and changing physical abilities, personal preferences or even financial means.
Brenna Argall is the June and Donald Brewer Junior Professor of Electrical Engineering & Computer Science at Northwestern University, and also an assistant professor in the Department of Mechanical Engineering and the Department of Physical Medicine & Rehabilitation. Her research lies at the intersection of robotics, machine learning and human rehabilitation. She is director of the assistive & rehabilitation robotics laboratory (argallab) at the Rehabilitation Institute of Chicago (RIC), the premier rehabilitation hospital in the United States, and her lab's mission is to advance human ability through robotics autonomy. Argall is a 2016 recipient of the NSF CAREER award. Her Ph.D. in Robotics (2009) was received from the Robotics Institute at Carnegie Mellon University, as well as her M.S. in Robotics (2006) and B.S. in Mathematics (2002). Prior to joining Northwestern, she was a postdoctoral fellow (2009-2011) at the École Polytechnique Fédérale de Lausanne (EPFL), and prior to graduate school she held a Computational Biology position at the National Institutes of Health (NIH).
Faculty Host: Stephen Nuske
Achieving optimality while staying safe is one of the key problems that arise when planning under uncertainty. We specifically focus on path planning for aerial vehicles, where the uncertainties arise due to unobserved winds and other air traffic. A flight plan or a policy that doesn’t take into account such uncertainties can not only result in highly inefficient flight paths but can also jeopardize safety. In this talk, we will first focus on how to reduce uncertainty in wind predictions by using airplanes in flight as a large-scale sensor network. In particular, we explore how information from existing commercial aircraft on their normal business can be harnessed to observe and predict weather phenomena at a continental scale in greater detail that currently available. In the second part of the talk, we consider the problem of path planning under uncertain winds and traffic conditions. Specifically we propose planning algorithms that trade off exploration and exploitation in near-optimal manner and have appealing no-regret properties. Further, we will also discuss how Probabilistic Signal Temporal Logic (PrSTL) can be adapted to the robotic path planning problems in order to guarantee safety. We will present results from longitudinal real-world studies that demonstrate effectiveness of the framework.
Ashish Kapoor is a senior researcher at Microsoft Research, Redmond. Currently, his research focuses on Aerial Informatics and Robotics with an emphasis on building intelligent and autonomous flying agents that are safe and enable applications that can positively influence our society. The research builds upon cutting edge research in machine intelligence, robotics and human-centered computation in order to enable an entire fleet of flying robots that range from micro-UAVs to commercial jetliners. Various applications scenarios include Weather Sensing, Monitoring for Precision Agriculture, Safe Cyber-Physical Systems etc. Ashish received his PhD from MIT Media Laboratory in 2006. He also holds FAA Commercial Pilot certificate (SEL), FAA Flight Instructor certificate (Airplane Single Engine and Instrument Airplane) and is an avid amateur aircraft builder (see build blog).
Faculty Host: Louis-Philippe Morency
Reception follows at 5:00 pm in Newell-Simon 1513
Learn more about the major and minor in Robotics!
We describe the development and testing of the Optical Coherence Tomography Microsurgical Augmented Reality System (OCT-MARS). This system allows surgeons to view real-time medical image data as an in-situ overlay within the surgical field. There are a number of clinical applications for which real time, in situ visualization of otherwise transparent structures of the eye would be beneficial to surgeons. The primary motivating application for this project is the surgical treatment of glaucoma. We have built a projection system capable of producing flat and tilted images in the normal field of view of the microscope with sufficient brightness and resolution to be viewed under magnification. We have also studied the perception of tilted surfaces under magnification and found that OCT images provide sufficient stereo information to be correctly perceived. Finally, we have tested stereo perception under magnification using surgically relevant tasks to evaluate the effectiveness of the system.
George Stetten (Co-chair)
John Galeotti (Co-chair)
Thomas Furness (University of Washington)
Humans effortlessly manipulate objects in cluttered and uncertain environments. In contrast, most robotic manipulators are limited to carefully engineered environments to circumvent the difficulty of manipulation under uncertainty. Contact sensors can provide robots with with the feedback vital to addressing this limitation.
This thesis proposes a framework for using feedback from contact sensors to reliably manipulate objects under uncertainty. We formalize manipulation as a partially observable Markov decision process that includes object pose uncertainty, proprioceptual error, and kinematic constraints. Our algorithms exploit the structure of contact to efficiently estimate state and plan with this model.
First, we introduce the manifold particle filter as a principled method of estimating object pose and robot configuration. This algorithm avoids degeneracy by drawing samples from the lower-dimensional manifold of states induced by contact. Next, we introduce two belief space planning algorithms that seek out contact with sensors when doing so is necessary to achieve the goal. One algorithm harnesses the decoupling effect of contact to share computation between problem instances. The second leverages lower-dimensional structure to plan around kinematic constraints.
Finally, we evaluate the efficacy of our approach in real-robot and simulation experiments. The results show that our state estimation and planning algorithms consistently outperform those that are not tailored to manipulation or contact sensing.
Siddhartha Srinivasa (Co-chair)
Nancy Pollard (Co-chair)
Tomas Lozano-Perez (Massachusetts Institute of Technology)
Improving robotic manipulation is critical for robots to be actively useful in real-world factories and homes. While some success has been shown in simulation and controlled environments, robots are slow, clumsy, and not general or robust enough when interacting with their environment. By contrast, humans effortlessly manipulate objects. One possible reason for this discrepancy is that humans have had years of experience to collect data to have good internal models of what happens when they manipulate objects. If robots could learn models from a large amount of real data, they could become more capable manipulators. In this thesis, we propose to improve robotic manipulation by solving two problems. First, we look at how robots can collect a large amount of manipulation data without human intervention. Second, we study how to build statistical models of robotic manipulation from the collected data. These data-driven models can then be used for planning more robust manipulation actions.
To solve the first problem of enabling large data collection, we perform several different robotic manipulation experiments and use these as case studies. We study bin-picking, post-grasp manipulation, pushing, regrasping, and planar grasping. These case studies allow us to gain insights on how robots can collect a large amount of accurate data with minimal human intervention.
To solve the second problem of statistically modeling manipulation actions, we propose models for different parts of various manipulation actions. First, we look at how to model post-grasp manipulation actions by modeling the probability distribution of where an object ends up in a robot's hand, and how this affects its success rate at various tasks such as placing or insertion. Second, we model how robots can change the pose of an object in their hand with regrasp actions. These learned data-driven models can then be used for planning more robust and accurate manipulation actions.
Matthew T. Mason (Chair)
Nancy S. Pollard
Geoffrey J. Gordon
Paul G. Backes (Jet Propulsion Laboratory)
In many application domains, robots co-exist in the same physical space with humans and aim to become trustworthy partners. We particularly envision personal robots arranging furniture with a human partner, manufacturing robots performing spar assembly with human co-workers, or rehabilitation robots assisting spinal cord injury patients. In such collaborative settings, humans often have inaccurate models of the robot capabilities, which leads the team towards suboptimal strategies. On the other hand, the robot frequently knows the optimal way of executing the task based on some objective performance metric. This thesis proposes a set of decision-theoretic models of human teammates, that allow the robot to reason in a principled way over the effects of its actions on the future human behavior, and guide the human towards new, optimal strategies, unknown to them in advance. We formalize human adaptability, that is their willingness to adapt to a robot strategy, and propose a human-robot mutual adaptation formalism based on a bounded-memory model. We evaluate the impact of adaptability on collaboration paradigms: a shared-location collaborative task, and a shared-autonomy setting. We show that the formalism significantly improves team performance when the starting human preference of executing the task is suboptimal. We expect that the proposed models will increase task performance, human trust in the robot and perceived collaboration on a variety of joint-action collaborative tasks.
Siddhartha Srinivasa (Chair)
David Hsu (National University of Singapore)
While the applications for robotics are plentiful in theory, matching technical capabilities to real world customer needs at high reliability and practical price points is incredibly difficult, leaving behind large numbers of ambitious, but ultimately failed, attempts to apply robotics to consumer applications. In this talk we will share a bit of our journey with Anki, a company we started working on in 2008 with the goal of identifying and entering markets where robotics and AI can have a real, measurable impact in a short time frame, and then using the technologies and learnings developed for one product as building blocks for the next.
We enjoyed an eventful path from our early days as three Robotics Institute PhD students working out of a Pittsburgh living room to a 150 person company (with over a dozen CMU RI grads!) with offices in San Francisco, London, Munich and Shenzhen. We will share a few of the stories and learnings along the journey through multiple product releases, four rounds of venture funding, challenges at the overlap of many disciplines, large scale mass production, and seemingly endless strings of highs and lows.
Finally, we are excited to share our next product, Cozmo, a robot character that uses a deep combination of robotics, AI, game design, and animated film-style animation with the aim of bringing a physical character to life with a level of personality, emotion and interaction that has never been possible outside of a screen. This interdisciplinary approach has led us to build a small animation studio within a robotics company with a novel approach to animating physical characters, showing intense levels of attachment and emotional response in all of our early testing. Along with a look at the many years of research and development leading to this product, we will discuss why the SDK that will be released with the launch in October could unlock one of the most capable and affordable robotic platforms for research and education.
Boris Sofman is co-founder and CEO of Anki, an artificial intelligence and robotics company focused on using these technologies to reinvent everyday consumer experiences. With an initial focus on entertainment, Anki's first product line, Overdrive, is a battle-racing game that allowed a level of physical gameplay and interaction previously not possible outside of video games and was one of the top selling toys of the 2015 holiday season. Anki is releasing its next product line, Cozmo, this fall. Boris has a background in building diverse robotic systems from consumer products to off-road autonomous vehicles and bomb-disposal robots. He earned a B.S., M.S. and Ph.D. from the Robotics Institute of Carnegie Mellon University.
Hanns Tappeiner is co-founder and President of Anki, an artificial intelligence and robotics company focused on creating groundbreaking consumer products. Anki's first product line, Overdrive, is a battle-racing game that allowed a level of physical gameplay and interaction previously not possible outside of video games and was one of the top selling toys of the 2015 holiday season. Anki is releasing its next product line, Cozmo, this fall. Before moving to the US for his MS and PhD in Robotics at Carnegie Mellon, Hanns earned a Dipl. Ing. in Computer Science in Europe with minors in Mechanical and Electrical Engineering. His is currently on LOA from the PhD program at CMU and hopes to find the time to finish his thesis in the not to far future. He is mainly interested in the application of Robotics and AI in real world consumer products.
Faculty Host: Martial Hebert