Learning and Introspection for Effective and Reliable Planning Under Uncertainty: Towards Household Robots Comfortable with Missing Knowledge
In the pursuit of more capable robots, much of our research in the Robotic Anticipatory Intelligence & Learning Group (
See also our full publication list.
So far, my work has focused on the tasks of navigation and exploration: problem settings in which humans have incredibly powerful heuristics yet robots have historically struggled to perform as well. More generally, I am interested in the ways that we can imbue a robot with the ability to predict what lies beyond what it cas see, so that it can make more informed decisions when planning when much of the environment is not known.
Best Paper Finalist at CoRL 2018; Best Oral Presentation at CoRL 2018.
We also have a body of work on multi-robot planning, which further develops our abstractions for long-horizon planning under uncertainty to support concurrent progress made by a centrally-coordinated team of heterogeneous robots. Uncertainty further complicates the challenge of multi-robot planning: good behavior involves deciding what each robot should do over long time horizons, which requires careful consideration of when and how each should reveal unseen space and make progress towards completing the team’s joint objective.
Building on our work in learning-informed planning for single robots, our work in this forcus develops multi-robot state and actions abstractions that integrate learning with model-based planning for effective team behaviors. Our work has so far targeted complex multi-object retrieval and interaction tasks, and affords both reliability and performance, outperforming competitive learned and non-learned baseline strategies.
Robots will be expected to complete tasks in environments that persist over time, which means that the robots actions to complete an immediate task may impact subsequent tasks the robot has not yet been assigned. Most planners myopically aim to quickly immediate task without regard to what the robot may need to do next, owing to both the lack of advance knowledge of subsequent tasks and the would-be computational challenges associated with considering them.
Instead, our anticipatory planning approach uses learning to estimate the expected future cost associated with a particular solution strategy, information we can use to find plans that both quickly complete the current objective and also preemptively pay down cost on possible future tasks. Experiments of ours have shown that our approach results in many forward-thinking behaviors difficult to achieve otherwise—organization, tidiness, preparation, and more—exciting opportunities for improving robots that will serve as long-lived assistive companions.
Particularly when acting in uncertain and unknown environments, it is not sufficient to simply deploy robots and expect them to perform well. Robust long-horizon performance requires that robots regularly self-evaluate and introspect and tune their behavior appropriately based on their experience.
This focus develops tools for deployment-time introspection, enabling robots to quickly and reliably select the best-performing of a set of policies despite pervasive uncertainty. Our work is designed to operate across task planning settings and is flexible to support a variety of strategies that make predictions about what the robot does not know—from expert-designed heuristics to learned and language-model-informed predictive models—affording robust and data efficient adaptation despite uncertainty.
Aspects of this work are supported by a $500k grant from the
If we are to welcome robots into our homes and trust them to make decisions on their own, they should perform well despite uncertainty and be able to clearly explain both what they are doing and why they are doing it in terms that both expert and non-expert users can understand. As learning (particularly deep learning) is increasingly used to inform decisions, it is both a practical and an ethical imperative that we ensure that data-driven systems are easy to inspect and audit both during development and after deployment. Yet designing an agent that both achieves state-of-the-art performance and can meaningfully explain its actions has so far proven out of reach for existing approaches to long-horizon planning under uncertainty. In our work, we develop new representations for planning that allow us to leverage machine learning to inform good decisions and have sufficient structure so that the agent’s planning process is easily explained.
In our Planning under Uncertainty work, high-level (topological) strategies for navigation meaningfully reduce the space of possible actions available to a robot, allowing use of heuristic priors or learning to enable computationally efficient, intelligent planning. The challenges in estimating structure with many existing techniques that aim to build a map of the environment from monocular vision in low texture or highly cluttered environments have precluded their use for topological planning in the past. In our research, we proposed a robust, sparse map representation that we built with monocular vision and that overcomes these shortcomings. Using a learned sensor, we estimated high-level structure of an environment from streaming images by detecting sparse vertices (e.g., boundaries of walls) and reasoning about the structure between them. We also estimate the known free space in our map, a necessary feature for planning through previously unknown environments. Our mapping technique can also be used with real data and is sufficient for planning and exploration in simulated multi-agent search and Learned Subgoal Planning applications.