Learning The State Of The World Object-based World Modeling For Mobile-Manipulation Robots
Mobile-manipulation robots performing service tasks in human-centric indoor environments need to know about relevant aspects of their spatial surroundings. However, service robots rarely know the exact state of the world, unlike industrial robots in structured environments. Additionally, as the world is shared with humans, uncertainty in the complete state of the world is inevitable over time. Mobile-manipulation robots therefore need to continuously perform state estimation, using perceptual information to maintain a representation of the state, and its uncertainty, of task-relevant aspects of the world. Because indoor tasks frequently require interacting with objects, objects should be given critical emphasis in spatial representations for service robots. In my Ph.D. work, I propose a world model based on objects, their semantic attributes (task-relevant properties such as type and pose), and their geometric realizations in the physical world.
Objects are challenging to keep track of because there is significant uncertainty in their states. Object detection and recognition using robotic vision is still error-prone. Objects can also be inherently ambiguous because they have similar attributes. Besides detection noise, other agents may change the state of the world. Compounded over multitudes of objects and long temporal horizons, the above sources of uncertainty give rise to a challenging estimation problem. Fortunately, most objects do not change quickly, and sensing is relatively cheap, so we can leverage information from multiple diverse snapshots of similar world states. However, putting the information together introduces a data association problem, which I tackle with constrained Bayesian nonparametric models. By carefully aggregating information across different viewpoints, times, and sensors, I show that robots can reduce their uncertainty in the state of the world and maintain more accurate object-based world models.
Lawson L.S. Wong is a Ph.D. candidate at the Massachusetts Institute of Technology, working in the Learning and Intelligent Systems Group under the supervision of Leslie Pack Kaelbling and Tomás Lozano-Pérez. Previously, he received his B.S. (with Honors) and M.S. in Computer Science at Stanford University, both in 2009. His current research focuses on acquiring, representing, and estimating knowledge about the world that an autonomous robot may find useful. More broadly, Lawson is interested in, and follows many topics within, the fields of robotics, machine learning, and artificial intelligence. He was recently awarded a AAAI Robotics Student Fellowship and a Croucher Foundation Fellowship for Postdoctoral Research. He will begin his postdoctoral appointment at Brown University in 2016, working with Stefanie Tellex.