Presented By: Michigan Robotics
Uncertainty Propagation In Robot Perception
Robotics PhD Defense, Parker Ewen
Chair: Ram Vasudevan
November 15, 3pm
FRB Atrium & Zoom
Abstract
Autonomous construction of shelters on Mars, robotic assistants for life-saving surgeries, and home robots integrated into our lives, once viewed as science fiction, will be attainable in the near-future thanks to advancements in robotics. A connecting thread between all these tasks, and a source of numerous challenges, is the inherent uncertainty in perceptual data. This dissertation considers how to manage rather than ignore perceptual uncertainty, enabling the creation of safe, robust, and reliable systems that will allow robots to leave the lab and enter the world. This dissertation introduces four novel approaches for perceptual uncertainty quantification. The first is a means of computing geometric uncertainty for radiance field models. Second, we propose a framework for estimating semantic uncertainty using vision and, third, we build off this framework to enable robots to use multiple sensing modalities such as vision and touch to update this semantic uncertainty. Lastly, we propose a means of propagating uncertainty over Riemannian manifolds with guarantees on numerical accuracy and with applications for robotic state estimation. Combined, these methods demonstrate a comprehensive means of quantifying uncertainty for robotic perception.
November 15, 3pm
FRB Atrium & Zoom
Abstract
Autonomous construction of shelters on Mars, robotic assistants for life-saving surgeries, and home robots integrated into our lives, once viewed as science fiction, will be attainable in the near-future thanks to advancements in robotics. A connecting thread between all these tasks, and a source of numerous challenges, is the inherent uncertainty in perceptual data. This dissertation considers how to manage rather than ignore perceptual uncertainty, enabling the creation of safe, robust, and reliable systems that will allow robots to leave the lab and enter the world. This dissertation introduces four novel approaches for perceptual uncertainty quantification. The first is a means of computing geometric uncertainty for radiance field models. Second, we propose a framework for estimating semantic uncertainty using vision and, third, we build off this framework to enable robots to use multiple sensing modalities such as vision and touch to update this semantic uncertainty. Lastly, we propose a means of propagating uncertainty over Riemannian manifolds with guarantees on numerical accuracy and with applications for robotic state estimation. Combined, these methods demonstrate a comprehensive means of quantifying uncertainty for robotic perception.
Related Links
Explore Similar Events
-
Loading Similar Events...