
Open access
Date
2021-05-30Type
- Working Paper
ETH Bibliography
yes
Altmetrics
Abstract
Introducing semantically meaningful objects to visual Simultaneous Localization and Mapping (SLAM) has the potential to improve both the accuracy and reliability of pose estimates, especially in challenging scenarios with significant viewpoint and appearance changes. However, how semantic objects should be represented for an efficient inclusion in optimization-based SLAM frameworks is still an open question. Superquadrics (SQs) are an efficient and compact object representation, able to represent most common object types to a high degree, and typically retrieved from 3D point-cloud data. However, accurate 3D point-cloud data might not be available in all applications. Recent advancements in machine learning enabled robust object recognition and semantic mask measurements from camera images under many different appearance conditions. We propose a pipeline to leverage such semantic mask measurements to fit SQ parameters to multi-view camera observations using a multi-stage initialization and optimization procedure. We demonstrate the system's ability to retrieve randomly generated SQ parameters from multi-view mask observations in preliminary simulation experiments and evaluate different initialization stages and cost functions. Show more
Permanent link
https://doi.org/10.3929/ethz-b-000487527Publication status
publishedPublisher
ETH Zurich, Autonomous System LabOrganisational unit
03737 - Siegwart, Roland Y. / Siegwart, Roland Y.
02261 - Center for Sustainable Future Mobility / Center for Sustainable Future Mobility
More
Show all metadata
ETH Bibliography
yes
Altmetrics