Image-to-image translation for enhanced feature matching, image retrieval and visual localization

Open access
Date
2019Type
- Conference Paper
Abstract
The performance of machine learning and deep learning algorithms for image analysis depends significantly on the quantity and quality of the training data. The generation of annotated training data is often costly, time-consuming and laborious. Data augmentation is a powerful option to overcome these drawbacks. Therefore, we augment training data by rendering images with arbitrary poses from 3D models to increase the quantity of training images. These training images usually show artifacts and are of limited use for advanced image analysis. Therefore, we propose to use image-to-image translation to transform images from a rendered domain to a captured domain. We show that translated images in the captured domain are of higher quality than the rendered images. Moreover, we demonstrate that image-to-image translation based on rendered 3D models enhances the performance of common computer vision tasks, namely feature matching, image retrieval and visual localization. The experimental results clearly show the enhancement on translated images over rendered images for all investigated tasks. In addition to this, we present the advantages utilizing translated images over exclusively captured images for visual localization. Show more
Permanent link
https://doi.org/10.3929/ethz-b-000391178Publication status
publishedExternal links
Journal / series
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information SciencesVolume
Pages / Article No.
Publisher
CopernicusEvent
Subject
Image-to-Image Translation; Convolutional Neural Networks; Generative Adversarial Networks; Data Augmentation; 3D Models; Feature Matching; Image Retrieval; Visual LocalizationOrganisational unit
03766 - Pollefeys, Marc / Pollefeys, Marc
More
Show all metadata