Published October 31, 2019 | Version 1.0.0
Conference paper Open

360° Surface Regression with a Hyper-Sphere Loss

Description

Omnidirectional vision is becoming increasingly relevant as more efficient 360° image acquisition is now possible. However, the lack of annotated 360° datasets has hindered the application of deep learning techniques on spherical content. This is further exaggerated on tasks where ground truth acquisition is difficult, such as monocular surface estimation. While recent research approaches on the 2D domain overcome this challenge by relying on generating normals from depth cues using RGB-D sensors, this is very difficult to apply on the spherical domain. In this work, we address the unavailability of sufficient 360° ground truth normal data, by leveraging existing 3D datasets and remodelling them via rendering. We present a dataset of 360° images of indoor spaces with their corresponding ground truth surface normal, and train a deep convolutional neural network (CNN) on the task of monocular 360° surface estimation. We achieve this by minimizing a novel angular loss function defined on the hyper-sphere using simple quaternion algebra. We put an effort to appropriately compare with other state of the art methods trained on planar datasets and finally, present the practical applicability of our trained model on a spherical image re-lighting task using completely unseen data by qualitatively showing the promising generalization ability of our dataset and model.

Notes

Dataset available @ https://vcl3d.github.io/3D60/

Files

arxiv_version.pdf

Files (6.6 MB)

Name Size Download all
md5:09ddb331e4b62bdc614e96331ead0f32
6.6 MB Preview Download

Additional details

Related works

Has part
Dataset: 10.5281/zenodo.3407840 (DOI)
Dataset: 10.5281/zenodo.3408441 (DOI)
Dataset: 10.5281/zenodo.3407875 (DOI)

Funding

European Commission
Hyper360 – Enriching 360 media with 3D storytelling and personalisation elements 761934