Browse by author
Lookup NU author(s): Dr Amir Atapour AbarghoueiORCiD
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND).
In this work, the issue of depth filling is addressed using a self-supervised feature learning model that predicts missing depth pixel values based on the context and structure of the scene. A fully-convolutional generative model is conditioned on the available depth information and full RGB colour information from the scene and trained in an adversarial fashion to complete scene depth. Since ground truth depth is not readily available, synthetic data is instead used with a separate model developed to predict where holes would appear in a sensed (non-synthetic) depth image based on the contents of the RGB image. The resulting synthetic data with realistic holes is utilized in training the depth filling model which makes joint use of a reconstruction loss which employs the Discrete Cosine Transform for more realistic outputs, an adversarial loss which measures the distribution distances via the Wasserstein metric and a bottleneck feature loss that aids in better contextual feature execration. Additionally, the model is adversarially adapted to perform well on naturally-obtained data with no available ground truth. Qualitative and quantitative evaluations demonstrate the efficacy of the approach compared to contemporary depth filling techniques. The strength of the feature learning capabilities of the resulting deep network model is also demonstrated by performing the task of monocular depth estimation using our pre-trained depth hole filling model as the initialization for subsequent transfer learning.
Author(s): Atapour-Abarghouei A, Akcay S, Payen de La Garanderie G, Breckon TP
Publication type: Article
Publication status: Published
Journal: Pattern Recognition
Year: 2019
Volume: 91
Pages: 232-244
Print publication date: 23/04/2019
Online publication date: 04/03/2019
Acceptance date: 17/02/2019
Date deposited: 02/02/2021
ISSN (print): 0031-3203
Publisher: Elsevier
URL: https://doi.org/10.1016/j.patcog.2019.02.010
DOI: 10.1016/j.patcog.2019.02.010
Altmetrics provided by Altmetric