Toggle Main Menu Toggle Search

Open Access padlockePrints

Generative adversarial framework for depth filling via Wasserstein metric, cosine transform and domain transfer

Lookup NU author(s): Dr Amir Atapour AbarghoueiORCiD



This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND).


In this work, the issue of depth filling is addressed using a self-supervised feature learning model that predicts missing depth pixel values based on the context and structure of the scene. A fully-convolutional generative model is conditioned on the available depth information and full RGB colour information from the scene and trained in an adversarial fashion to complete scene depth. Since ground truth depth is not readily available, synthetic data is instead used with a separate model developed to predict where holes would appear in a sensed (non-synthetic) depth image based on the contents of the RGB image. The resulting synthetic data with realistic holes is utilized in training the depth filling model which makes joint use of a reconstruction loss which employs the Discrete Cosine Transform for more realistic outputs, an adversarial loss which measures the distribution distances via the Wasserstein metric and a bottleneck feature loss that aids in better contextual feature execration. Additionally, the model is adversarially adapted to perform well on naturally-obtained data with no available ground truth. Qualitative and quantitative evaluations demonstrate the efficacy of the approach compared to contemporary depth filling techniques. The strength of the feature learning capabilities of the resulting deep network model is also demonstrated by performing the task of monocular depth estimation using our pre-trained depth hole filling model as the initialization for subsequent transfer learning.

Publication metadata

Author(s): Atapour-Abarghouei A, Akcay S, Payen de La Garanderie G, Breckon TP

Publication type: Article

Publication status: Published

Journal: Pattern Recognition

Year: 2019

Volume: 91

Pages: 232-244

Print publication date: 23/04/2019

Online publication date: 04/03/2019

Acceptance date: 17/02/2019

Date deposited: 02/02/2021

ISSN (print): 0031-3203

Publisher: Elsevier


DOI: 10.1016/j.patcog.2019.02.010


Altmetrics provided by Altmetric