Toggle Main Menu Toggle Search

Open Access padlockePrints

Adversarial unseen visual feature synthesis for Zero-shot Learning

Lookup NU author(s): Dr Yang Long



This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND).


© 2018 Due to the extreme imbalance of training data between seen classes and unseen classes, most existing methods fail to achieve satisfactory results in the challenging task of Zero-shot Learning (ZSL). To avoid the need for labelled data of unseen classes, in this paper, we investigate how to synthesize visual features for ZSL problem. The key challenge is how to capture the realistic feature distribution of unseen classes without training samples. To this end, we propose a hybrid model consists of Random Attribute Selection (RAS) and conditional Generative Adversarial Network (cGAN). RAS aims to learn the realistic generation of attributes by their correlations in nature. To improve the discrimination for the large number of classes, we add a reconstruction loss in the generative network, which can solve the domain shift problem and significantly improve the classification accuracy. Extensive experiments on four benchmarks demonstrate that our method can outperform all the state-of-the-art methods. Qualitative results show that, compared to conventional generative models, our method can capture more realistic distribution and remarkably improve the variability of the synthesized data.

Publication metadata

Author(s): Zhang H, Long Y, Liu L, Shao L

Publication type: Article

Publication status: Published

Journal: Neurocomputing

Year: 2019

Volume: 329

Pages: 12-20

Print publication date: 15/02/2019

Online publication date: 24/10/2018

Acceptance date: 21/10/2018

Date deposited: 17/01/2019

ISSN (print): 0925-2312

ISSN (electronic): 1872-8286

Publisher: Elsevier BV


DOI: 10.1016/j.neucom.2018.10.043


Altmetrics provided by Altmetric


Funder referenceFunder name