Browse by author
Lookup NU author(s): Dr Yang Long
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (CC BY-NC-ND).
© 2018 Due to the extreme imbalance of training data between seen classes and unseen classes, most existing methods fail to achieve satisfactory results in the challenging task of Zero-shot Learning (ZSL). To avoid the need for labelled data of unseen classes, in this paper, we investigate how to synthesize visual features for ZSL problem. The key challenge is how to capture the realistic feature distribution of unseen classes without training samples. To this end, we propose a hybrid model consists of Random Attribute Selection (RAS) and conditional Generative Adversarial Network (cGAN). RAS aims to learn the realistic generation of attributes by their correlations in nature. To improve the discrimination for the large number of classes, we add a reconstruction loss in the generative network, which can solve the domain shift problem and significantly improve the classification accuracy. Extensive experiments on four benchmarks demonstrate that our method can outperform all the state-of-the-art methods. Qualitative results show that, compared to conventional generative models, our method can capture more realistic distribution and remarkably improve the variability of the synthesized data.
Author(s): Zhang H, Long Y, Liu L, Shao L
Publication type: Article
Publication status: Published
Print publication date: 15/02/2019
Online publication date: 24/10/2018
Acceptance date: 21/10/2018
Date deposited: 17/01/2019
ISSN (print): 0925-2312
ISSN (electronic): 1872-8286
Publisher: Elsevier BV
Altmetrics provided by Altmetric