Toggle Main Menu Toggle Search

Open Access padlockePrints

A general transductive regularizer for zero-shot learning

Lookup NU author(s): Dr Yang Long


Full text for this publication is not currently held within this repository. Alternative links are provided below where available.


© 2019. The copyright of this document resides with its authors.Zero Shot Learning (ZSL) has attracted much attention due to its ability to recognize objects of unseen classes, which is realized by transferring knowledge from seen classes through semantic embeddings. Since the seen classes and unseen classes usually have different distributions, conventional inductive ZSL often suffers from the domain shift problem. Transductive ZSL is a type of method for solving such a problem. However, the regularizers of conventional transductive methods are different from each other, and cannot be applied to other methods. In this paper, we propose a General Transductive Regularizer (GTR), which assigns each unlabeled sample to a fixed attribute by defining a Kullback-Leibler Divergence (KLD) objective. To this end, GTR can be easily applied to many compatible linear and deep inductive ZSL models. Extensive experiments on both linear and deep methods are conducted on four popular datasets, and the results show that GTR can significantly improve the performance comparing to its original inductive method, and also outperform some state-of-the-art methods, especially the extension on deep model.

Publication metadata

Author(s): Mao H, Zhang H, Long Y, Wang S, Yang L

Publication type: Conference Proceedings (inc. Abstract)

Publication status: Published

Conference Name: 30th British Machine Vision Conference 2019, BMVC 2019

Year of Conference: 2019

Acceptance date: 02/04/2016

Publisher: BMVA Press