A Simple Discriminative Dual Semantic Auto-Encoder for Zero-Shot Classification

Yang Liu, Jin Li, Xinbo Gao; Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2020, pp. 940-941

Abstract


Most existing ZSL models focus on searching the mapping between visual space and semantic space directly. However, few models study whether the human-designed semantic information is discriminative enough to recognize different categories. On the other hand, one-way mapping typically suffers from the project domain shift problem. Inspired by the encoder-decoder paradigm, we propose a novel solution to ZSL based on learning a Discriminative Dual Semantic Auto-encoder (DDSA). DDSA aims to build an aligned space to bridge the visual space and the semantic space by learning two bidirectional mappings, which provides us the required discriminative information about the visual and semantic features in the aligned space. The key to the proposed model is that we implicitly exact the principal information from visual and semantic space to construct aligned features, which is not only semantic-preserving but also discriminative. Extensive experiments on five benchmark data sets demonstrate the effectiveness of the proposed approach.

Related Material


[pdf]
[bibtex]
@InProceedings{Liu_2020_CVPR_Workshops,
author = {Liu, Yang and Li, Jin and Gao, Xinbo},
title = {A Simple Discriminative Dual Semantic Auto-Encoder for Zero-Shot Classification},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month = {June},
year = {2020}
}