location: Current position: Home >> Scientific Research >> Paper Publications

Cooperative Coupled Generative Networks for Generalized Zero-Shot Learning

Hits:

Indexed by:Journal Papers

Date of Publication:2020-01-01

Journal:IEEE ACCESS

Included Journals:SCIE

Volume:8

Page Number:119287-119299

ISSN No.:2169-3536

Key Words:Visualization; Semantics; Generative adversarial networks; Neural networks; Correlation; Training; Task analysis; Zero-shot learning; generalized zero-shot learning; generative adversarial network; neural network; residual module

Abstract:Compared with zero-shot learning (ZSL), the generalized zero-shot learning (GZSL) is more challenging since its test samples are taken from both seen and unseen classes. Most previous mapping-based methods perform well on ZSL, while their performance degrades on GZSL. To solve this problem, inspired by the ensemble learning, this paper proposes a model with cooperative coupled generative networks (CCGN). Firstly, to alleviate the hubness problem, the reverse visual feature space is taken as the embedding space, with the mapping achieved by a visual feature center generation network. To learn a proper visual representation of each class, we propose a coupled of generative networks, which cooperate with each other to synthesize a visual feature center template of the class. Secondly, to improve the generative ability of the coupled networks, we further employ a deeper network to generate. Meanwhile, to alleviate loss semantic information problem caused by multiple network layers, a residual module is employed. Thirdly, to mitigate overfitting and to increase scalability, an adversarial network is introduced to discriminate the generation of visual feature centers. Finally, a reconstruction network, which reverses the generation process, is employed to restrict the structural correlation between the generated visual feature center and the original semantic representation of each class. Extensive experiments on five benchmark datasets (AWA1, AWA2, CUB, SUN, APY) demonstrate that the proposed algorithm yields satisfactory results, as compared with the state-of-the-art methods.

Pre One:Exploring Overall Contextual Information for Image Captioning in Human-Like Cognitive Style

Next One:Exploring Latent Information for Unsupervised Person Re-Identification by Discriminative Learning Networks