孙亮
开通时间:..
最后更新时间:..
点击次数:
论文类型:期刊论文
发表时间:2018-01-01
发表刊物:IEEE ACCESS
收录刊物:SCIE
卷号:6
页面范围:43315-43326
ISSN号:2169-3536
关键字:Multi-view learning; generative model; unsupervised learning; generative adversarial networks
摘要:The problem of multi-view transformation is associated with transforming available source views of a given object into unknown target views. To solve this problem, a Mutual-Encoding InfoGenerative Adversarial Networks (MEIGANs)-based algorithm is proposed in this paper. A mutual-encoding representation learning network is proposed to obtain multi-view representations, i.e., it guarantees through encoders different views of the same object are mapped to the common representation, which carries enough information with respect to the object itself. An InfoGenerative Adversarial Networks-based transformation network is proposed to transform multi-views of the given object, which carries the representation information in the generative models and discriminative models, guaranteeing the synthetic transformed view matches the source view. The advantages of the MEIGAN are that it bypasses direct mappings among different views, and can solve the problem of missing views in training data and the problem of mapping between transformed views and source views. Finally, experiments on incomplete data to complete data restoration tasks on MNIST, CelebA, and multi-view angle transformation tasks on 3-D rendered chairs and multi-view clothing show the proposed algorithm yields satisfactory transformation results.