![]() |
个人信息Personal Information
教授
博士生导师
硕士生导师
主要任职:teaching
性别:男
毕业院校:重庆大学
学位:博士
所在单位:软件学院、国际信息与软件学院
学科:软件工程. 计算机软件与理论
办公地点:开发区综合楼405
联系方式:Email: zkchen@dlut.edu.cn Moble:13478461921 微信:13478461921 QQ:1062258606
电子邮箱:zkchen@dlut.edu.cn
Deep Discrete Cross-Modal Hashing for Cross-Media Retrieval
点击次数:
论文类型:期刊论文
发表时间:2018-11-01
发表刊物:PATTERN RECOGNITION
收录刊物:SCIE
卷号:83
页面范围:64-77
ISSN号:0031-3203
关键字:Cross-modal retrieval; deep learning; discrete hashing; alternative optimization
摘要:Cross-modal hashing has drawn increasing research interests in multimedia retrieval due to the explosive growth of multimedia big data. It is such a challenging topic due to the heterogeneity gap and high storage cost. However, most of the previous methods based on conventional linear projections and relaxation scheme fail to capture the nonlinear relationship among samples and suffers from large quantization loss, which result in an unsatisfactory performance of cross-modal retrieval. To address these issues, this paper is dedicated to learning discrete nonlinear hash functions by deep learning. A novel framework of cross-modal deep neural networks is proposed to learn binary codes directly. We formulate the similarity preserving in the framework, and also bit-independent as well as binary constraints are imposed on the hash codes. Specifically, we consider intra-modality similarity preserving at each hidden layer of the networks. Inter-modality similarity preserving is formulated by the output of each individual network. By so doing, the cross correlation can be encoded into the network training (i.e. hash functions learning) by back propagation algorithm. The final objective is solved by alternative optimization in an iterative fashion. Experimental results on four datasets i.e. NUS-WIDE, MIR Flickr, Pascal VOC, and LabelMe demonstrate the effectiveness of the proposed method, which is significantly superior to state-of-the-art cross-modal hashing approaches. (C) 2018 Elsevier Ltd. All rights reserved.