标题:Scratch: A scalable discrete matrix factorization hashing for cross-modal retrieval
作者:Li, Chuan-Xiang ;Chen, Zhen-Duo ;Zhang, Peng-Fei ;Luo, Xin ;Nie, Liqiang ;Zhang, Wei ;Xu, Xin-Shun
通讯作者:Xu, XinShun
作者机构:[Li, Chuan-Xiang ;Chen, Zhen-Duo ;Zhang, Peng-Fei ;Luo, Xin ;Xu, Xin-Shun ] School of Software, Shandong University, Jinan; 250101, China;[Zhang, Wei 更多
会议名称:26th ACM Multimedia conference, MM 2018
会议日期:22 October 2018 through 26 October 2018
来源:MM 2018 - Proceedings of the 2018 ACM Multimedia Conference
出版年:2018
页码:1-9
DOI:10.1145/3240508.3240547
关键词:Cross-Modal Retrieval; Discrete Optimization; Hashing; Matrix Factorization
摘要:In recent years, many hashing methods have been proposed for the cross-modal retrieval task. However, there are still some issues that need to be further explored. For example, some of them relax the binary constraints to generate the hash codes, which may generate large quantization error. Although some discrete schemes have been proposed, most of them are time-consuming. In addition, most of the existing supervised hashing methods use an n × n similarity matrix during the optimization, making them unscalable. To address these issues, in this paper, we present a novel supervised cross-modal hashing method-Scalable disCRete mATrix faCtorization Hashing, SCRATCH for short. It leverages the collective matrix factorization on the kernelized features and the semantic embedding with labels to find a latent semantic space to preserve the intra- and inter-modality similarities. In addition, it incorporates the label matrix instead of the similarity matrix into the loss function. Based on the proposed loss function and the iterative optimization algorithm, it can learn the hash functions and binary codes simultaneously. Moreover, the binary codes can be generated discretely, reducing the quantization error generated by the relaxation scheme. Its time complexity is linear to the size of the dataset, making it scalable to large-scale datasets. Extensive experiments on three benchmark datasets, namely, Wiki, MIRFlickr-25K, and NUS-WIDE, have verified that our proposed SCRATCH model outperforms several state-of-the-art unsupervised and supervised hashing methods for cross-modal retrieval. © 2018 Copyright held by the owner/author(s). Publication rights licensed to ACM.
收录类别:EI;SCOPUS
资源类型:会议论文;期刊论文
原文链接:https://www.scopus.com/inward/record.uri?eid=2-s2.0-85058218591&doi=10.1145%2f3240508.3240547&partnerID=40&md5=e792c137fa8cec68e54191344cce09bc
TOP