Distributed Graph Hashing

被引:21
作者
Wang, Shengnan [1 ,2 ]
Li, Chunguang [1 ,2 ]
Shen, Hui-Liang [1 ,2 ]
机构
[1] Zhejiang Univ, Coll Informat Sci & Elect Engn, Hangzhou 310027, Zhejiang, Peoples R China
[2] Zhejiang Univ, Zhejiang Prov Key Lab Informat Proc Commun & Netw, Hangzhou 310027, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Alternating direction method of multipliers (ADMMs); distributed hashing; graph hashing; large-scale image retrieval; learning to hash (LH); ITERATIVE QUANTIZATION; PROCRUSTEAN APPROACH; BIG DATA; RETRIEVAL; SEARCH;
D O I
10.1109/TCYB.2018.2816791
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, hashing-based approximate nearest neighbors search has attracted considerable attention, especially in big data applications, due to its low computation cost and fast retrieval speed. In the literature, most of the existing hashing algorithms are centralized. However, in many large-scale applications, the data are often stored or collected in a distributed manner. In this situation, the centralized hashing methods are not suitable for learning hash functions. In this paper, we consider the distributed learning to hash problem. We propose a novel distributed graph hashing model for learning efficient hash functions based on the data distributed across multiple agents over network. The graph hashing model involves a graph matrix, which contains the similarity information in the original space. We show that the graph matrix in the proposed distributed hashing model can be decomposed into multiple local graph matrices, and each local graph matrix can be constructed by a specific agent independently, with moderate communication and computation cost. Then, the whole objective function of the distributed hashing model can be represented by the sum of local objective functions of multiple agents, and the hashing problem can be formulated as a nonconvex constrained distributed optimization problem. For tractability, we transform the nonconvex constrained distributed optimization problem into an equivalent bi-convex distributed optimization problem. Then we propose two algorithms based on the idea of alternating direction method of multipliers to solve this problem in a distributed manner. We show that the proposed two algorithms have moderate communication and computational complexities, and both of them are scalable. Experiments on benchmark datasets are given to demonstrate the effectiveness of the proposed methods.
引用
收藏
页码:1896 / 1908
页数:13
相关论文
共 55 条
[21]  
Kulis B., 2009, ADV NEURAL INFORM PR, P1042
[22]   Fast Similarity Search for Learned Metrics [J].
Kulis, Brian ;
Jain, Prateek ;
Grauman, Kristen .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (12) :2143-2157
[23]  
Leng C, 2015, PR MACH LEARN RES, V37, P1642
[24]   Large-scale information retrieval with latent semantic indexing [J].
Letsche, TA ;
Berry, MW .
INFORMATION SCIENCES, 1997, 100 (1-4) :105-137
[25]   Distributed Vector Quantization over Sensor Network [J].
Li, Chunguang ;
Luo, Yiliang .
INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2014,
[26]   Diffusion Information Theoretic Learning for Distributed Estimation Over Network [J].
Li, Chunguang ;
Shen, Pengcheng ;
Liu, Ying ;
Zhang, Zhaoyang .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2013, 61 (16) :4011-4024
[27]  
Li X. L., 2003, International Journal of Computers & Applications, V25, P136
[28]  
Li Z., 2009, P ECAI, P64
[29]   Fast Supervised Hashing with Decision Trees for High-Dimensional Data [J].
Lin, Guosheng ;
Shen, Chunhua ;
Shi, Qinfeng ;
van den Hengel, Anton ;
Suter, David .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :1971-1978
[30]  
Liu H, 2017, AAAI CONF ARTIF INTE, P2238