Residual Vector Product Quantization for approximate nearest neighbor search

被引:9
作者
Niu, Lushuai [1 ]
Xu, Zhi [1 ]
Zhao, Longyang [1 ]
He, Daojing [2 ]
Ji, Jianqiu [1 ,3 ]
Yuan, Xiaoli [4 ]
Xue, Mian [4 ]
机构
[1] Guilin Univ Elect Technol, Sch Comp Informat & Secur, Guilin 541004, Peoples R China
[2] Harbin Inst Technol, Sch Comp Sci & Technol, Shenzhen 518055, Peoples R China
[3] Doodod Technol Co Ltd, Beijing, Peoples R China
[4] Hohai Univ, Nanjing 210024, Peoples R China
基金
中国国家自然科学基金;
关键词
Vector quantization; Residual structure; Index structure; Approximate nearest neighbor search;
D O I
10.1016/j.eswa.2023.120832
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Vector quantization is one of the most popular techniques for approximate nearest neighbor (ANN) search. Over the past decade, many vector quantization methods have been proposed for ANN search. However, these methods do not strike a satisfactory balance between accuracy and efficiency because of their defects in quantization structures. To overcome this problem, a quantization method, named as Residual Vector Product Quantization (RVPQ), is proposed in this study. Under this method, the data space is decomposed into several subspaces, and a residual structure consisting of several ordered residual codebooks is constructed for each subspace. Learned by using an effective joint training algorithm, the quantization structure of RVPQ is much better than other methods, and it greatly enhances the performance of ANN search. Except that, an efficient residual quantization encoding method H-Variable Beam Search is also proposed to achieve higher encoding efficiency with negligible loss of accuracy. Furthermore, Inverted Multi-Index based on RVPQ is also designed to effectively solve the ANN search for a very large-scale database. Experimental results and theoretical evaluations show that RVPQ outperforms the-state-of-the-art methods on retrieval accuracy while retaining a comparable computational complexity.
引用
收藏
页数:11
相关论文
共 40 条
[1]   Optimized residual vector quantization for efficient approximate nearest neighbor search [J].
Ai, Liefu ;
Yu, Junqing ;
Wu, Zebin ;
He, Yunfeng ;
Guan, Tao .
MULTIMEDIA SYSTEMS, 2017, 23 (02) :169-181
[2]   Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions [J].
Andoni, Alexandr ;
Indyk, Piotr .
COMMUNICATIONS OF THE ACM, 2008, 51 (01) :117-122
[3]   Efficient Indexing of Billion-Scale datasets of deep descriptors [J].
Babenko, Artem ;
Lempitsky, Victor .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2055-2063
[4]   Additive Quantization for Extreme Vector Compression [J].
Babenko, Artem ;
Lempitsky, Victor .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :931-938
[5]   The Inverted Multi-Index [J].
Babenko, Artem ;
Lempitsky, Victor .
2012 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2012, :3069-3076
[6]   Revisiting the Inverted Indices for Billion-Scale Approximate Nearest Neighbors [J].
Baranchuk, Dmitry ;
Babenko, Artem ;
Malkov, Yury .
COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 :209-224
[7]  
Beyer K, 1999, LECT NOTES COMPUT SC, V1540, P217
[8]   Searching in high-dimensional spaces -: Index structures for improving the performance of multimedia Databases [J].
Böhm, C ;
Berchtold, S ;
Keim, D .
ACM COMPUTING SURVEYS, 2001, 33 (03) :322-373
[9]   Tensor Quantization: High-Dimensional Data Compression [J].
Chang, Shih Yu ;
Wu, Hsiao-Chun .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) :5566-5580
[10]  
Charikar M.S., 2002, STOC, P380, DOI DOI 10.1145/509907.509965