Flexible product quantization for fast approximate nearest neighbor search

被引:0
作者
Jingya Fan
Yang Wang
Wenwen Song
Zhibin Pan
机构
[1] Xi’an Jiaotong University,Faculty of Electronic and Information Engineering
[2] Xi’an University of Technology,Department of Information Science
[3] Research Institute of Xi’an Jiaotong University,undefined
来源
Multimedia Tools and Applications | 2024年 / 83卷
关键词
Approximate nearest neighbor search; Product quantization; The number of sub-vectors; Sub-vector concatenation; Flexible product quantization;
D O I
暂无
中图分类号
学科分类号
摘要
Product quantization (PQ) is an effective solution to approximate nearest neighbor (ANN) search. The idea of PQ is to decompose the space into Cartesian product of several low-dimensional subspaces and quantize each subspace separately. Database vectors are represented by short codes with different length composed of their subspace quantization indices. A vector is encoded to a short code consisting of its subspace quantization indices. PQ allows to generate a large size codebook with very low memory and time cost, and the selection of the number M of sub-vectors is the most essential and important problem, but is easily neglected. Motivated by this observation, we propose a method named Flexible Product quantization (FPQ) to optimize PQ-based methods that fix M for all vectors. FPQ achieves flexible M by concatenating the sub-vector pairs through constrains on quantization distortion. The concatenating operation is implemented individually among different database vectors and can lead to far less additions in the search process. Different from existing PQ-based fast search methods, this operation of our proposed FPQ is to reduce the computational complexity for searching each single database vector, which provides a new perspective to better accelerate ANN search. Experimental results show that our method for product quantization based methods can significantly reduce nearly 25%\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\%$$\end{document} computational complexity by using appropriate parameters, while the search accuracy can be guaranteed.
引用
收藏
页码:53243 / 53261
页数:18
相关论文
共 52 条
  • [1] Zhang S(2017)Efficient knn classification with different numbers of nearest neighbors IEEE Trans Neural Netw Learn Syst 29 1774-1785
  • [2] Li X(2017)Compact hash codes for efficient visual descriptors retrieval in large scale databases IEEE Trans Multimed 19 2521-2532
  • [3] Zong M(2012)Semi-supervised hashing for large-scale search IEEE Trans Patt Anal Mach Intell 34 2393-2406
  • [4] Zhu X(2020)Deep semantic multimodal hashing network for scalable image-text and video-text retrievals IEEE Trans Neural Netw Learn Syst 34 1838-1851
  • [5] Wang R(2020)Weakly-supervised semantic guided hashing for social image retrieval Int J Comp Vision 128 2265-2278
  • [6] Ercoli S(2011)Product quantization for nearest neighbor search IEEE Trans Patt Anal Mach Intell 33 117-128
  • [7] Bertini M(2020)Product quantization with dual codebooks for approximate nearest neighbor search Neurocomputing 401 59-68
  • [8] Bimbo AD(2014)Optimized product quantization IEEE Trans Patt Anal Mach Intell 36 744-755
  • [9] Wang J(2020)Asymmetric mapping quantization for nearest neighbor search IEEE Trans Patt Anal Mach Intell 42 1783-1790
  • [10] Kumar S(2019)Distance encoded product quantization for approximate k-nearest neighbor search in high-dimensional space IEEE Trans Patt Anal Mach Intell 41 2084-2097