Latent Structure Preserving Hashing

Liu, Li, Yu, Mengyang and Shao, Ling (2017) Latent Structure Preserving Hashing. International Journal of Computer Vision, 122 (3). pp. 439-457. ISSN 0920-5691

[img]
Preview
Text (Final published version)
Liu2017_Article_LatentStructurePreservingHashi.pdf - Published Version
Available under License Creative Commons Attribution 4.0.

Download (4MB) | Preview
[img]
Preview
Text (Advance online version)
art%3A10.1007%2Fs11263-016-0931-4.pdf - Published Version
Available under License Creative Commons Attribution 4.0.

Download (4MB) | Preview
Official URL: https://doi.org/10.1007/s11263-016-0931-4

Abstract

Aiming at efficient similarity search, hash functions are designed to embed high-dimensional feature descriptors to low-dimensional binary codes such that similar descriptors will lead to binary codes with a short distance in the Hamming space. It is critical to effectively maintain the intrinsic structure and preserve the original information of data in a hashing algorithm. In this paper, we propose a novel hashing algorithm called Latent Structure Preserving Hashing (LSPH), with the target of finding a well-structured low-dimensional data representation from the original high-dimensional data through a novel objective function based on Nonnegative Matrix Factorization (NMF) with their corresponding Kullback-Leibler divergence of data distribution as the regularization term. Via exploiting the joint probabilistic distribution of data, LSPH can automatically learn the latent information and successfully preserve the structure of high-dimensional data. To further achieve robust performance with complex and nonlinear data, in this paper, we also contribute a more generalized multi-layer LSPH (ML-LSPH) framework, in which hierarchical representations can be effectively learned by a multiplicative up-propagation algorithm. Once obtaining the latent representations, the hash functions can be easily acquired through multi-variable logistic regression. Experimental results on three large-scale retrieval datasets, i.e., SIFT 1M, GIST 1M and 500 K TinyImage, show that ML-LSPH can achieve better performance than the single-layer LSPH and both of them outperform existing hashing techniques on large-scale data.

Item Type: Article
Uncontrolled Keywords: Hashing, Nonnegative matrix factorization, Latent structure, Dimensionality reduction, Multi-layer extension
Subjects: G400 Computer Science
Department: Faculties > Engineering and Environment > Computer and Information Sciences
Depositing User: Ellen Cole
Date Deposited: 28 Jul 2016 14:40
Last Modified: 31 Jul 2021 13:35
URI: http://nrl.northumbria.ac.uk/id/eprint/27249

Actions (login required)

View Item View Item

Downloads

Downloads per month over past year

View more statistics