en
Conference proceedings
Open access
English

Sparse ternary codes for similarity search have higher coding gain than dense binary codes

Publication date2017
Abstract

This paper addresses the problem of Approximate Nearest Neighbor (ANN) search in pattern recognition where feature vectors in a database are encoded as compact codes in or- der to speed-up the similarity search in large-scale databases. Considering the ANN problem from an information-theoretic perspective, we interpret it as an encoding, which maps the original feature vectors to a less entropic sparse represen- tation while requiring them to be as informative as possi- ble. We then define the coding gain for ANN search using information-theoretic measures. We next show that the clas- sical approach to this problem, which consists of binarization of the projected vectors is sub-optimal. Instead, a properly designed ternary encoding achieves higher coding gains and lower complexity.

Keywords
  • Approximate Nearest Neighbor search
  • Content identification
  • Binary hashing
  • Coding gain
  • Sparse representation
Citation (ISO format)
FERDOWSI, Sohrab et al. Sparse ternary codes for similarity search have higher coding gain than dense binary codes. [s.l.] : [s.n.], 2017.
Main files (1)
Proceedings (Published version)
accessLevelPublic
Identifiers
  • PID : unige:94026
447views
171downloads

Technical informations

Creation05/05/2017 4:24:00 PM
First validation05/05/2017 4:24:00 PM
Update time03/15/2023 1:39:15 AM
Status update03/15/2023 1:39:14 AM
Last indexation10/19/2023 3:44:36 AM
All rights reserved by Archive ouverte UNIGE and the University of GenevaunigeBlack