fr
Actes de conférence
Accès libre
Anglais

Sparse ternary codes for similarity search have higher coding gain than dense binary codes

Date de publication2017
Résumé

This paper addresses the problem of Approximate Nearest Neighbor (ANN) search in pattern recognition where feature vectors in a database are encoded as compact codes in or- der to speed-up the similarity search in large-scale databases. Considering the ANN problem from an information-theoretic perspective, we interpret it as an encoding, which maps the original feature vectors to a less entropic sparse represen- tation while requiring them to be as informative as possi- ble. We then define the coding gain for ANN search using information-theoretic measures. We next show that the clas- sical approach to this problem, which consists of binarization of the projected vectors is sub-optimal. Instead, a properly designed ternary encoding achieves higher coding gains and lower complexity.

Mots-clés
  • Approximate Nearest Neighbor search
  • Content identification
  • Binary hashing
  • Coding gain
  • Sparse representation
Citation (format ISO)
FERDOWSI, Sohrab et al. Sparse ternary codes for similarity search have higher coding gain than dense binary codes. [s.l.] : [s.n.], 2017.
Fichiers principaux (1)
Proceedings (Published version)
accessLevelPublic
Identifiants
  • PID : unige:94026
455vues
185téléchargements

Informations techniques

Création05/05/2017 16:24:00
Première validation05/05/2017 16:24:00
Heure de mise à jour15/03/2023 01:39:15
Changement de statut15/03/2023 01:39:14
Dernière indexation16/01/2024 23:53:34
All rights reserved by Archive ouverte UNIGE and the University of GenevaunigeBlack