A Lower Bound of Hash Codes' Performance

October 12, 2022 ยท Declared Dead ยท ๐Ÿ› Neural Information Processing Systems

๐Ÿ“œ CAUSE OF DEATH: Death by README
Repo has only a README

Repo contents: .gitignore, LICENSE, README.md

Authors Xiaosu Zhu, Jingkuan Song, Yu Lei, Lianli Gao, Heng Tao Shen arXiv ID 2210.05899 Category cs.CV: Computer Vision Citations 2 Venue Neural Information Processing Systems Repository https://github.com/VL-Group/LBHash โญ 5 Last Checked 1 month ago
Abstract
As a crucial approach for compact representation learning, hashing has achieved great success in effectiveness and efficiency. Numerous heuristic Hamming space metric learning objectives are designed to obtain high-quality hash codes. Nevertheless, a theoretical analysis of criteria for learning good hash codes remains largely unexploited. In this paper, we prove that inter-class distinctiveness and intra-class compactness among hash codes determine the lower bound of hash codes' performance. Promoting these two characteristics could lift the bound and improve hash learning. We then propose a surrogate model to fully exploit the above objective by estimating the posterior of hash codes and controlling it, which results in a low-bias optimization. Extensive experiments reveal the effectiveness of the proposed method. By testing on a series of hash-models, we obtain performance improvements among all of them, with an up to $26.5\%$ increase in mean Average Precision and an up to $20.5\%$ increase in accuracy. Our code is publicly available at https://github.com/VL-Group/LBHash.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

๐Ÿ“œ Similar Papers

In the same crypt โ€” Computer Vision

Died the same way โ€” ๐Ÿ“œ Death by README