作者
Ziming Zhang, Yuting Chen, Venkatesh Saligrama
发表日期
2016
研讨会论文
Proceedings of the IEEE conference on computer vision and pattern recognition
页码范围
1487-1495
简介
In this paper, we propose training very deep neural networks (DNNs) for supervised learning of hash codes. Existing methods in this context train relatively" shallow" networks limited by the issues arising in back propagation (eg vanishing gradients) as well as computational efficiency. We propose a novel and efficient training algorithm inspired by alternating direction method of multipliers (ADMM) that overcomes some of these limitations. Our method decomposes the training process into independent layer-wise local updates through auxiliary variables. Empirically we observe that our training algorithm always converges and its computational complexity is linearly proportional to the number of edges in the networks. Empirically we manage to train DNNs with 64 hidden layers and 1024 nodes per layer for supervised hashing in about 3 hours using a single GPU. Our proposed very deep supervised hashing (VDSH) method significantly outperforms the state-of-the-art on several benchmark datasets.
引用总数
2014201520162017201820192020202120222023202417252724231512113
学术搜索中的文章
Z Zhang, Y Chen, V Saligrama - Proceedings of the IEEE conference on computer …, 2016
Z Zhang, Y Chen, V Saligrama - arXiv preprint arXiv:1511.04524, 2015