Please use this identifier to cite or link to this item:
|Scopus||Web of Science®||Altmetric|
|Title:||Fast supervised hashing with decision trees for high-dimensional data|
Van Den Hengel, A.
|Citation:||Proceedings, 2014 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014, 24-27 June 2014, Columbus, Ohio, USA / pp.1963-1970|
|Series/Report no.:||IEEE Conference on Computer Vision and Pattern Recognition|
|Conference Name:||IEEE Conference on Computer Vision and Pattern Recognition (2014 : Columbus, Ohio)|
|Guosheng Lin, Chunhua Shen, Qinfeng Shi, Anton van den Hengel, David Suter|
|Abstract:||Supervised hashing aims to map the original features to compact binary codes that are able to preserve label based similarity in the Hamming space. Non-linear hash functions have demonstrated their advantage over linear ones due to their powerful generalization capability. In the literature, kernel functions are typically used to achieve non-linearity in hashing, which achieve encouraging retrieval performance at the price of slow evaluation and training time. Here we propose to use boosted decision trees for achieving non-linearity in hashing, which are fast to train and evaluate, hence more suitable for hashing with high dimensional data. In our approach, we first propose sub-modular formulations for the hashing binary code inference problem and an efficient GraphCut based block search method for solving large-scale inference. Then we learn hash functions by training boosted decision trees to fit the binary codes. Experiments demonstrate that our proposed method significantly outperforms most state-of-the-art methods in retrieval precision and training time. Especially for highdimensional data, our method is orders of magnitude faster than many methods in terms of training time.|
|Rights:||© 2014 IEEE|
|Appears in Collections:||Aurora harvest 4|
Computer Science publications
Files in This Item:
|RA_hdl_84278.pdf||Restricted Access||440.25 kB||Adobe PDF||View/Open|
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.