Please use this identifier to cite or link to this item:
|Scopus||Web of Science®||Altmetric|
|Title:||Training effective node classifiers for cascade classification|
Van Den Hengel, A.
|Citation:||International Journal of Computer Vision, 2013; 103(3):326-347|
|Publisher:||Kluwer Academic Publ|
|Chunhua Shen, Peng Wang, Sakrapee Paisitkriangkrai, Anton van den Hengel|
|Abstract:||Cascade classifiers are widely used in real-time object detection. Different from conventional classifiers that are designed for a low overall classification error rate, a classifier in each node of the cascade is required to achieve an extremely high detection rate and moderate false positive rate. Although there are a few reported methods addressing this requirement in the context of object detection, there is no principled feature selection method that explicitly takes into account this asymmetric node learning objective. We provide such an algorithm here. We show that a special case of the biased minimax probability machine has the same formulation as the linear asymmetric classifier (LAC) of Wu et al (2005). We then design a new boosting algorithm that directly optimizes the cost function of LAC. The resulting totally-corrective boosting algorithm is implemented by the column generation technique in convex optimization. Experimental results on object detection verify the effectiveness of the proposed boosting algorithm as a node classifier in cascade object detection, and show performance better than that of the current state-of-the-art.|
|Keywords:||AdaBoost; Minimax Probability Machine; Cascade Classifier; Object Detection; Human Detection|
|Description:||Extent: 23p. The final publication is available at www.springerlink.com: http://link.springer.com/article/10.1007/s11263-013-0608-1|
|Rights:||© Springer Science+Business Media New York 2013|
|Appears in Collections:||Computer Science publications|
Files in This Item:
|hdl_74723.pdf||Published version||809.48 kB||Adobe PDF||View/Open|
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.