2008 MultiClassCostSensitiveBoosting

From GM-RKB
Jump to navigation Jump to search

Subject Headings:

Notes

Cited By

Quotes

Author Keywords

Abstract

We propose a family of novel cost-sensitive boosting methods for multi-class classification by applying the theory of gradient boosting to p-norm based cost functionals. We establish theoretical guarantees including proof of convergence and convergence rates for the proposed methods. Our theoretical treatment provides interpretations for some of the existing algorithms in terms of the proposed family, including a generalization of the costing algorithm, DSE and GBSE-t, and the Average Cost method. We also experimentally evaluate the performance of our new algorithms against existing methods of cost sensitive boosting, including AdaCost, CSB2, and AdaBoost. M2 with cost-sensitive weight initialization. We show that our proposed scheme generally achieves superior results in terms of cost minimization and, with the use of higher order p-norm loss in certain cases, consistently outperforms the comparison methods, thus establishing its empirical advantage.

References

,

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2008 MultiClassCostSensitiveBoostingNaoki Abe
Aurélie C. Lozano
Multi-class Cost-sensitive Boosting with P-norm Loss FunctionsKDD-2008 Proceedings10.1145/1401890.14019532008