SPLBoost: An improved robust boosting algorithm based on self-paced learning

K Wang, Y Wang, Q Zhao, D Meng… - IEEE Transactions on …, 2019 - ieeexplore.ieee.org
K Wang, Y Wang, Q Zhao, D Meng, X Liao, Z Xu
IEEE Transactions on Cybernetics, 2019ieeexplore.ieee.org
It is known that boosting can be interpreted as an optimization technique to minimize an
underlying loss function. Specifically, the underlying loss being minimized by the traditional
AdaBoost is the exponential loss, which proves to be very sensitive to random noise/outliers.
Therefore, several boosting algorithms, eg, LogitBoost and SavageBoost, have been
proposed to improve the robustness of AdaBoost by replacing the exponential loss with
some designed robust loss functions. In this article, we present a new way to robustify …
It is known that boosting can be interpreted as an optimization technique to minimize an underlying loss function. Specifically, the underlying loss being minimized by the traditional AdaBoost is the exponential loss, which proves to be very sensitive to random noise/outliers. Therefore, several boosting algorithms, e.g., LogitBoost and SavageBoost, have been proposed to improve the robustness of AdaBoost by replacing the exponential loss with some designed robust loss functions. In this article, we present a new way to robustify AdaBoost, that is, incorporating the robust learning idea of self-paced learning (SPL) into the boosting framework. Specifically, we design a new robust boosting algorithm based on the SPL regime, that is, SPLBoost, which can be easily implemented by slightly modifying off-the-shelf boosting packages. Extensive experiments and a theoretical characterization are also carried out to illustrate the merits of the proposed SPLBoost.
ieeexplore.ieee.org