TY - GEN
T1 - Distribution based ensemble for class imbalance learning
AU - Mustafa, Ghulam
AU - Niu, Zhendong
AU - Yousif, Abdallah
AU - Tarus, John
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2015/7/30
Y1 - 2015/7/30
N2 - MultiBoost ensemble has been well acknowledged as an effective learning algorithm which able to reduce both bias and variance in error and has high generalization performance. However, to deal with the class imbalanced learning, the Multi- Boost shall be amended. In this paper, a new hybrid machine learning method called Distribution based MultiBoost (DBMB) for class imbalanced problems is proposed, which combines Distribution based balanced sampling with the MultiBoost algo- rithm to achieve better minority class performance. It minimizes the within class and between class imbalance by learning and sampling different distributions (Gaussian and Poisson) and reduces bias and variance in error by employing the MultiBoost ensemble. Therefore, DBMB could output the final strong learner that is more proficient ensemble of weak base learners for imbalanced data sets. We prove that the G-mean, F1 measure and AUC of the DBMB is significantly superior to others. The experimental verification has shown that the proposed DBMB outperforms other state-of-the-art algorithms on many real world class imbalanced problems. Furthermore, our proposed method is scalable as compare to other boosting methods.
AB - MultiBoost ensemble has been well acknowledged as an effective learning algorithm which able to reduce both bias and variance in error and has high generalization performance. However, to deal with the class imbalanced learning, the Multi- Boost shall be amended. In this paper, a new hybrid machine learning method called Distribution based MultiBoost (DBMB) for class imbalanced problems is proposed, which combines Distribution based balanced sampling with the MultiBoost algo- rithm to achieve better minority class performance. It minimizes the within class and between class imbalance by learning and sampling different distributions (Gaussian and Poisson) and reduces bias and variance in error by employing the MultiBoost ensemble. Therefore, DBMB could output the final strong learner that is more proficient ensemble of weak base learners for imbalanced data sets. We prove that the G-mean, F1 measure and AUC of the DBMB is significantly superior to others. The experimental verification has shown that the proposed DBMB outperforms other state-of-the-art algorithms on many real world class imbalanced problems. Furthermore, our proposed method is scalable as compare to other boosting methods.
KW - Class imbalance learning
KW - MultiBoost
KW - distribution based resampling
KW - ensemble learning
UR - http://www.scopus.com/inward/record.url?scp=84946575132&partnerID=8YFLogxK
U2 - 10.1109/INTECH.2015.7173365
DO - 10.1109/INTECH.2015.7173365
M3 - Conference contribution
AN - SCOPUS:84946575132
T3 - 5th International Conference on Innovative Computing Technology, INTECH 2015
SP - 5
EP - 10
BT - 5th International Conference on Innovative Computing Technology, INTECH 2015
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 5th International Conference on Innovative Computing Technology, INTECH 2015
Y2 - 20 May 2015 through 22 May 2015
ER -