Authors: Jia Wu; Zhi-hua Cai
Addresses: School of Computer Science, China University of Geosciences, No. 388 Lumo Road, Wuhan, China ' School of Computer Science, China University of Geosciences, No. 388 Lumo Road, Wuhan, China
Abstract: Naïve Bayes (NB) is a probability-based classification model based on the conditional independence assumption. However, in many real-world applications, this assumption is often violated. Responding to this fact, superparent-one-dependence estimators (SPODEs) weaken the attribute independence assumption by using each attribute of the database as the superparent. Aggregating one-dependence estimators (AODEs), which estimates the corresponding parameters for every SPODE, has been proved to be one of the most efficient models due to its high accuracy among those improvements for NB classifier. This paper investigates a novel approach to ensemble the single SPODE based on the boosting strategy, Boosting for superparent-one-dependence estimators, simply, BODE. BODE first endows every instance a weight, and then find an optimal SPODE with highest accuracy in each iteration as a weak classifier. By doing so, BODE boosts all the selected weak classifiers to do the classification in the test processing. Experiments on UCI datasets demonstrate the algorithm performance.
Keywords: naive Bayes; superparent-one-dependence estimator; SPODE; aggregating one-dependence estimator; AODE; boosting; classification.
International Journal of Computing Science and Mathematics, 2013 Vol.4 No.3, pp.277 - 286
Received: 16 May 2013
Accepted: 27 Jun 2013
Published online: 10 May 2014 *