Authors: Sotiris B. Kotsiantis
Addresses: Educational Software Development Laboratory, Department of Mathematics, University of Patras, University Campus, 26504, Rio, Patras, Greece
Abstract: Naive Bayes algorithm captures the assumption that every attribute is independent from the rest of the attributes, given the state of the class attribute. In this study, we attempted to increase the prediction accuracy of the simple Bayes model. Because the concept of combining classifiers is proposed as a new direction for the improvement of the performance of individual classifiers, we propose a technique of localised multiple simple Bayes models. The ensemble consists of multiple simple Bayes models constructed locally by pseudorandomly selecting subsets of components of the feature vector, that is, simple Bayes models constructed in randomly chosen subspaces. Finally, we performed a large-scale comparison with other attempts that have tried to improve the accuracy of the naive Bayes algorithm as well as other state-of-the-art algorithms and ensembles on 26 standard benchmark datasets and the proposed method gave better accuracy in most cases.
Keywords: naive Bayes classifiers; instance-based learner; classifier ensemble; random subspace.
International Journal of Data Mining, Modelling and Management, 2009 Vol.1 No.4, pp.375 - 392
Published online: 29 Oct 2009 *Full-text access for editors Access for subscribers Purchase this article Comment on this article