An efficient digital VLSI implementation of Gaussian mixture models-based classifier
IEEE Transactions on Very Large Scale Integration (VLSI) Systems
NB+: An improved Naïve Bayesian algorithm
Knowledge-Based Systems
Boosting performance of gene mention tagging system by hybrid methods
Journal of Biomedical Informatics
Parallel multi-objective Ant Programming for classification using GPUs
Journal of Parallel and Distributed Computing
Hi-index | 0.00 |
In this paper, we address the problem of how to classify a set of query vectors that belong to the same unknown class. Sets of data known to be sampled from the same class are naturally available in many application domains, such as speaker recognition. We refer to these sets as homologous sets. We show how to take advantage of homologous sets in classification to obtain improved accuracy over classifying each query vector individually. Our method, called homologous naive Bayes (HNB), is based on the naive Bayes classifier, a simple algorithm shown to be effective in many application domains. RNB uses a modified classification procedure that classifies multiple instances as a single unit. Compared with a voting method and several other variants of naive Bayes classification, HNB significantly outperforms these methods in a variety of test data sets, even when the number of query vectors in the homologous sets is small. We also report a successful application of HNB to speaker recognition. Experimental results show that HNB can achieve classification accuracy comparable to the Gaussian mixture model (GMM), the most widely used speaker recognition approach, while using less time for both training and classification