The Combination of Evidence in the Transferable Belief Model
IEEE Transactions on Pattern Analysis and Machine Intelligence
Artificial Intelligence
Machine Learning
Pattern Recognition with Fuzzy Objective Function Algorithms
Pattern Recognition with Fuzzy Objective Function Algorithms
Ensembling evidential k-nearest neighbor classifiers through multi-modal perturbation
Applied Soft Computing
ECM: An evidential version of the fuzzy c-means algorithm
Pattern Recognition
Estimation and decision fusion: A survey
Neurocomputing
Rule extraction from support vector machines: A review
Neurocomputing
A defect in Dempster-Shafer theory
UAI'94 Proceedings of the Tenth international conference on Uncertainty in artificial intelligence
Leveraging k-NN for generic classification boosting
Neurocomputing
Belief C-Means: An extension of Fuzzy C-Means algorithm in belief functions framework
Pattern Recognition Letters
An evidence-theoretic k-NN rule with parameter optimization
IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews
EVCLUS: evidential clustering of proximity data
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
Classification Using Belief Functions: Relationship Between Case-Based and Model-Based Approaches
IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics
A neural network classifier based on Dempster-Shafer theory
IEEE Transactions on Systems, Man, and Cybernetics, Part A: Systems and Humans
Measuring ambiguity in the evidence theory
IEEE Transactions on Systems, Man, and Cybernetics, Part A: Systems and Humans
Hi-index | 0.01 |
In this paper, we present a new belief cxK neighbor (BCKN) classifier based on evidence theory for data classification when the available attribute information appears insufficient to correctly classify objects in specific classes. In BCKN, the query object is classified according to its K nearest neighbors in each class, and cxK neighbors are involved in the BCKN approach (c being the number of classes). BCKN works with the credal classification introduced in the belief function framework. It allows to commit, with different masses of belief, an object not only to a specific class, but also to a set of classes (called meta-class), or eventually to the ignorant class characterizing the outlier. The objects that lie in the overlapping zone of different classes cannot be reasonably committed to a particular class, and that is why such objects will be assigned to the associated meta-class defined by the union of these different classes. Such an approach allows to reduce the misclassification errors at the price of the detriment of the overall classification precision, which is usually preferable in some applications. The objects too far from the others will be naturally considered as outliers. The credal classification is interesting to explore the imprecision of class, and it can also provide a deeper insight into the data structure. The results of several experiments are given and analyzed to illustrate the potential of this new BCKN approach.