Information Sciences: an International Journal
Analysis of an evolutionary RBFN design algorithm, CO2RBFN, for imbalanced data sets
Pattern Recognition Letters
IEEE Transactions on Evolutionary Computation
HAIS'11 Proceedings of the 6th international conference on Hybrid artificial intelligent systems - Volume Part I
IWANN'11 Proceedings of the 11th international conference on Artificial neural networks conference on Advances in computational intelligence - Volume Part I
Expert Systems with Applications: An International Journal
Pattern Recognition Letters
Class imbalance and the curse of minority hubs
Knowledge-Based Systems
Cost-sensitive decision tree ensembles for effective imbalanced classification
Applied Soft Computing
Hi-index | 0.00 |
A two-class data set is said to be imbalanced when one (minority) class is heavily under-represented with respect to the other (majority) class. In the presence of a significant overlapping, the task of learning from imbalanced data can be a very difficult problem. Additionally, if the overall imbalance ratio is different from local imbalance ratios in overlap regions, the task can become in a major challenge. This paper explains the behaviour of the k-nearest neighbour (k-NN) rule when learning from such a complex scenario. This local model is compared to other machine learning algorithms, attending to how their behaviour depends on a number of data complexity features (global imbalance, size of overlap region, and its local imbalance). As a result, several conclusions useful for classifier design are inferred.