On combining graph-partitioning with non-parametric clustering for image segmentation
Computer Vision and Image Understanding
Salient Closed Boundary Extraction with Ratio Contour
IEEE Transactions on Pattern Analysis and Machine Intelligence
Graph Partitioning Active Contours (GPAC) for Image Segmentation
IEEE Transactions on Pattern Analysis and Machine Intelligence
An Approximate Distribution for the Normalized Cut
Journal of Mathematical Imaging and Vision
A Graph Clustering Algorithm Based on Minimum and Normalized Cut
ICCS '07 Proceedings of the 7th international conference on Computational Science, Part I: ICCS 2007
Fundamenta Informaticae
Hypergraph Cuts & Unsupervised Representation for Image Segmentation
Fundamenta Informaticae
Contour grouping by clustering with multi-feature similarity measure
SSPR&SPR'10 Proceedings of the 2010 joint IAPR international conference on Structural, syntactic, and statistical pattern recognition
On the complexity of isoperimetric problems on trees
Discrete Applied Mathematics
Weighted adaptive neighborhood hypergraph partitioning for image segmentation
ICAPR'05 Proceedings of the Third international conference on Pattern Recognition and Image Analysis - Volume Part II
Selective attention guided perceptual grouping model
ICNC'05 Proceedings of the First international conference on Advances in Natural Computation - Volume Part I
Fundamenta Informaticae
Hi-index | 0.14 |
In recent years, one of the effective engines for perceptual organization of low-level image features is based on the partitioning of a graph representation that captures Gestalt inspired local structures, such as similarity, proximity, continuity, parallelism, and perpendicularity, over the low-level image features. Mainly motivated by computational efficiency considerations, this graph partitioning process is usually implemented as a recursive bipartitioning process, where, at each step, the graph is broken into two parts based on a partitioning measure. We focus on three such measures, namely, the minimum, average, and normalized cuts. The minimum cut partition seeks to minimize the total link weights cut. The average cut measure is proportional to the total link weight cut, normalized by the sizes of the partitions. The normalized cut measure is normalized by the product of the total connectivity (valencies) of the nodes in each partition. We provide theoretical and empirical insight into the nature of the three partitioning measures in terms of the underlying image statistics. In particular, we consider for what kinds of image statistics would optimizing a measure, irrespective of the particular algorithm used, result in correct partitioning. Are the quality of the groups significantly different for each cut measure? Are there classes of images for which grouping by partitioning does not work well? Also, can the recursive bipartitioning strategy separate out groups corresponding to K objects from each other? In the analysis, we draw from probability theory and the rich body of work on stochastic ordering of random variables. Our major conclusion is that optimization of none of the three measures is guaranteed to result in the correct partitioning of K objects, in the strict stochastic order sense, for all image statistics.