Hidden Markov Model} Induction by Bayesian Model Merging
Advances in Neural Information Processing Systems 5, [NIPS Conference]
Model-Based Hierarchical Clustering
UAI '00 Proceedings of the 16th Conference on Uncertainty in Artificial Intelligence
A permutation-augmented sampler for DP mixture models
Proceedings of the 24th international conference on Machine learning
Hierarchical mixture models: a probabilistic analysis
Proceedings of the 13th ACM SIGKDD international conference on Knowledge discovery and data mining
Agglomerative independent variable group analysis
Neurocomputing
Unsupervised Text Learning Based on Context Mixture Model with Dirichlet Prior
Advanced Web and NetworkTechnologies, and Applications
A new multimedia information data mining method
Proceedings of the first ACM/SIGEVO Summit on Genetic and Evolutionary Computation
Bayesian clustering for email campaign detection
ICML '09 Proceedings of the 26th Annual International Conference on Machine Learning
Non-parametric Bayesian areal linguistics
NAACL '09 Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics
Efficient Bayesian task-level transfer learning
IJCAI'07 Proceedings of the 20th international joint conference on Artifical intelligence
The nested chinese restaurant process and bayesian nonparametric inference of topic hierarchies
Journal of the ACM (JACM)
Modeling and Visualizing Uncertainty in Gene Expression Clusters Using Dirichlet Process Mixtures
IEEE/ACM Transactions on Computational Biology and Bioinformatics (TCBB)
The Indian Buffet Process: An Introduction and Review
The Journal of Machine Learning Research
A robust approach to multi-feature based mesh segmentation using adaptive density estimation
CAIP'11 Proceedings of the 14th international conference on Computer analysis of images and patterns - Volume Part I
Hierarchical verb clustering using graph factorization
EMNLP '11 Proceedings of the Conference on Empirical Methods in Natural Language Processing
Document hierarchies from text and links
Proceedings of the 21st international conference on World Wide Web
On two-way Bayesian agglomerative clustering of gene expression data
Statistical Analysis and Data Mining
Document-topic hierarchies from document graphs
Proceedings of the 21st ACM international conference on Information and knowledge management
Modeling topic hierarchies with the recursive chinese restaurant process
Proceedings of the 21st ACM international conference on Information and knowledge management
Online video segmentation by bayesian split-merge clustering
ECCV'12 Proceedings of the 12th European conference on Computer Vision - Volume Part IV
Learning bi-clustered vector autoregressive models
ECML PKDD'12 Proceedings of the 2012 European conference on Machine Learning and Knowledge Discovery in Databases - Volume Part II
PHA: A fast potential-based hierarchical agglomerative clustering method
Pattern Recognition
Data Field for Hierarchical Clustering
International Journal of Data Warehousing and Mining
Clustering using principal component analysis applied to autonomy-disability of elderly people
Decision Support Systems
CoBaFi: collaborative bayesian filtering
Proceedings of the 23rd international conference on World wide web
Analysing microarray expression data through effective clustering
Information Sciences: an International Journal
Hi-index | 0.00 |
We present a novel algorithm for agglomerative hierarchical clustering based on evaluating marginal likelihoods of a probabilistic model. This algorithm has several advantages over traditional distance-based agglomerative clustering algorithms. (1) It defines a probabilistic model of the data which can be used to compute the predictive distribution of a test point and the probability of it belonging to any of the existing clusters in the tree. (2) It uses a model-based criterion to decide on merging clusters rather than an ad-hoc distance metric. (3) Bayesian hypothesis testing is used to decide which merges are advantageous and to output the recommended depth of the tree. (4) The algorithm can be interpreted as a novel fast bottom-up approximate inference method for a Dirichlet process (i.e. countably infinite) mixture model (DPM). It provides a new lower bound on the marginal likelihood of a DPM by summing over exponentially many clusterings of the data in polynomial time. We describe procedures for learning the model hyperpa-rameters, computing the predictive distribution, and extensions to the algorithm. Experimental results on synthetic and real-world data sets demonstrate useful properties of the algorithm.