The Journal of Machine Learning Research
ICML '06 Proceedings of the 23rd international conference on Machine learning
Topics over time: a non-Markov continuous-time model of topical trends
Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining
Representing documents with named entities for story link detection (SLD)
CIKM '06 Proceedings of the 15th ACM international conference on Information and knowledge management
Proceedings of the 13th ACM SIGKDD international conference on Knowledge discovery and data mining
The dynamic hierarchical Dirichlet process
Proceedings of the 25th international conference on Machine learning
Introduction to Information Retrieval
Introduction to Information Retrieval
Dynamic hyperparameter optimization for bayesian topical trend analysis
Proceedings of the 18th ACM conference on Information and knowledge management
On smoothing and inference for topic models
UAI '09 Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence
An n-gram topic model for time-stamped documents
ECIR'13 Proceedings of the 35th European conference on Advances in Information Retrieval
Hi-index | 0.00 |
In this paper, we propose a new method for topical trend analysis We model topical trends by per-topic Beta distributions as in Topics over Time (TOT), proposed as an extension of latent Dirichlet allocation (LDA) However, TOT is likely to overfit to timestamp data in extracting latent topics Therefore, we apply prior distributions to Beta distributions in TOT Since Beta distribution has no conjugate prior, we devise a trick, where we set one among the two parameters of each per-topic Beta distribution to one based on a Bernoulli trial and apply Gamma distribution as a conjugate prior Consequently, we can marginalize out the parameters of Beta distributions and thus treat timestamp data in a Bayesian fashion In the evaluation experiment, we compare our method with LDA and TOT in link detection task on TDT4 dataset We use word predictive probabilities as term weights and estimate document similarities by using those weights in a TFIDF-like scheme The results show that our method achieves a moderate fitting to timestamp data.