Probabilistic reasoning in intelligent systems: networks of plausible inference
Probabilistic reasoning in intelligent systems: networks of plausible inference
Fundamental concepts of qualitative probabilistic networks
Artificial Intelligence
Adaptive Probabilistic Networks with Hidden Variables
Machine Learning - Special issue on learning with probabilistic representations
A tutorial on learning with Bayesian networks
Learning in graphical models
Creating an empirical basis for adaptation decisions
Proceedings of the 5th international conference on Intelligent user interfaces
Expert Systems and Probabiistic Network Models
Expert Systems and Probabiistic Network Models
Local learning in probabilistic networks with hidden variables
IJCAI'95 Proceedings of the 14th international joint conference on Artificial intelligence - Volume 2
Accelerating EM: an empirical study
UAI'99 Proceedings of the Fifteenth conference on Uncertainty in artificial intelligence
Elicitation of probabilities for belief networks: combining qualitative and quantitative information
UAI'95 Proceedings of the Eleventh conference on Uncertainty in artificial intelligence
Update rules for parameter estimation in Bayesian networks
UAI'97 Proceedings of the Thirteenth conference on Uncertainty in artificial intelligence
Learning Bayesian nets that perform well
UAI'97 Proceedings of the Thirteenth conference on Uncertainty in artificial intelligence
Tutorial and selected approaches on parameter learning in bayesian network with incomplete data
ISNN'12 Proceedings of the 9th international conference on Advances in Neural Networks - Volume Part I
Hi-index | 0.01 |
Algorithms for learning the conditional probabilities of Bayesian networks with hidden variables typically operate within a high-dimensional search space and yield only locally optimal solutions. One way of limiting the search space and avoiding local optima is to impose qualitative constraints that are based on background knowledge concerning the domain. We present a method for integrating formal statements of qualitative constraints into two learning algorithms, APN and EM. In our experiments with synthetic data, this method yielded networks that satisfied the constraints almost perfectly. The accuracy of the learned networks was consistently superior to that of corresponding networks learned without constraints. The exploitation of qualitative constraints therefore appears to be a promising way to increase both the interpretability and the accuracy of learned Bayesian networks with known structure.