Abstract
Probabilistic belief change is an operation that takes a probability distribution representing a belief state along with an input sentence representing some information to be accommodated or removed, and maps it to a new probability distribution. In order to choose from many such mappings possible, techniques from information theory such as the principle of minimum cross-entropy have previously been used. Central to this principle is the Kullback-Leibler (KL) divergence. In this short study, we focus on the contraction of a belief state P by a belief a, which is the process of turning the belief a into a non-belief. The contracted belief state \(P^-_a\) can be represented as a mixture of two states: the original belief state P, and the resultant state \(P^*_{\lnot a}\) of revising P by \(\lnot a\). Crucial to this mixture is the mixing factor \(\epsilon \) which determines the proportion of P and \(P^*_{\lnot a}\) that are to be used in this process. We show that once \(\epsilon \) is determined, the KL divergence of \(P^-_a\) from P is given by a function whose only argument is \(\epsilon \). We suggest that \(\epsilon \) is not only a mixing factor but also captures relevant aspects of P and \(P^*_{\lnot a}\) required for computing the KL divergence.
We thank the referees for valuable comments. We also acknowledge the help of the Australian Research Council (ARC) for partially funding this research via ARC grant DP150104133.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Alchourrón, C.E., Gärdenfors, P., Makinson, D.: On the logic of theory change: Partial meet contraction and revision functions. J. Symb. Log. 50(2), 510–530 (1985)
Chhogyal, K., Nayak, A.C., Zhuang, Z., Sattar, A.: Probabilistic belief contraction using argumentation. In: Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, IJCAI 2015, Buenos Aires, Argentina, pp. 2854–2860, July 25–31, 2015
Gärdenfors, P.: Knowledge in Flux. Modelling the Dymanics of Epistemic States. MIT Press (1988)
Jaynes, E.T.: Information theory and statistical mechanics. Physical Review 106(4), 620 (1957)
Kern-Isberner, G.: Linking iterated belief change operations to nonmonotonic reasoning. In: Principles of Knowledge Representation and Reasoning: Proceedings of the Eleventh International Conference, KR 2008, pp. 166–176 (2008)
Kullback, S., Leibler, R.A.: On information and sufficiency. The Annals of Mathematical Statistics, 79–86 (1951)
Potyka, N., Beierle, C., Kern-Isberner, G.: Changes of relational probabilistic belief states and their computation under optimum entropy semantics. In: Timm, I.J., Thimm, M. (eds.) KI 2013. LNCS, vol. 8077, pp. 176–187. Springer, Heidelberg (2013)
Ramachandran, R., Ramer, A., Nayak, A.C.: Probabilistic belief contraction. Minds and Machines 22(4), 325–351 (2012)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Chhogyal, K., Nayak, A., Sattar, A. (2015). On the KL Divergence of Probability Mixtures for Belief Contraction. In: Hölldobler, S., , Peñaloza, R., Rudolph, S. (eds) KI 2015: Advances in Artificial Intelligence. KI 2015. Lecture Notes in Computer Science(), vol 9324. Springer, Cham. https://doi.org/10.1007/978-3-319-24489-1_20
Download citation
DOI: https://doi.org/10.1007/978-3-319-24489-1_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-24488-4
Online ISBN: 978-3-319-24489-1
eBook Packages: Computer ScienceComputer Science (R0)