Skip to Main content Skip to Navigation
Conference papers

An Unforeseen Equivalence Between Uncertainty and Entropy

Abstract : Uncertainty and entropy are related concepts, so we would expect there to be some overlap, but the equality that is shown in this paper is unexpected. In Beta models, interactions between agents are evidence used to construct Beta distributions. In models based on the Beta Model, such as Subjective Logic, uncertainty is defined to be inversely proportional to evidence. Entropy measures measure how much information is lacking in a distribution. Uncertainty was neither intended nor expected to be an entropy measure. We discover that a specific entropy measure we call EDRB coincides with uncertainty whenever uncertainty is defined. EDRB is the expected Kullback-Leibler divergence between two Bernouilli trials with parameters randomly selected from the distribution. EDRB allows us to apply the notion of uncertainty to other distributions that may occur in the context of Beta models.
Document type :
Conference papers
Complete list of metadata
Contributor : Hal Ifip <>
Submitted on : Friday, March 26, 2021 - 2:33:15 PM
Last modification on : Friday, March 26, 2021 - 2:39:15 PM


 Restricted access
To satisfy the distribution rights of the publisher, the document is embargoed until : 2022-01-01

Please log in to resquest access to the document


Distributed under a Creative Commons Attribution 4.0 International License



Tim Muller. An Unforeseen Equivalence Between Uncertainty and Entropy. 13th IFIP International Conference on Trust Management (IFIPTM), Jul 2019, Copenhagen, Denmark. pp.57-72, ⟨10.1007/978-3-030-33716-2_5⟩. ⟨hal-03182607⟩



Record views