HAL will be down for maintenance from Friday, June 10 at 4pm through Monday, June 13 at 9am. More information
Skip to Main content Skip to Navigation
Conference papers

An Unforeseen Equivalence Between Uncertainty and Entropy

Abstract : Uncertainty and entropy are related concepts, so we would expect there to be some overlap, but the equality that is shown in this paper is unexpected. In Beta models, interactions between agents are evidence used to construct Beta distributions. In models based on the Beta Model, such as Subjective Logic, uncertainty is defined to be inversely proportional to evidence. Entropy measures measure how much information is lacking in a distribution. Uncertainty was neither intended nor expected to be an entropy measure. We discover that a specific entropy measure we call EDRB coincides with uncertainty whenever uncertainty is defined. EDRB is the expected Kullback-Leibler divergence between two Bernouilli trials with parameters randomly selected from the distribution. EDRB allows us to apply the notion of uncertainty to other distributions that may occur in the context of Beta models.
Document type :
Conference papers
Complete list of metadata

Contributor : Hal Ifip Connect in order to contact the contributor
Submitted on : Friday, March 26, 2021 - 2:33:15 PM
Last modification on : Friday, March 26, 2021 - 2:39:15 PM
Long-term archiving on: : Sunday, June 27, 2021 - 6:47:26 PM


Files produced by the author(s)


Distributed under a Creative Commons Attribution 4.0 International License



Tim Muller. An Unforeseen Equivalence Between Uncertainty and Entropy. 13th IFIP International Conference on Trust Management (IFIPTM), Jul 2019, Copenhagen, Denmark. pp.57-72, ⟨10.1007/978-3-030-33716-2_5⟩. ⟨hal-03182607⟩



Record views


Files downloads