LESSWRONG
LW

1319
Nina Katz-Christy
0010
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No posts to display.
No wikitag contributions to display.
Deconfusing Landauer's Principle
Nina Katz-Christy9mo10

how much information (in bits) we gain by finding out the exact value of X given some prior beliefs described by p(x).

Just to clarify, this is the expected/average information, right? 

If we observe X = x for some fixed x, we get exactly -log_2 p(x) bits of information, and entropy is the average of this information gain taken over the distribution p(x). For example, suppose X ~ Bern(0.5). Then we have a 50% chance of getting a 0, and thus gaining -log2(0.5) = 1 bit of information, and a 50% chance of getting a 1, and thus gaining -log2(0.5) = 1 bits of information, meaning we will necessarily gain 1 bit of information upon finding the value of X. But, if X ~ Bern(0.25) instead, then finding out X is 1 gives -log2(0.25) = 2 bits of information and finding out X is 0 gives log2(0.75) bits of information. So, on average, we expect to gain (0.25)(-log2(0.25)) + (0.75)(-log2(0.75)) bits of information.

Is my understanding of this correct?

Reply