Information theory MOC

Shannon information

The Shannon information or surprisal of a discrete random variable 𝑋 :πœ‰ →𝑀 is the random function info

𝐼𝑋:𝑀→ℝπ‘₯β†¦βˆ’log𝑏⁑𝑝𝑋(π‘₯)=βˆ’log𝑏⁑ℙ(𝑋=π‘₯)

where 𝑏 =2 corresponds to the unit Sh, 𝑏 =𝑒 corresponds to the unit nat, and 𝑏 =10 corresponds to the unit Hart. Up to choice of 𝑏, the Shannon information function is the unique function satisfying the following properties:

  1. If 𝑝𝑋(π‘₯) =1 then 𝐼𝑋(π‘₯) =0, i.e. a certain event is perfectly unsurprising.
  2. If 𝑝𝑋(π‘₯1) <𝑝𝑋(π‘₯2) then 𝐼𝑋(π‘₯1) >𝐼𝑋(π‘₯2), i.e. the more unlikely an event the more surprising.
  3. If two independent events are measured, the total Shannon information gained is the sum of the Shannon information of the individual events.


develop | en | SemBr