Mutual information

cosmos 17th May 2019 at 1:35am
Information measures

In Information theory, the mutual information between Random variables XX, and YY is defined as:

I(X;Y)=x,yp(x,y)logp(x,y)p(x)p(y)=Elogp(X,Y)p(X)p(Y)I(X;Y) = \sum_{x,y} p(x,y) \log{\frac{p(x,y)}{p(x)p(y)}} = E \log{\frac{p(X,Y)}{p(X)p(Y)}}

where EE denotes expectation. The mutual information measures the amount of information we obtain about XX by knowing YY (see result below).

video

The mutual information between a random variable and itself is equal to its entropy

Some results (video):

I(X;Y)=H(X)H(XY)=H(Y)H(YX)I(X;Y) = H(X) - H(X|Y) = H(Y) - H(Y|X)

H(XY)H(X|Y) is the Conditional entropy and thus gives you the information about XX that YY doesn't give you.


estimating mutual info with neural nets: https://arxiv.org/abs/1801.04062