Mutual information is a measure of the inherent dependence expressed in the [[joint distribution]] of <math>X</math> and <math>Y</math> relative to the joint distribution of <math>X</math> and <math>Y</math> under the assumption of independence. Mutual information therefore measures dependence in the following sense: <math>\operatorname{I}(X;Y)=0</math> [[if and only if]] <math>X</math> and <math>Y</math> are independent random variables. This is easy to see in one direction: if <math>X</math> and <math>Y</math> are independent, then <math>p_{(X,Y)}(x,y)=p_X(x) \cdot p_Y(y)</math>, and therefore: | Mutual information is a measure of the inherent dependence expressed in the [[joint distribution]] of <math>X</math> and <math>Y</math> relative to the joint distribution of <math>X</math> and <math>Y</math> under the assumption of independence. Mutual information therefore measures dependence in the following sense: <math>\operatorname{I}(X;Y)=0</math> [[if and only if]] <math>X</math> and <math>Y</math> are independent random variables. This is easy to see in one direction: if <math>X</math> and <math>Y</math> are independent, then <math>p_{(X,Y)}(x,y)=p_X(x) \cdot p_Y(y)</math>, and therefore: |