# What is Mutual Information?

### Pablo Aznar

#### 31/03/2021

In the field of machine learning, when it comes to extracting relationships between variables, we often use Pearson correlation. The problem is that this measure only finds linear relationships, which can lead sometimes to a bad interpretation of the relation between two variables. Nevertheless, other statistics measure non-linear relationships, such as mutual information.

Therefore, in this post, we are going to explain mutual information, how it is calculated, and an example of its use.

## Mutual Information

The Mutual Information between two random variables measures non-linear relations between them. Besides, it indicates how much information can be obtained from a random variable by observing another random variable.

It is closely linked to the concept of entropy. This is because it can also be known as the reduction of uncertainty of a random variable if another is known. Therefore, a high mutual information value indicates a large reduction of uncertainty whereas a low value indicates a small reduction. If the mutual information is zero, that means that the two random variables are independent.

### But, how is mutual information calculated?

The following formula shows the calculation of the mutual information for two discrete random variables.

$$I(X;Y) = \sum_{y \in Y}\sum_{x \in X}p_{(X,Y)}(x,y) \cdot log\left(\frac{p_{(X,Y)}(x,y)}{p_{X}(x)p_{Y}(y)} \right )$$

Where $$p_{x}$$ and $$p_{y}$$ are the marginal probability density functions and $$p_{xy}$$ the joint probability density function.

Whereas to compute the mutual information for continuous random variables the summations have to be replaced by the integrals.

$$I(X;Y) = \int_{Y}\int_{X} p_{(X,Y)}(x,y) \cdot log\left(\frac{p_{(X,Y)}(x,y)}{p_{X}(x)p_{Y}(y))} \right ) dxdy$$

As explained before, it is related to entropy. This relation is shown in the following formula:

\begin{align*} I(X; Y) & = H(X) – H(X|Y) \\ & = H(Y) – H(Y|X) \\ & = H(X) + H(Y) – H(X,Y) \\ & = H(X,Y) – H(X|Y) – H(Y|X) \end{align*}

Entropy (H) measures the level of expected uncertainty in a random variable. Therefore, H(X) is approximately how much information can be learned of the random variable X by observing just one sample.

$$H(X) = -\sum_{x_{i}\in X}P(X=x_{i})\cdot log(P(X=x_{i}))$$

The joint entropy measures the uncertainty when considering together two random variables.

$$H(X,Y) = – \sum_{x_{i} \in X} \sum_{y_{j}\in Y} P(X=x_{i}, Y=x_{j})\cdot log(P(X=x_{i}, Y=x_{j}))$$

The conditional entropy measures how much uncertainty has the random variable X when the value of Y is known.

$$H(X|Y) = – \sum_{x,y} p(x,y)\cdot log p(x|y)$$

For better understanding, the relationship between entropy and mutual information has been depicted in the following Venn diagram, where the area shared by the two circles is the mutual information:

### Properties of Mutual Information

The main properties of the Mutual Information are the following:

• Non-negative: $$I(X; Y) \geq 0$$
• Symmetric: $$I(X; Y) = I(Y; X)$$
• $$I(X; Y) = 0 \leftrightarrow X, Y$$ independent, because in that case $$P(x, y) = P(x) \cdot P(y)$$

Since mutual information has only lower boundaries, sometimes it is difficult to interpret the obtained result. Looking at the equation that relates mutual information with entropy and the Venn diagram, we can see that it is possible to obtain the maximum value of the mutual information.

\begin{align*} I(X; Y) = H(X) + H(Y) – H(X,Y) \\ \\ I_{max}(X; Y) = min(H(X), H(Y)) \\ \\ 0 \leq I(X; Y) \leq min(H(X), H(Y)) \\ \\ I_{NORM}(X; Y) = \frac{I(X; Y)}{I_{max}(X; Y)} = \frac{I(X; Y)}{min(H(X), H(Y)) } \end{align*}

### So, what is the difference between Mutual Information and correlation?

The main difference is that correlation is a measure of linear dependence, whereas mutual information measures general dependence (including non-linear relations). Therefore, mutual information detects dependencies that do not only depend on the covariance. Thus, mutual information is zero when the two random variables are strictly independent.

## Uses

In the field of machine learning, one of its main uses is in decision trees. It is used for looking for the optimum split of the features to choose the nodes that compose the tree (also called information gain).

Another use is for feature selection. When having a big dataset with a big range of features, mutual information can help to select a subset of those features in order to discard the irrelevant ones.

In other fields, mutual information is also widely used. For example, in telecommunications, it is used to calculate the channel capacity.

### Example

Let’s see an example of how mutual information can be used for feature selection. For that purpose, we are going to generate a synthetic dataset and then, calculate the mutual information between the features and the target. Then, the features with higher scores will be the selected ones.

Using the make_classification function of the python library scikit-learn, we generate a synthetic dataset, which is a binary classification problem. This function allows us to choose the number of informative, repeated, redundant, and random features.

Once generated the dataset with 6 informative features, 1 redundant, 2 repeated, and 1 random, we calculate the mutual information between each feature and the target. Using the mutual_info_classif also of scikit-learn, we obtain the following results:

As can be seen, features 7 and 8 have the same mutual information as 2 and 5, respectively, so they are the repeated ones. Furthermore, features 9 and 10 have a mutual information value of 0, indicating that they are both independent of the target, so they are random features. Moreover, the rest of the features are the informative ones and the redundant ones. Note that only by calculating the mutual information between the features and the target, we can no distinguish between informative and redundant features.

## Conclusions

In this post, we have seen what is mutual information, how it is calculated, its differences with correlation, and an example of how to use it for feature selection.

## References

 https://en.wikipedia.org/wiki/Mutual_information
 Li, W. (1990). Mutual information functions versus correlation functions. Journal of statistical physics60(5), 823-837.
 Kraskov, A., Stögbauer, H., & Grassberger, P. (2004). Estimating mutual information. Physical review E69(6), 066138.