Multivariate Mutual Information based Feature Selection for Predicting Histone Post-Translational Modifications in Epigenetic Datasets

This article has been Reviewed by the following groups

Read the full article See related articles

Listed in

Log in to save this article

Abstract

Mutual information (MI) has been traditionally employed in many areas including biology to identify the non-linear relationships between features. This technique is particularly useful in the biological context to identify features such as genes, histone post-translational modifications (PTMs), transcriptional factors etc . In this work, instead of considering the conventional pairwise MI between PTM features, we evaluate multivariate mutual information (MMI) between PTM triplets, to identify a set of outlier features. This enables us to form a small subset of PTMs that serve as principal features for the prediction of the values of any histone PTM across the epigenome. We also compare the principal MMI features with those from the traditional feature selection techniques such as PCA and Orthogonal Matching Pursuit. We predict all the remaining histone PTM intensities using XGBoost based regression on the selected features. The accuracy of this technique is demonstrated on the ChIP-seq datasets from the yeast and the human epigenomes.The results indicate that the proposed MMI based feature selection technique can serve as a useful method across various biological datasets.

Article activity feed

  1. significant negative MMI for the feature set seems to be a reasonable idea to build an independent featureset.

    One concern about only looking at negative MMI is that you will miss other types of interactions. For example, high positive MMI (or interaction information) is associated with 'common cause' effects and represent another important class of interactions. Did you consider including both cases or including signed representations?

  2. 𝐼 (𝑋 ; 𝑌 ; 𝑍 ) = 𝐻 (𝑋 ) + 𝐻 (𝑌 ) + 𝐻 (𝑍 ) − 𝐻 (𝑋, 𝑌 ) + 𝐻 (𝑋, 𝑍 ) + 𝐻 (𝑌, 𝑍 ) + 𝐻 (𝑋, 𝑌, 𝑍 ) (2)Eq. 3 can be rewritten in terms of MI as𝐼 (𝑋 ; 𝑌 ; 𝑍 ) = 𝐼 (𝑋 ; 𝑌 ) − 𝐼 (𝑋 ; 𝑌 |𝑍 )

    This generalization of MI to more than two variables is also known as interaction information and captures the unique information gained by knowing all three variables beyond only knowing subsets of the variables. There are other extensions of MI to more than 2 variables. In particular total correlation allows quantification of the shared information between all three variables. As a result, total correlation has been used for feature selection. Did you consider using total correlation in this case?