Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2018 Jan 1;19(1):136-147.
doi: 10.1093/bib/bbw086.

Transferring entropy to the realm of GxG interactions

Transferring entropy to the realm of GxG interactions

Paola G Ferrario et al. Brief Bioinform. .

Abstract

Genome-wide association studies are moving to genome-wide interaction studies, as the genetic background of many diseases appears to be more complex than previously supposed. Thus, many statistical approaches have been proposed to detect gene-gene (GxG) interactions, among them numerous information theory-based methods, inspired by the concept of entropy. These are suggested as particularly powerful and, because of their nonlinearity, as better able to capture nonlinear relationships between genetic variants and/or variables. However, the introduced entropy-based estimators differ to a surprising extent in their construction and even with respect to the basic definition of interactions. Also, not every entropy-based measure for interaction is accompanied by a proper statistical test. To shed light on this, a systematic review of the literature is presented answering the following questions: (1) How are GxG interactions defined within the framework of information theory? (2) Which entropy-based test statistics are available? (3) Which underlying distribution do the test statistics follow? (4) What are the given strengths and limitations of these test statistics?

Keywords: entropy; estimation; genetic interactions; information theory.

PubMed Disclaimer

Figures

Figure 1
Figure 1
H(X1): Entropy of a random variable X1:=E[log2(1P(X))]=i=1mpilog2(pi)
Figure 2
Figure 2
H(X1,X2): Joint entropy of the two random variables X1 and X2:=E[log2(1P(X1,X2))]=i=1mj=1mpijlog2(pij)
Figure 3
Figure 3
H(X1|X2): Conditional entropy of the variable X1 given the variable X2:=E[log2(P(X2)P(X1,X2))]=i=1mj=1mpijlog2(pijp·j)=H(X1,X2)H(X2)
Figure 4
Figure 4
I(X1,X2): Mutual information of the variables X1 and X2:=E[log2(P(X1,X2)P(X1)P(X2))]=i=1mj=1mpijlog2(pijpi·p·j)=H(X2)H(X2|X1)=H(X1)H(X1|X2)
Figure 5
Figure 5
TCI(X1,X2,X3): Total correlation information of three random variables H(X1)+H(X2)+H(X3)H(X1,X2,X3)
Figure 6
Figure 6
3WII(X1,X2,X3): Three-way interaction information of three random variables H(X1,X2)+H(X1,X3)+H(X2,X3)H(X1)H(X2)H(X3)H(X1,X2,X3)
Figure 7
Figure 7
Flow diagram of the search process.

Similar articles

Cited by

References

    1. Cordell HJ. Detecting gene-gene interactions that underlie human diseases. Nat Rev Genet 2009;10:392–404. - PMC - PubMed
    1. Gola D, Mahachie John JM, van Steen K, et al.A roadmap to multifactor dimensionality reduction methods. Brief Bioinform 2016;17(2):293–308. - PMC - PubMed
    1. Ritchie MD, Hahn LW, Roodi N, et al.Multifactor-dimensionality reduction reveals high-order interactions among estrogen-metabolism genes in sporadic breast cancer. Am J Hum Genet 2001;69:138–47. - PMC - PubMed
    1. Calle ML, Urrea V, Vellalta G, et al.Improving strategies for detecting genetic patterns of disease susceptibility in association studies. Stat Med 2008;27:6532–46. - PubMed
    1. Breiman L. Random forests. Mach Learn 2001;45:5–32.

Publication types