Open Access

minet: A R/Bioconductor Package for Inferring Large Transcriptional Networks Using Mutual Information

  • Patrick E Meyer1Email author,
  • Frédéric Lafitte1 and
  • Gianluca Bontempi1
BMC Bioinformatics20089:461

DOI: 10.1186/1471-2105-9-461

Received: 02 July 2008

Accepted: 29 October 2008

Published: 29 October 2008



This paper presents the R/Bioconductor package minet (version 1.1.6) which provides a set of functions to infer mutual information networks from a dataset. Once fed with a microarray dataset, the package returns a network where nodes denote genes, edges model statistical dependencies between genes and the weight of an edge quantifies the statistical evidence of a specific (e.g transcriptional) gene-to-gene interaction. Four different entropy estimators are made available in the package minet (empirical, Miller-Madow, Schurmann-Grassberger and shrink) as well as four different inference methods, namely relevance networks, ARACNE, CLR and MRNET. Also, the package integrates accuracy assessment tools, like F-scores, PR-curves and ROC-curves in order to compare the inferred network with a reference one.


The package minet provides a series of tools for inferring transcriptional networks from microarray data. It is freely available from the Comprehensive R Archive Network (CRAN) as well as from the Bioconductor website.


Modelling transcriptional interactions by large networks of interacting elements and determining how these interactions can be effectively learned from measured expression data are two important issues in system biology [1]. It should be noted that by focusing only on transcript data, the inferred network should not be considered as a proper biochemical regulatory network, but rather as a gene-to-gene network where many physical connections between macromolecules might be hidden by short-cuts. In spite of some evident limitations the bioinformatics community made important advances in this domain over the last few years [2, 3]. In particular, mutual information networks have been succesfully applied to transcriptional network inference [46]. Such methods, which typically rely on the estimation of mutual information between all pairs of variables, have recently held the attention of the bioinformatics community for the inference of very large networks (up to several thousands nodes) [4, 79].

R is a widely used open source language and environment for statistical computing and graphics [10] which has become a de-facto standard in statistical modeling, data analysis, biostatistics and machine learning [11]. An important feature of the R environment is that it integrates generic data analysis and visualization functionalities with off-the-shelf packages implementing the latest advances in computational statistics. Bioconductor is an open source and open development software project for the analysis and comprehension of genomic data [12] mainly based on the R programming language. This paper introduces the new R and Bioconductor package minet, where the acronym stands for Mutual Information NETwork inference. This package is freely available on the R CRAN package resource [10] as well as on the Bioconductor website [12].

1 Mutual information networks

Mutual information networks are a subcategory of network inference methods. The rationale of this family of methods is to infer a link between a couple of nodes if it has a high score based on mutual information [9].

Mutual informaton network inference proceeds in two steps. The first step is the computation of the mutual information matrix (MIM), a square matrix whose i, j-th element

M I M ij = I(X i ; X j ) (1)

is the mutual information between X i and X j , where X i X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ , i = 1,...,n, is a discrete random variable denoting the expression level of the i th gene. The second step is the computation of an edge score for each pair of nodes by an inference algorithm that takes the MIM matrix as input.

The adoption of mutual information in network inference tasks can be traced back to the Chow and Liu's tree algorithm [13, 14]. Mutual information provides a natural generalization of the correlation since it is a non-linear measure of dependency. Hence with mutual information generalized correlation networks (relevance networks [7]) and also conditional independence graphs (e.g. ARACNE [8]) can be built. An advantage of these methods is their ability to deal with up to several thousands of variables also in the presence of a limited number of samples. This is made possible by the fact that the MIM computation requires only n ( n 1 ) 2 MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqWGUbGBcqGGOaakcqWGUbGBcqGHsislcqaIXaqmcqGGPaqkaeaacqaIYaGmaaaaaa@33BE@ estimations of a bivariate mutual information term. Since each bivariate estimation can be computed fastly and is low variant also for a small number of samples, this family of methods is adapted for dealing with microarray data. Note that since mutual information is a symmetric measure, it is not possible to derive the direction of an edge using a mutual information network inference technique. Notwithstanding the orientation of the edges can be obtained by using algorithms like IC which are well known in the graphical modelling community [15].

1.1 Relevance Network

The relevance network approach [7] has been introduced in gene clustering and was successfully applied to infer relationships between RNA expressions and chemotherapeutic susceptibility [6]. The approach consists in inferring a genetic network where a pair of genes {X i , X j } is linked by an edge if the mutual information I(X i ; X j ) is larger than a given threshold I0. The complexity of the method is O(n2) since all pairwise interactions are considered.

Note that this method does not eliminate all the indirect interactions between genes. For example, if gene X1 regulates both gene X2 and gene X3, this would cause a high mutual information between the pairs {X1, X2}, {X1, X3} and {X2, X3}. As a consequence, the algorithm will set an edge between X2 and X3 although these two genes interact only through gene X1.

1.2 CLR Algorithm

The CLR algorithm [4] is an extension of the relevance network approach. This algorithm computes the mutual information for each pair of genes and derives a score related to the empirical distribution of the MI values. In particular, instead of considering the information I(X i ; X j ) between genes X i and X j , it takes into account the score z i j = z i 2 + z j 2 MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOEaO3aaSbaaSqaaiabdMgaPjabdQgaQbqabaGccqGH9aqpdaGcaaqaaiabdQha6naaDaaaleaacqWGPbqAaeaacqaIYaGmaaGccqGHRaWkcqWG6bGEdaqhaaWcbaGaemOAaOgabaGaeGOmaidaaaqabaaaaa@3A32@ where
z i = max ( 0 , I ( X i ; X j ) μ i σ i ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOEaO3aaSbaaSqaaiabdMgaPbqabaGccqGH9aqpcyGGTbqBcqGGHbqycqGG4baEdaqadaqaaiabicdaWiabcYcaSKqbaoaalaaabaGaemysaKKaeiikaGIaemiwaG1aaSbaaeaacqWGPbqAaeqaaiabcUda7iabdIfaynaaBaaabaGaemOAaOgabeaacqGGPaqkcqGHsislcqaH8oqBdaWgaaqaaiabdMgaPbqabaaabaGaeq4Wdm3aaSbaaeaacqWGPbqAaeqaaaaaaOGaayjkaiaawMcaaaaa@48F1@

and μ i and σ i are respectively the sample mean and standard deviation of the empirical distribution of the values I(X i , X k ), k = 1,...,n. The CLR algorithm was successfully applied to decipher the E. Coli TRN [4]. CLR has a complexity in O(n2) once the MIM is computed.


The Algorithm for the Reconstruction of Accurate Cellular Networks (ARACNE) [8] is based on the Data Processing Inequality [16]. This inequality states that, if gene X1 interacts with gene X3 through gene X2, thenI(X1; X3) ≤ min (I(X1; X2), I(X2; X3)).

ARACNE starts by assigning to each pair of nodes a weight equal to the mutual information. Then, as in relevance networks, all edges for which I(X i ; X j ) <I0 are removed, with I0 a given threshold. Eventually, the weakest edge of each triplet is interpreted as an indirect interaction and is removed if the difference between the two lowest weights is above a threshold W0. Note that by increasing I0 the number of inferred edges is decreased while the opposite effect is obtained by increasing W0.

If the network is a tree and only pairwise interactions are present, the method guarantees the reconstruction of the original network, once it is provided with the exact MIM. ARACNE's complexity is O(n3) since the algorithm considers all triplets of genes. In [8] the method was able to recover components of the TRN in mammalian cells and outperformed Bayesian networks and relevance networks on several inference tasks [8].


MRNET [9] infers a network using the maximum relevance/minimum redundancy (MRMR) feature selection method [17, 18]. The idea consists in performing a series of supervised MRMR gene selection procedures where each gene in turn plays the role of the target output.

The MRMR method has been introduced in [17, 18] together with a best-first search strategy for performing filter selection in supervised learning problems. Consider a supervised learning task where the output is denoted by Y and V is the set of input variables. The method ranks the set V of inputs according to a score that is the difference between the mutual information with the output variable Y (maximum relevance) and the average mutual information with all the previously ranked variables (minimum redundancy). The rationale is that direct interactions (i.e. the most informative variables to the target Y) should be well ranked whereas indirect interactions (i.e. the ones with redundant information with the direct ones) should be badly ranked by the method. The greedy search starts by selecting the variable X i having the highest mutual information to the target Y. The second selected variable X j will be the one with a high information I(X j ; Y) to the target and at the same time a low information I(X j ; X i ) to the previously selected variable. In the following steps, given a set S of selected variables, the criterion updates S by choosing the variable
X j M R M R = arg max X j V \ S ( u j r j ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemiwaG1aa0baaSqaaiabdQgaQbqaaiabd2eanjabdkfasjabd2eanjabdkfasbaakiabg2da9iGbcggaHjabckhaYjabcEgaNnaaxababaGagiyBa0MaeiyyaeMaeiiEaGhaleaacqWGybawdaWgaaadbaGaemOAaOgabeaaliabgIGiolabdAfawjabcYfaCjabdofatbqabaGccqGGOaakcqWG1bqDdaWgaaWcbaGaemOAaOgabeaakiabgkHiTiabdkhaYnaaBaaaleaacqWGQbGAaeqaaOGaeiykaKcaaa@4DA8@

that maximizes the scores j = u j - r j ,

where u j is a relevance term and r j is a redundancy term. More precisely,u j = I(X j ; Y)

is the mutual information of X j with the target variable Y, and
r j = 1 | S | X k S I ( X j ; X k ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOCai3aaSbaaSqaaiabdQgaQbqabaGccqGH9aqpjuaGdaWcaaqaaiabigdaXaqaaiabcYha8jabdofatjabcYha8baakmaaqafabaGaemysaKKaeiikaGIaemiwaG1aaSbaaSqaaiabdQgaQbqabaGccqGG7aWocqWGybawdaWgaaWcbaGaem4AaSgabeaakiabcMcaPaWcbaGaemiwaG1aaSbaaWqaaiabdUgaRbqabaWccqGHiiIZcqWGtbWuaeqaniabggHiLdaaaa@46FB@

measures the average redundancy of X j to each already selected variables X k S. At each step of the algorithm, the selected variable is expected to allow an efficient trade-off between relevance and redundancy. It has been shown in [19] that the MRMR criterion is an optimal "pairwise" approximation of the conditional mutual information between any two genes X i and X j given the set S of selected variables I(X i ; X j |S).

The MRNET approach consists in repeating this selection procedure for each target gene by putting Y = X i and V = X \ {X i }, i = 1,...,n, where X is the set of the expression levels of all genes. For each pair {X i , X j }, MRMR returns two (not necessarily equal) scores s i and s j according to (4). The score of the pair {X i , X j } is then computed by taking the maximum of s i and s j . A specific network can then be inferred by deleting all the edges whose score lies below a given threshold I0 (as in relevance networks, CLR and ARACNE). Thus, the algorithm infers an edge between X i and X j either when X i is a well-ranked predictor of X j (s i > I0) or when X j is a well-ranked predictor of X i (s j > I0).

An effective implementation of the best-first search for quadratic problems is available in [20]. This implementation demands an O(f × n) complexity for selecting f features using a best first search strategy. It follows that MRNET has an O(f × n2) complexity since the feature selection step is repeated for each of the n genes. In other terms, the complexity ranges between O(n2) and O(n3) according to the value of f. In practice the selection of features stops once a variable obtains a negative score.

Implementation of the inference algorithms in minet

All the algorithms discussed above are available in the minet package. The RELNET algorithm is implemented by simply running the command build.mim which returns the MIM matrix which can be considered as a weighted adjacency matrix of the network. CLR, ARACNE and MRNET are implemented by the commands aracne(mim), clr(mim), mrnet(mim) respectively that return a weighted adjacency matrix of the network.

It should be noted, that the modularity of the minet package makes possible to assess network inference methods on similarity matrices other than MIM [21].

2 Mutual information estimation

An information-theoretic network inference technique aims at identifying connections between two genes (variables) by estimating the amount of information common to any pair of genes. Mutual information is a measure which calculates dependencies between two discrete random variables. An important property of this measure is that it is not restricted to the identification of linear relations between the random variables [16].

If X is a continuous random variable taking values between a and b, the interval [a, b] can be discretized by partitioning it into | X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ | subintervals, called bins, where the symbol X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ denotes the bin index vector. We use also nb(x k ) to denote the number of data points in the k th bin and the symbol m = k X n b ( x k ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemyBa0Maeyypa0ZaaabeaeaacqWGUbGBcqWGIbGycqGGOaakcqWG4baEdaWgaaWcbaGaem4AaSgabeaakiabcMcaPaWcbaGaem4AaSMaeyicI48enfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfabeqdcqGHris5aaaa@4618@ to denote the number of samples. If X is a random vector each element X i can be discretized separately into | X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ | bins with index vector X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ .

Let X be a random vector and p a probability measure. The i, j-th element of the mutual information matrix (MIM) is defined by
M I M i j = H ( X i ) + H ( X j ) H ( X i , X j ) = I ( X i ; X j ) = k i X i k j X j p ( x k i , x k j ) log ( p ( x k i , x k j ) p ( x k i ) p ( x k j ) ) , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGceaabbeaacqWGnbqtcqWGjbqscqWGnbqtdaWgaaWcbaGaemyAaKMaemOAaOgabeaakiabg2da9iabdIeaijabcIcaOiabdIfaynaaBaaaleaacqWGPbqAaeqaaOGaeiykaKIaey4kaSIaemisaGKaeiikaGIaemiwaG1aaSbaaSqaaiabdQgaQbqabaGccqGGPaqkcqGHsislcqWGibascqGGOaakcqWGybawdaWgaaWcbaGaemyAaKgabeaakiabcYcaSiabdIfaynaaBaaaleaacqWGQbGAaeqaaOGaeiykaKcabaGaeyypa0JaemysaKKaeiikaGIaemiwaG1aaSbaaSqaaiabdMgaPbqabaGccqGG7aWocqWGybawdaWgaaWcbaGaemOAaOgabeaakiabcMcaPaqaaiabg2da9maaqafabaWaaabuaeaacqWGWbaCcqGGOaakcqWG4baEdaWgaaWcbaGaem4AaS2aaSbaaWqaaiabdMgaPbqabaaaleqaaOGaeiilaWIaemiEaG3aaSbaaSqaaiabdUgaRnaaBaaameaacqWGQbGAaeqaaaWcbeaakiabcMcaPiGbcYgaSjabc+gaVjabcEgaNnaabmaajuaGbaWaaSaaaeaacqWGWbaCcqGGOaakcqWG4baEdaWgaaqaaiabdUgaRnaaBaaabaGaemyAaKgabeaaaeqaaiabcYcaSiabdIha4naaBaaabaGaem4AaS2aaSbaaeaacqWGQbGAaeqaaaqabaGaeiykaKcabaGaemiCaaNaeiikaGIaemiEaG3aaSbaaeaacqWGRbWAdaWgaaqaaiabdMgaPbqabaaabeaacqGGPaqkcqWGWbaCcqGGOaakcqWG4baEdaWgaaqaaiabdUgaRnaaBaaabaGaemOAaOgabeaaaeqaaiabcMcaPaaaaOGaayjkaiaawMcaaaWcbaGaem4AaS2aaSbaaWqaaiabdQgaQbqabaWccqGHiiIZt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwdaWgaaadbaGaemOAaOgabeaaaSqab0GaeyyeIuoaaSqaaiabdUgaRnaaBaaameaacqWGPbqAaeqaaSGaeyicI4Sae83fXJ1aaSbaaWqaaiabdMgaPbqabaaaleqaniabggHiLdGccqGGSaalaaaa@A390@
where the entropy of a random variable X is defined as
H ( X ) = k X p ( x k ) log p ( x k ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemisaGKaeiikaGIaemiwaGLaeiykaKIaeyypa0JaeyOeI0YaaabuaeaacqWGWbaCcqGGOaakcqWG4baEdaWgaaWcbaGaem4AaSgabeaakiabcMcaPaWcbaGaem4AaSMaeyicI48enfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfabeqdcqGHris5aOGagiiBaWMaei4Ba8Maei4zaCMaemiCaaNaeiikaGIaemiEaG3aaSbaaSqaaiabdUgaRbqabaGccqGGPaqkaaa@533C@

and I(X i ; X j ) is the mutual information between the random variables X i and X j .

Hence, each mutual information calculus demands the estimation of three entropy terms (Eq. 5). A fast entropy estimation is therefore essential for an effective network inference based on MI. Entropy estimation has gained much interest in feature selection and network inference over the last decade [22]. Most approaches focus on reducing the bias inherent to entropy estimation. In this section, some of the fastest and most used entropy estimators are stressed. Other interesting approaches can be found in [2226].

2.1 Empirical and Miller-Madow corrected estimators

The empirical estimator (also called "plug-in", "maximum likelihood" or "naïve", see [23]) is the entropy of the empirical distribution.
H ^ e m p = k X n b ( x k ) m log n b ( x k ) m . MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGafmisaGKbaKaadaahaaWcbeqaaiabdwgaLjabd2gaTjabdchaWbaakiabg2da9iabgkHiTmaaqafajuaGbaWaaSaaaeaacqWGUbGBcqWGIbGycqGGOaakcqWG4baEdaWgaaqaaiabdUgaRjabcMcaPaqabaaabaGaemyBa0gaaaWcbaGaem4AaSMaeyicI48enfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfabeqdcqGHris5aOGagiiBaWMaei4Ba8Maei4zaCwcfa4aaSaaaeaacqWGUbGBcqWGIbGycqGGOaakcqWG4baEdaWgaaqaaiabdUgaRbqabaGaeiykaKcabaGaemyBa0gaaOGaeiOla4caaa@5C0F@
Note that, because of the convexity of the logarithmic function, an underestimate of p(x k ) causes an error on H(X = x k ) that is larger than the one given by an overestimation of the same quantity. As a result, entropy estimators are biased downwards, that is
E [ H ^ e m p ( p X ) ] H ( p X ) . MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemyrauKaei4waSLafmisaGKbaKaadaahaaWcbeqaaiabdwgaLjabd2gaTjabdchaWbaakiabcIcaOiabdchaWjabdIfayjabcMcaPiabc2faDjabgsMiJkabdIeaijabcIcaOiabdchaWjabdIfayjabcMcaPiabc6caUaaa@418F@

It has been shown that the variance of the empirical estimator is upper-bounded by v a r ( H ^ e m p ) ( ( log m ) 2 m ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemODayNaemyyaeMaemOCaiNaeiikaGIafmisaGKbaKaadaahaaWcbeqaaiabdwgaLjabd2gaTjabdchaWbaakiabcMcaPiabgsMiJoaabmaajuaGbaWaaSaaaeaacqGGOaakcyGGSbaBcqGGVbWBcqGGNbWzcqWGTbqBcqGGPaqkdaahaaqabeaacqaIYaGmaaaabaGaemyBa0gaaaGccaGLOaGaayzkaaaaaa@44C3@ which depends only on the number of samples whereas the asymptotic bias of the estimate b i a s ( H ^ e m p ) = | X | 1 2 m MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOyaiMaemyAaKMaemyyaeMaem4CamNaeiikaGIafmisaGKbaKaadaahaaWcbeqaaiabdwgaLjabd2gaTjabdchaWbaakiabcMcaPiabg2da9iabgkHiTKqbaoaalaaabaGaeiiFaW3enfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJLaeiiFaWNaeyOeI0IaeGymaedabaGaeGOmaiJaemyBa0gaaaaa@4DCD@ depends also on the number of bins | X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ | [23]. As | X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ | m, this estimator can still have a low variance but the bias can become very large [23].

The Miller-Madow correction is then given by the following formula which is the empirical entropy corrected by the asymptotic bias,
H ^ m m = H ^ e m p + | X | 1 2 m . MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGafmisaGKbaKaadaahaaWcbeqaaiabd2gaTjabd2gaTbaakiabg2da9iqbdIeaizaajaWaaWbaaSqabeaacqWGLbqzcqWGTbqBcqWGWbaCaaGccqGHRaWkjuaGdaWcaaqaaiabcYha8nrtHrhAL1wy0L2yHvtyaeHbnfgDOvwBHrxAJfwnaGabaiab=Dr8yjabcYha8jabgkHiTiabigdaXaqaaiabikdaYiabd2gaTbaakiabc6caUaaa@4C10@

where | X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ | is the number of bins with non-zero probability. This correction, while adding no computational cost to the empirical estimator, reduces the bias without changing variance. As a result, the Miller-Madow estimator is often preferred to the naive empirical entropy estimator.

2.2 Shrink entropy estimator

The rationale of the shrink estimator, [27], is to combine two different estimators, one with low variance and one with low bias, by using a weighting factor λ [0,1]
p ^ λ ( x k ) = λ 1 | X | + ( 1 λ ) n b ( x k ) m . MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGafmiCaaNbaKaadaWgaaWcbaGaeq4UdWgabeaakiabcIcaOiabdIha4naaBaaaleaacqWGRbWAaeqaaOGaeiykaKIaeyypa0Jaeq4UdWwcfa4aaSaaaeaacqaIXaqmaeaacqGG8baFt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwcqGG8baFaaGccqGHRaWkcqGGOaakcqaIXaqmcqGHsislcqaH7oaBcqGGPaqkjuaGdaWcaaqaaiabd6gaUjabdkgaIjabcIcaOiabdIha4naaBaaabaGaem4AaSgabeaacqGGPaqkaeaacqWGTbqBaaGccqGGUaGlaaa@57AB@

Shrinkage is a general technique to improve an estimator for a small sample size [3]. As the value of λ tends to one, the estimated entropy is moved toward the maximal entropy (uniform probability) whereas when λ is zero the estimated entropy tends to the value of the empirical one.

Let λ* be the value minimizing the mean square function, see [27],
λ * = arg min λ [ 0 , 1 ] E [ k X ( p ^ λ ( x k ) p ( x k ) ) 2 ] . MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaeq4UdWMaeiOkaOIaeyypa0JagiyyaeMaeiOCaiNaei4zaC2aaCbeaeaacyGGTbqBcqGGPbqAcqGGUbGBaSqaaiabeU7aSjabgIGiolabcUfaBjabicdaWiabcYcaSiabigdaXiabc2faDbqabaGccqWGfbqrdaWadaqaamaaqafabaGaeiikaGIafmiCaaNbaKaadaWgaaWcbaGaeq4UdWgabeaakiabcIcaOiabdIha4naaBaaaleaacqWGRbWAaeqaaOGaeiykaKcaleaacqWGRbWAcqGHiiIZt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwaeqaniabggHiLdGccqGHsislcqWGWbaCcqGGOaakcqWG4baEdaWgaaWcbaGaem4AaSgabeaakiabcMcaPiabcMcaPmaaCaaaleqabaGaeGOmaidaaaGccaGLBbGaayzxaaGaeiOla4caaa@6751@
It has been shown in [28] that the optimal λ is given by
λ * = | X | ( m 2 k X n b ( x k ) 2 ) ( m 1 ) ( | X | k X n b ( x k ) 2 m 2 ) . MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaeq4UdWMaeiOkaOIaeyypa0tcfa4aaSaaaeaacqGG8baFt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwcqGG8baFcqGGOaakcqWGTbqBdaahaaqabeaacqaIYaGmaaGaeyOeI0YaaabeaeaacqWGUbGBcqWGIbGycqGGOaakcqWG4baEdaWgaaqaaiabdUgaRbqabaGaeiykaKYaaWbaaeqabaGaeGOmaidaaaqaaiabdUgaRjabgIGiolab=Dr8ybqabiabggHiLdGaeiykaKcabaGaeiikaGIaemyBa0MaeyOeI0IaeGymaeJaeiykaKIaeiikaGIaeiiFaWNae83fXJLaeeiFaW3aaabeaeaacqWGUbGBcqWGIbGycqGGOaakcqWG4baEdaWgaaqaaiabdUgaRbqabaGaeiykaKYaaWbaaeqabaGaeGOmaidaaaqaaiabdUgaRjabgIGiolab=Dr8ybqabiabggHiLdGaeyOeI0IaemyBa02aaWbaaeqabaGaeGOmaidaaiabcMcaPaaakiabc6caUaaa@71F1@
H ^ s h r i n k = k X p ^ λ ( x k ) log p ^ λ ( x k ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGafmisaGKbaKaadaahaaWcbeqaaiabdohaZjabdIgaOjabdkhaYjabdMgaPjabd6gaUjabdUgaRbaakiabg2da9iabgkHiTmaaqafabaGafmiCaaNbaKaadaWgaaWcbaGaeq4UdWgabeaakiabcIcaOiabdIha4naaBaaaleaacqWGRbWAaeqaaOGaeiykaKIagiiBaWMaei4Ba8Maei4zaCMafmiCaaNbaKaadaWgaaWcbaGaeq4UdWgabeaakiabcIcaOiabdIha4naaBaaaleaacqWGRbWAcqGGPaqkaeqaaaqaaiabdUgaRjabgIGioprtHrhAL1wy0L2yHvtyaeHbnfgDOvwBHrxAJfwnaGabaiab=Dr8ybqab0GaeyyeIuoaaaa@5CC1@

2.3 The Schurmann-Grassberger Estimator

The Dirichlet distribution can be used in order to estimate the entropy of a discrete random variable. The Dirichlet distribution is the multivariate generalization of the beta distribution. It is also the conjugate prior of the multinomial distribution in Bayesian statistics. More precisely, the density of a Dirichlet distribution takes the following form
f ( X ; β ) = k X Γ ( β k ) Γ ( k X β k ) k X x k β k 1 MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOzayMaeiikaGIaemiwaGLaei4oaSJaeqOSdiMaeiykaKIaeyypa0tcfa4aaSaaaeaadaqeqaqaaiabfo5ahjabcIcaOiabek7aInaaBaaabaGaem4AaSgabeaacqGGPaqkaeaacqWGRbWAcqGHiiIZt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwaeqacqGHpis1aaqaaiabfo5ahnaabmaabaWaaabeaeaacqaHYoGydaWgaaqaaiabdUgaRbqabaaabaGaem4AaSMaeyicI4Sae83fXJfabeGaeyyeIuoaaiaawIcacaGLPaaaaaGcdaqeqbqaaiabdIha4naaDaaaleaacqWGRbWAaeaacqaHYoGydaWgaaadbaGaem4AaSMaeyOeI0IaeGymaedabeaaaaaaleaacqWGRbWAcqGHiiIZcqWFxepwaeqaniabg+Givdaaaa@66B8@

where β i is the prior probability of an event x i and Γ(·) is the gamma function, (see [25, 27, 29] for more details).

In case of no a priori knowledge, the β k are assumed to be equal (β k = N, k X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ ) so as no event becomes more probable than another. Note that using a Dirichlet prior with parameters N is equivalent to adding N ≥ 0 "pseudo-counts" to each bin i X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ . The prior actually provides the estimator the information that | X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ |N counts have been observed in previous experiments. From that viewpoint, | X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ |N becomes the a priori sample size.

The entropy of a Dirichlet distribution can be computed directly with the following equation:
H ^ d i r ( X ) = 1 m + | X | N k X ( n b ( x k ) + N ) ( ψ ( m + | X | N + 1 ) ψ ( n b ( x k ) + N + 1 ) ) MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGafmisaGKbaKaadaahaaWcbeqaaiabdsgaKjabdMgaPjabdkhaYbaakiabcIcaOiabdIfayjabcMcaPiabg2da9KqbaoaalaaabaGaeGymaedabaGaemyBa0Maey4kaSIaeiiFaW3enfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJLaeiiFaWNaemOta4eaaOWaaabuaeaacqGGOaakcqWGUbGBcqWGIbGycqGGOaakcqWG4baEdaWgaaWcbaGaem4AaSgabeaakiabcMcaPiabgUcaRiabd6eaojabcMcaPiabcIcaOiabeI8a5jabcIcaOiabd2gaTjabgUcaRiabcYha8jab=Dr8yjabcYha8jabd6eaojabgUcaRiabigdaXiabcMcaPiabgkHiTiabeI8a5jabcIcaOiabd6gaUjabdkgaIjabcIcaOiabdIha4naaBaaaleaacqWGRbWAaeqaaOGaeiykaKIaey4kaSIaemOta4Kaey4kaSIaeGymaeJaeiykaKIaeiykaKcaleaacqWGRbWAcqGHiiIZcqqGybawaeqaniabggHiLdaaaa@79A4@

with ψ ( z ) = d ln Γ ( z ) d z MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaeqiYdKNaeiikaGIaemOEaONaeiykaKIaeyypa0tcfa4aaSaaaeaacqWGKbazcyGGSbaBcqGGUbGBcqqHtoWrcqGGOaakcqWG6bGEcqGGPaqkaeaacqWGKbazcqWG6bGEaaaaaa@3DF2@ the digamma function.

Various choices of prior parameters has been proposed in the literature [2931]. Schurmann and Grassberger have proposed the prior N = 1 | X | MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOta4Kaeyypa0tcfa4aaSaaaeaacqaIXaqmaeaacqGG8baFt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwcqGG8baFaaaaaa@3E30@ [32] that has been retained in the package.

Implementation of estimators in minet

The mutual information matrix is estimated by using the function build.mim(dataset, estimator). This function returns a matrix of paired mutual informations computed in nats (base e) and takes two arguments:
  1. 1.

    the data frame dataset which stores the gene expression dataset or a generic dataset where columns contain variables/features and rows contain outcomes/samples

  2. 2.

    the string mi, that denotes the routine used to perform mutual information estimator.


The package makes available four estimation routines : "mi.empirical", "mi.shrink", "","" (default:"mi.empirical") each referring to the estimators technique explained above.

3 Discretization methods

All the estimators discussed in the previous section have been designed for discrete variables. If the random variable X is continuous and takes values comprised between a and b, it is then required to partition the interval [a, b] into | X MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJfaaa@3777@ | sub-intervals in order to adopt a discrete entropy estimator. The two most used discretizing algorithm are the equal width and the equal frequency quantization. These are explained in the next sections. Other discretization methods can be found in [3335].

3.1 Equal Width

The principle of the equal width discretization is to divide the range [a i , b i ] of each variable X i , i {1, 2,...,n} in the dataset into | X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ | sub-intervals of equal size: [ a i , a i + b i a i | X i | [ , [ a i + b i a i | X i | , a i + 2 b i a i | X i | [ , ... [ a i + ( | X i | 1 ) ( b i a i ) | X i | , b i + ε [ MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaei4waSLaemyyae2aaSbaaSqaaiabdMgaPjabcYcaSaqabaGccqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabgUcaRKqbaoaalaaabaGaemOyai2aaSbaaeaacqWGPbqAaeqaaiabgkHiTiabdggaHnaaBaaabaGaemyAaKgabeaaaeaacqGG8baFt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwdaWgaaqaaiabdMgaPbqabaGaeiiFaWhaaOGaei4waSLaeiilaWIaei4waSLaemyyae2aaSbaaSqaaiabdMgaPbqabaGccqGHRaWkjuaGdaWcaaqaaiabdkgaInaaBaaabaGaemyAaKgabeaacqGHsislcqWGHbqydaWgaaqaaiabdMgaPbqabaaabaGaeiiFaWNae83fXJ1aaSbaaeaacqWGPbqAaeqaaiabcYha8baakiabcYcaSiabdggaHnaaBaaaleaacqWGPbqAaeqaaOGaey4kaSIaeGOmaitcfa4aaSaaaeaacqWGIbGydaWgaaqaaiabdMgaPbqabaGaeyOeI0Iaemyyae2aaSbaaeaacqWGPbqAaeqaaaqaaiabcYha8jab=Dr8ynaaBaaabaGaemyAaKgabeaacqGG8baFaaGccqGGBbWwcqGGSaalcqGGUaGlcqGGUaGlcqGGUaGlcqGGBbWwcqWGHbqydaWgaaWcbaGaemyAaKgabeaakiabgUcaRKqbaoaalaaabaGaeiikaGIaeiiFaWNae83fXJ1aaSbaaeaacqWGPbqAaeqaaiabcYha8jabgkHiTiabigdaXiabcMcaPiabcIcaOiabdkgaInaaBaaabaGaemyAaKgabeaacqGHsislcqWGHbqydaWgaaqaaiabdMgaPbqabaGaeiykaKcabaGaeiiFaWNae83fXJ1aaSbaaeaacqWGPbqAaeqaaiabcYha8baakiabcYcaSiabdkgaInaaBaaaleaacqWGPbqAaeqaaOGaey4kaSIaeqyTduMaei4waSfaaa@9DF9@ . Note that an ε is added in the last interval in order to include the maximal value in one of the | X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ | bins. This discretization scheme has a O(m) complexity cost (by variable).

3.2 Global Equal Width

The principle of the global equal width discretization is the same as the equal width (Sec. 3.1) except that the considered range [a, b] is not the range of each random variable such as in Sec. 3.1 but the range of the random vector composed of all the variables in the dataset. In other words, a and b are respectively the minimal and the maximal value of the dataset.

3.3 Equal Frequency

The equal frequency discretization scheme consists in partitioning the range [a i , b i ] of each variable X i in the dataset into | X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ | intervals, each having the same number m/| X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ | of data points points. As a result, the size of each interval can be different. Note that if the | X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ | intervals have equal frequencies, the computation of entropy is straightforward: it is log 1 | X i | MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqaIXaqmaeaacqGG8baFt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwdaWgaaqaaiabdMgaPbqabaGaeiiFaWhaaaaa@3D81@ . However, there can be more than m/| X i MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWenfgDOvwBHrxAJfwnHbqeg0uy0HwzTfgDPnwy1aaceaGae83fXJ1aaSbaaSqaaiabdMgaPbqabaaaaa@38FE@ | identical values in a vector of measurements. In such case, one of the bins will be more dense than the others and the resulting entropy will be different of log 1 | X i | MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaqcfa4aaSaaaeaacqaIXaqmaeaacqGG8baFt0uy0HwzTfgDPnwy1egaryqtHrhAL1wy0L2yHvdaiqaacqWFxepwdaWgaaqaaiabdMgaPbqabaGaeiiFaWhaaaaa@3D81@ . It should be noted that this discretization is reported in some papers as one of the most efficient method (e.g. for naive Bayes classification) [35].

Implementation of discretization strategies in minet

The discretization is performed in minet by the function

discretize(dataset, disc = "equalfreq", nbins = sqrt(nrow(dataset)))


  • dataset is the dataset to be discretized

  • disc is a string which can take three values: "equalfreq" "equalwidth" "globalequalwidth"(default is " equalfreq").

  • nbins, the number of bins to be used for discretization, which is by default set to m MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xH8viVGI8Gi=hEeeu0xXdbba9frFj0xb9qqpG0dXdb9aspeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaWaaOaaaeaacqWGTbqBaSqabaaaaa@2D53@ with m is the number of samples [35]. Note that there are functions used by the built-in R hist() function that can be used here such as nclass. FD(dataset), nclass. scott(dataset) and nclass. Sturges(dataset).

4 Assessment of the network inference algorithm

A network inference problem can be seen as a binary decision problem where the inference algorithm plays the role of a classifier: for each pair of nodes, the algorithm either returns an edge or not. Each pair of nodes can thus be assigned a positive label (an edge) or a negative one (no edge).

A positive label (an edge) predicted by the algorithm is considered as a true positive (TP) or as a false positive (FP) depending on the presence or not of the corresponding edge in the underlying true network, respectively. Analogously, a negative label is considered as a true negative (TN) or a false negative (FN) depending on whether the corresponding edge is present or not in the underlying true network, respectively. Note that all mutual information network inference methods use a threshold value in order to delete the arcs having a too low score. Hence, for each treshold value, a confusion matrix can be computed.

4.1 ROC curves

The false positive rate is defined as
F P R = F P T N + F P , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOrayKaemiuaaLaemOuaiLaeyypa0tcfa4aaSaaaeaacqWGgbGrcqWGqbauaeaacqWGubavcqWGobGtcqGHRaWkcqWGgbGrcqWGqbauaaGccqGGSaalaaa@39D0@
and the true positive rate as
T P R = T P T P + F N , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemivaqLaemiuaaLaemOuaiLaeyypa0tcfa4aaSaaaeaacqWGubavcqWGqbauaeaacqWGubavcqWGqbaucqGHRaWkcqWGgbGrcqWGobGtaaGccqGGSaalaaa@3A08@

also known as recall or sensitivity.

A Receiver Operating Characteristic (ROC) curve, is a graphical plot of the TPR (true positive rate) vs. FPR (false positive rate) for a binary classifier system as the threshold is varied [36]. A perfect classifier would yield a point in the upper left corner (having coordinates [0,1]) of the ROC space, representing 100% TPR (all true positives are found) and 0% FPR (no false positives are found). A completely random guess gives a point along the diagonal line (the so-called line of no-discrimination) which goes from the left bottom to the top right corners. Points above the diagonal line indicate good classification results, while points below the line indicate wrong results.

4.2 PR curves

It is generally recommended [37] to use receiver operator characteristic (ROC) curves when evaluating binary decision problems in order to avoid effects related to the chosen threshold. However, ROC curves can present an overly optimistic view of an algorithm's performance if there is a large skew in the class distribution, as typically encountered in transcriptional network inference because of sparseness. To tackle this problem, precision-recall (PR) curves have been cited as an alternative to ROC curves [38].

Let the precision quantity
p = T P T P + F P , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemiCaaNaeyypa0tcfa4aaSaaaeaacqWGubavcqWGqbauaeaacqWGubavcqWGqbaucqGHRaWkcqWGgbGrcqWGqbauaaGccqGGSaalaaa@37EE@
measure the fraction of real edges among the ones classified as positive and the recall quantity
r = T P T P + F N , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOCaiNaeyypa0tcfa4aaSaaaeaacqWGubavcqWGqbauaeaacqWGubavcqWGqbaucqGHRaWkcqWGgbGrcqWGobGtaaGccqGGSaalaaa@37EE@

also know as true positive rate (TPR), denote the fraction of real edges that are correctly inferred. These quantities depend on the threshold chosen to return a binary decision. The PR curve is a diagram which plots the precision (p) versus recall (r) for different values of the threshold on a two-dimensional coordinate system.

4.3 F-Scores

Note that a compact representation of the PR diagram is returned by the maximum and/or the average of the F-score quantity [39]:
F = 2 p r r + p , MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOrayKaeyypa0tcfa4aaSaaaeaacqaIYaGmcqWGWbaCcqWGYbGCaeaacqWGYbGCcqGHRaWkcqWGWbaCaaGccqGGSaalaaa@3746@

which is an harmonic average of precision and recall.

The general formula for non-negative real β is:
F β = ( 1 β ) ( p r ) β p + r MathType@MTEF@5@5@+=feaagaart1ev2aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacPC6xNi=xI8qiVKYPFjYdHaVhbbf9v8qqaqFr0xc9vqFj0dXdbba91qpepeI8k8fiI+fsY=rqGqVepae9pg0db9vqaiVgFr0xfr=xfr=xc9adbaqaaeGaciGaaiaabeqaaeqabiWaaaGcbaGaemOray0aaSbaaSqaaiabek7aIbqabaGccqGH9aqpjuaGdaWcaaqaaiabcIcaOiabigdaXiabgkHiTiabek7aIjabcMcaPiabcIcaOiabdchaWjabdkhaYjabcMcaPaqaaiabek7aIjabdchaWjabgUcaRiabdkhaYbaaaaa@3FC4@

where β is a parameter denoting the weight of the recall. Two commonly used F-scores are the F2-measure, which weights recall twice as much as precision, and the F0.5-measure, which weights precision twice as much as recall. In transcriptional network inference, precision is often a more desirable feature than recall since it is expensive to investigate if a gene regulates another.

Assesment functionalities in minet

In order to benchmark the inference methods, the package provides a number of assessment tools. The validate(net,, steps = 50) function allows to compare an inferred network net to a reference network, described by a Boolean adjacency matrix. The assessment process consists in removing the inferred edges having a score below a given threshold and in computing the related confusion matrix, for steps thresholds ranging from the minimum to the maximum value of edge weigths. A resulting dataframe table containing the list of all the steps confusion matrices is returned and made available for further analysis.

In particular, the function pr(table) returns the related precisions and recalls, rates(table) computes true positive and false positive rates while the function fscores(table, beta) returns the F β scores. The functions and show.roc(table) allow the user to plot PR-curves and ROC-curves respectively (Figure 3) from a list of confusion matrices.
Figure 3

Precision-Recall curves plotted with

5 Example

Once the R platform is launched, the package, its description and its vignette can be loaded using the following commands:


library(help = minet)


A demo script (demo(demo)) shows the main functionalities of the package that we describe in the following.

In order to infer a network with the minet package, four steps are required:

  • data discretization,

  • MIM computation,

  • network inference,

  • normalization of the network (optional).

The main function of the package is minet which sequentially executes the four steps mentioned above, see Figure 1).
Figure 1

The four steps in the minet function (discretization disc, mutual information matrix build.mim, inference mrnet, aracne, clr and normalization norm.

The function minet(dataset, method, estimator, disc, nbins) takes the following arguments: dataset, a matrix or a dataframe containing the microarray data, method, the inference algorithm (such as ARACNE, CLR or MRNET), estimator, the entropy estimator used for the computation of mutual information (empirical, Miller-Madow, shrink, Schurmannn-Grassberger), disc the binning algorithm (i.e. equal frequency or equal size interval) and the parameter nbins which sets the number of bins to use. The final step of the minet function is the normalization using the norm(net) function. This step normalizes all the weights of the inferred adjancy matrix between 0 and 1. Hence, the minet function returns the inferred network as a weighted adjacency matrix with values ranging from 0 to 1 where the higher is a weight, the higher is the evidence that a gene-gene interaction exists.

For demo purposes the package makes available also the dataset representing the expression of 50 genes in 100 experiments. This dataset has been synthetically generated from the network using the microarray data generator Syntren [40]. This dataset can be loaded with data( and the corresponding original network with data(

Note that the command res<-minet(,"mrnet","mi.shrink","equalwidth",10) is a compact way to execute the following sequence of instructions:





In order to plot a PR-curve (see Figure 3), the functions and validate can be used.

table <- validate(res,

In order to display the inferred network, the Rgraphviz package [41] can be used with the following commands (see Fig. 2):
Figure 2

Graph generated with minet and plotted with Rgraphviz .


graph <- as(res, "graphNEL")


Note that, for the sake of computational efficiency, all the inference functions as well as the entropy estimators are implemented in C++. As a reference, a network of five hundreds variables may be inferred in less than one minute on an Intel Pentium 4 with 2 Ghz and 512 DDR SDRAM.

6 Conclusion

Transcriptional network inference is a key issue toward the understanding of the relationships between the genes of an organism. Notwithstanding, few public domain tools are available once a thourough comparison of existing approaches is at stake. A new R/Bioconductor package, freely available, has been introduced in this paper. This package makes available to biologists and bioinformatics practicioneers a set of tools to infer networks from microarray datasets with a large number (several thousands) of genes. Four information-theoretic methods of network inference (i.e. Relevance Networks, CLR, ARACNE and MRNET), four different entropy estimators (i.e. empirical, Miller-Madow, Schurmann-Grassberger and shrink) and three validation tools (i.e. F-scores, PR curves and ROC curves) are implemented in the package. We deem that this tool is an effective answer to the increasing need of comparative tools in the growing domain of transcriptional network inference from expression data.

Availability and requirements

The R-package minet is freely available from the Comprehensive R Archive Network (CRAN) at as well as from the Bioconductor website The package runs on Linux, Mac OS and MS Windows using an installed version of R.
Table 1

Available functions of the package minet (version 1.1.6)



minet(data, method, estimator, disc, nbins)

Network inference from data

discretize(data, disc, nbins)

Unsupervised discretization

build.mim(data, estimator)

Mutual information matrix estimation

Estimator can be ""mi.empirical","","mi.shrink" and "".


MRNET algorithm


ARACNE algorithm


CLR algorithm


matrix/network normalization

validate(net1, net2, steps)

Computes confusion matrices


Computes precisions and recalls from confusion matrices


Computes true positive rates and false positive rates from confusion matrices

Displays precision-recall curves from confusion matrices


Displays receiver operator caracteristic curves from confusion matrices


Returns a vector of F β -scores from confusion matrices



This work was partially funded by the Communauté Française de Belgique under ARC grant no. 04/09-307. The authors thank their collegue Catharina Olsen for her appreciable comments, suggestions and testing of package functionalities. The authors also thank Korbinian Strimmer as well as the reviewers for their useful comments on the package and the paper.

Authors’ Affiliations

Machine Learning Group, Computer Science Department, Faculty of Science, Université Libre de Bruxelles


  1. van Someren EP, Wessels LFA, Backer E, Reinders MJT: Genetic network modeling. Pharmacogenomics 2002, 3(4):507–525. 10.1517/14622416.3.4.507View ArticlePubMedGoogle Scholar
  2. Gardner TS, Faith J: Reverse-engineering transcription control networks. Physics of Life Reviews 2 2005.Google Scholar
  3. Schäfer J, Strimmer K: An empirical Bayes approach to inferring large-scale gene association networks. Bioinformatics 2005, 21(6):754–764. 10.1093/bioinformatics/bti062View ArticlePubMedGoogle Scholar
  4. Faith J, Hayete B, Thaden J, Mogno I, Wierzbowski J, Cottarel G, Kasif S, Collins J, Gardner T: Large-Scale Mapping and Validation of Escherichia coli Transcriptional Regulation from a Compendium of Expression Profiles. PLoS Biology 2007., 5:Google Scholar
  5. Basso K, Margolin A, Stolovitzky G, Klein U, Dalla-Favera R, Califano A: Reverse engineering of regulatory networks in human B cells. Nature Genetics 2005., 37:Google Scholar
  6. Butte AJ, PT , Slonim D, Golub T, Kohane I: Discovering functional relationships between RNA expression and chemotherapeutic susceptibility using relevance networks. Proceedings of the National Academy of Sciences 2000, 97(22):12182–12186. 10.1073/pnas.220392197View ArticleGoogle Scholar
  7. Butte AJ, Kohane IS: Mutual Information Relevance Networks: Functional Genomic Clustering Using Pairwise Entropy Measurments. Pac Symp Biocomput 2000, 418–429.Google Scholar
  8. Margolin AA, Nemenman I, Basso K, Wiggins C, Stolovitzky G, Favera RD, Califano A: ARACNE: an algorithm for the reconstruction of gene regulatory networks in a mammalian cellular context. BMC Bioinformatics 2006., 7:Google Scholar
  9. Meyer PE, Kontos K, Lafitte F, Bontempi G: Information-Theoretic Inference of Large Transcriptional Regulatory Networks. EURASIP J Bioinform Syst Biol 2007, 79879.Google Scholar
  10. Gentleman RIR: R: A language for data analysis and graphics. Journal of Computational and Graphical Statistics 1996., 5: []Google Scholar
  11. Venables WN, Ripley BD: Modern Applied Statistics with S. Fourth edition. Springer; 2002.View ArticleGoogle Scholar
  12. Gentleman RC, Carey VJ, Bates DJ, Bolstad BM, Dettling M, Dudoit S, Ellis B, Gautier L, Ge Y, Gentry J, Hornik K, Hothorn T, Huber W, Iacus S, Irizarry R, Leisch F, Li C, Maechler M, Rossini AJ, Sawitzki G, Smith C, Smyth GK, Tierney L, Yang YH, Zhang J: Bioconductor: Open software development for computational biology and bioinformatics. Genome Biology 2004., 5:Google Scholar
  13. Cheng J, Greiner R, Kelly J, Bell D, Liu W: Learning Bayesian Networks from Data: An Information-Theory Based Approach. Artificial Intelligence 2002., 137:Google Scholar
  14. Chow C, Liu C: Approximating discrete probability distributions with dependence trees. Information Theory, IEEE Transactions on 1968 14:
  15. Pearl J: Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference. Morgan Kaufmann Publishers Inc; 1988.Google Scholar
  16. Cover TM, Thomas JA: Elements of Information Theory. New York: John Wiley; 1990.Google Scholar
  17. Tourassi GD, Frederick ED, Markey MK, C E, Floyd J: Application of the mutual information criterion for feature selection in computer-aided diagnosis. Medical Physics 2001, 28(12):2394–2402. 10.1118/1.1418724View ArticlePubMedGoogle Scholar
  18. Peng H, Long F, Ding C: Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy. IEEE Transactions on Pattern Analysis and Machine Intelligence 2005, 27(8):1226–1238. 10.1109/TPAMI.2005.159View ArticlePubMedGoogle Scholar
  19. Ding C, Peng H: Minimum Redundancy Feature Selection From Microarray Gene Expression Data. Journal of Bioinformatics and Computational Biology 2005, 3(2):185–205. 10.1142/S0219720005001004View ArticlePubMedGoogle Scholar
  20. Merz P, Freisleben B: Greedy and Local Search Heuristics for Unconstrained Binary Quadratic Programming. Journal of Heuristics 2002, 8(2):1381–1231. 10.1023/A:1017912624016View ArticleGoogle Scholar
  21. Olsen C, Meyer PE, Bontempi G: On the Impact of Entropy Estimator in Transcriptional Regulatory Network Inference. In 5th International Workshop on Computational Systems Biology (WSCB 08). Edited by: Ahdesmäki M, Strimmer K, Radde N, Rahnenf hrer J, Klemm K, L hdesm ki H, Yli-Harja O. Tampere International Center for Signal Processing; 2008:41.Google Scholar
  22. Daub CO, Steuer R, Selbig J, Kloska S: Estimating mutual information using B-spline functions – an improved similarity measure for analysing gene expression data. BMC Bioinformatics 2004., 5:Google Scholar
  23. Paninski L: Estimation of entropy and mutual information. Neural Computation 2003, 15(6):1191–1253. 10.1162/089976603321780272View ArticleGoogle Scholar
  24. Beirlant J, Dudewica EJ, Gyofi L, Meulen E: Nonparametric Entropy Estimation: An Overview. Journal of Statistics 97.
  25. Nemenman I, Bialek W, de Ruyter van Steveninck R: Entropy and information in neural spike trains: Progress on the sampling problem. Phys Rev E Stat Nonlin Soft Matter Phys 2004, 69(5 Pt 2):056111.View ArticlePubMedGoogle Scholar
  26. Darbellay G, Vajda I: Estimation of the information by an adaptive partitioning of the observation space. IEEE Transactions on Information Theory 1999.Google Scholar
  27. Hausser J: Improving entropy estimation and inferring genetic regulatory networks. Master's thesis National Institute of Applied Sciences Lyon; 2006. []Google Scholar
  28. Schäfer J, Strimmer K: A shrinkage approach to large-scale covariance matrix estimation and implications for functional genomics. Stat Appl Genet Mol Biol 2005., 4(Article 32):
  29. Wu L, Neskovic P, Reyes E, Festa E, Heindel W: Classifying n-back EEG data using entropy and mutual information features. European Symposium on Artificial Neural Networks 2007.Google Scholar
  30. Beerenwinkel N, Schmidt B, Walter H, Kaiser R, Lengauer T, Hoffmann D, Korn K, Selbig J: Diversity and complexity of HIV-1 drug resistance: A bioinformatics approach to predicting phenotype from genotype. Proc Natl Acad Sci U S A 2002, 99(12):8271–8276. 10.1073/pnas.112177799PubMed CentralView ArticlePubMedGoogle Scholar
  31. Krichevsky R, Trofimov V: The performance of universal coding. IEEE Transactions in Information Theory 1981.Google Scholar
  32. Schurmann T, Grassberger P: Entropy estimation of symbol sequences. Chaos 1996.Google Scholar
  33. Dougherty J, Kohavi R, Sahami M: Supervised and Unsupervised Discretization of Continuous Features. International Conference on Machine Learning 1995, 194–202.Google Scholar
  34. Liu H, Hussain F, Tan CL, Dash M: Discretization: An Enabling Technique. Data Mining and Knowledge Discovery 2002., 6:Google Scholar
  35. Yang Y, Webb GI: On why discretization works for naive-bayes classifiers. Proceedings of the 16th Australian Joint Conference on Artificial Intelligence 2003.Google Scholar
  36. Davis J, Goadrich M: The Relationship Between Precision-Recall and ROC Curves. Proceedings of the 23rd international conference on Machine learning 2006.Google Scholar
  37. Provost F, Fawcett T, Kohavi R: The case against accuracy estimation for comparing induction algorithms. In Proceedings of the Fifteenth International Conference on Machine Learning. Morgan Kaufmann, San Francisco, CA; 1998:445–453.Google Scholar
  38. Bockhorst J, Craven M: Markov Networks for Detecting Overlapping Elements in Sequence Data. In Advances in Neural Information Processing Systems 17. Edited by: Saul LK, Weiss Y, Bottou L. Cambridge, MA: MIT Press; 2005:193–200.Google Scholar
  39. Sokolova M, Japkowicz N, Szpakowicz S: Beyond Accuracy, F-score and ROC: a Family of Discriminant Measures for Performance Evaluation. Proceedings of the AAAI'06 workshop on Evaluation Methods for Machine Learning 2006.Google Scholar
  40. den Bulcke TV, Leemput KV, Naudts B, van Remortel P, Ma H, Verschoren A, Moor BD, Marchal K: SynTReN: a generator of synthetic gene expression data for design and analysis of structure learning algorithms. BMC Bioinformatics 2006, 7: 43. 10.1186/1471-2105-7-43View ArticleGoogle Scholar
  41. Carey VJ, Gentry J, Whalen E, Gentleman R: Network Structures and Algorithms in Bioconductor. Bioinformatics 2005, 21: 135–136. 10.1093/bioinformatics/bth458View ArticlePubMedGoogle Scholar


© Meyer et al; licensee BioMed Central Ltd. 2008

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.