- Methodology article
- Open Access
- Published:

# Detecting multivariate differentially expressed genes

*BMC Bioinformatics***volume 8**, Article number: 150 (2007)

## Abstract

### Background

Gene expression is governed by complex networks, and differences in expression patterns between distinct biological conditions may therefore be complex and multivariate in nature. Yet, current statistical methods for detecting differential expression merely consider the univariate difference in expression level of each gene in isolation, thus potentially neglecting many genes of biological importance.

### Results

We have developed a novel algorithm for detecting multivariate expression patterns, named Recursive Independence Test (RIT). This algorithm generalizes differential expression testing to more complex expression patterns, while still including genes found by the univariate approach. We prove that RIT is consistent and controls error rates for small sample sizes. Simulation studies confirm that RIT offers more power than univariate differential expression analysis when multivariate effects are present. We apply RIT to gene expression data sets from diabetes and cancer studies, revealing several putative disease genes that were not detected by univariate differential expression analysis.

### Conclusion

The proposed RIT algorithm increases the power of gene expression analysis by considering multivariate effects while retaining error rate control, and may be useful when conventional differential expression tests yield few findings.

## Background

The problem of detecting a change in expression between two biological states, such as healthy *vs.* diseased, is central to microarray data analysis. There are two main approaches to this problem: statistical difference tests [1] or feature selection by machine learning techniques [2]. The former alternative provides a solid statistical foundation and allows proper control of false positive rates, but is limited to detecting differences in the expression level of single genes. We herein refer to this as *univariate differential expression* (UDE). The machine learning techniques on the other hand can potentially discover more complex, multivariate effects, herein referred to as *multivariate differential expression* (MDE). Figure 1 provides some examples illustrating the distinction between UDE and MDE. However, machine learning techniques usually aim to discover small, predictive gene sets and do not control error rates. Consequently, the gene lists obtained are often unreliable [3, 4]. Thus, there is a need for novel methods that combine the best of the two worlds – allowing detection of MDE patterns within a sound statistical framework.

Existing solutions to this problem center around the idea of performing statistical tests on *gene sets* rather than on individual genes. Examples of this include the popular Gene Set Enrichment Analysis [5] method and various multivariate tests [6, 7]. This approach permits detection of multivariate effects, but it requires the user to choose the gene sets involved in advance. The latter simplifies the problem considerably, but consequently only permits detection of previously known functional groups such as KEGG pathways [8] or GeneOntology categories [9].

Earlier, Szabo et al. [10] proposed to find novel gene sets by maximizing a test statistic *D* using subset search. Unfortunately, since exhaustive subset search is intractable for high-dimensional data, Szabo et al. are forced to resort to heuristics, considering only gene sets of some *a priori* fixed size and using randomized search methods. Xiao et al. [11] developed this procedure further and used a permutation test to ensure that *D* is significantly larger than what would be expected by random. In this way the error rate *over gene sets* can be controlled, at least approximately. Dettling et al. [12] proposed a different approach for finding gene sets, but only considered sets of size 2 and restricted attention to certain types of bivariate normal distributions.

In our view, a limitation with all of the above methods is that the error rates necessarily concern gene sets, not individual genes. Since a gene set may be significant even when containing only a single true positive gene [10], the false positive rate over individual genes may be much higher than the false positive rate over gene sets. For example, if a single, true gene set containing 10 genes is selected, then the false positive rate over gene sets is zero, while the false positive rate over the genes involved may be as high as 90%. This is appropriate if the investigator is indeed primarily interested in gene sets; but if the primary interest is individual genes, then these methods may be misleading.

In this paper we focus on finding individual MDE genes, not gene sets. We present a principled, generally applicable method which can be seen as a direct generalization of univariate differential expression to the multivariate case. We prove that our procedure is *consistent* (i.e., given enough samples, it discovers exactly the true MDE genes) for a realistic class of data distributions. Further, we show that our method produces correct p-values for small samples, and thus controls error rates while offering more power than univariate differential expression testing. Finally, we apply our method to two microarray data sets and conduct a literature validation of the gene lists generated, revealing many potentially important genes that were not detected by the UDE tests.

### Theoretical background

Throughout, we assume that samples (arrays) *x*^{(1)},... *x*^{(l)}are independent observations of an *n*-dimensional random vector *X* = (*X*_{1},..., *X*_{
n
}), with class-conditional density *f*(*x*|*y*), where *y* ∈ {-1, +1} is a class variable denoting the biological condition. For simplicity we here restrict ourselves to the two-class case, although the theory and methods presented can easily be extended to multiple classes or even continuous *Y*. For microarray data, we typically have *n* ≫ *l*.

We say that a gene *X*_{
i
}is *univariate differentially expressed* (UDE) if the mean expression level for the two conditions differ. Thus we define the set of UDE genes

In figure 1A, this holds for gene *X*_{2} but not for gene *X*_{1}; in figures 1B,C neither gene is UDE. In higher dimensions, similar situations may render a large fraction of the genes MDE yet not UDE. A more general definition is the following: we say that a gene *X*_{
i
}is *multivariate differentially expressed* (MDE) if there exists a gene set *S* ⊂ {*X*_{1},... *X*_{
n
}} such that *X*_{
i
}is conditionally dependent on *Y* given *S*. Thus the set of MDE genes is defined formally as

where $\overline{)\perp}$ denotes conditional dependence. This definition recognizes all genes in figures 1A,B,C as MDE. It was initially proposed by Kohavi and John [13]. Clearly, the criterion (1) implies criterion (2), so we have *U* ⊆ *M*. Therefore MDE is a generalization of UDE. Genes which are MDE but not UDE cannot be detected by univariate differential expression tests, as shown in figure 1. The goal of this paper is to estimate the set *M* from expression data.

The set *M* is typically larger than the set of genes optimal for *predicting Y*, because some genes in *M* may be "redundant" – their predictive information can be obtained from other genes, and hence they can safely be excluded from the predictor [13]. Therefore, machine learning techniques that attempt to optimize a predictor (such as the Recursive Feature Elimination [14] used herein) tend to select only a subset of *M*. Hence, these methods are generally not suitable for our purpose.

Unfortunately, for *arbitrary* data distributions, determining whether equation (2) holds for a given gene *X*_{
i
}requires exhaustive subset search, which is known to be intractable for high-dimensional data. However, for a large class of data distributions we herein refer to as the *PCWT* class (short for Positive/Composition/Weak Transitivity; see Additional file 1 for a rigorous definition), we will prove that the problem is tractable. We will then show that this PCWT class is sufficiently general to be used as a model for biological data, and in particular microarray data.

## Results

### The RIT algorithm

We developed a recursive algorithm named Recursive Independence Test (RIT) based on pairwise tests for marginal independencies. The algorithm pseudocode is given in figure 2A. In the first round, RIT tests for the marginal independencies *X*_{
i
}⊥ *Y* | ∅ for each gene *X*_{
i
}and obtains a gene set *S* of significant findings. Next, for each *X*_{
i
}∈ *S* we recursively call RIT to test for the marginal independencies *X*_{
i
}⊥ *X*_{
j
}| ∅ against each gene *X*_{
j
}∉ *S*, and add the significant findings to *S*. We continue in this fashion until no more dependencies are found.

An illustrating example of an RIT run is given in figure 2B. Here, the MDE genes are *M* = {*X*_{1},..., *X*_{11}}, the UDE genes are *U* = {*X*_{1},..., *X*_{4}}, and the remaining genes are unrelated to *Y*. In the first round of RIT we obtain the set *S*_{1}. In this case *S*_{1} differs from *U*, which of course may happen for small sample sizes since the statistical tests used have limited power. In the next recursion, RIT tests the genes in *S*_{1} against *X* \ *S*_{1} = {*X*_{4},..., *X*_{20}}; this discovers the set *S*_{2}, which is dependent on *X*_{2}. Continuing the recursion, RIT eventually finds two more gene sets *S*_{3}, *S*_{4}, after which no more significant genes are found and the algorithm terminates. The final output of RIT is then the estimate $\widehat{M}$ = *S*_{1} ∪ *S*_{2} ∪ *S*_{3} ∪ *S*_{4}. In *S*_{3} we obtain a false positive *X*_{12}, and since *X*_{4} ∉ *S*_{1}, we also fail to detect *X*_{9} because the required test is never made. Since the RIT algorithm only visits each *X*_{
i
}∈ $\widehat{M}$ once, it is easy to see that the number of tests made is on the order of |$\widehat{M}$|*n*. Thus, for reasonably small $\widehat{M}$, the algorithm scales approximately linearly with the number of genes. This is important not only for computational speed, but also to reduce multiplicity problems (see below). Note also that since the first round of RIT is a univariate differential expression test, the set $\widehat{M}$ found by RIT always includes the genes found by UDE testing. Hence, RIT always has at least as much statistical power as a UDE test.

Typically, one needs to use two different independence tests with RIT, since the class variable *Y* is different from the genes *X*_{
i
}. For simplicity, in our simulations (below) we have used the well-known Student's t-test for *X*_{
i
}⊥ *Y* | ∅ and Fisher's z-transformation for testing *X*_{
i
}⊥ *X*_{
j
}| ∅. The t-test is optimal (unbiased most powerful) for gaussian marginal distributions [15], but on the other hand is correct only for these distributions, which constitute a subset of larger PCWT distribution class (see Additional file 1). Fisher's z is consistent regardless of distribution. While the gaussian assumption is frequently made in gene expression analysis, for example in gene network inference [16], we emphasize that RIT is itself not restricted to this class. As an example, we applied RIT to 100 samples drawn from the non-linear distribution in figure 1B using the distribution-free Kolmogorov-Smirnov test [17] for *X*_{
i
}⊥ *Y* and the Spearman Rank Test [17] for *X*_{
i
}⊥ *X*_{
j
}, both at the 5% level. This produced the correct result (*S*_{1} = {*X*_{2}}, *S*_{2} = {*X*_{1}}) in 99 runs out of 100 (the single error made was a false positive *X*_{2} in *S*_{1}). Applying same tests to the distribution in figure 1C gave no significant findings, since this distribution is not in the PCWT class and therefore not detectable by RIT.

### Consistency of the RIT algorithm

Remarkably, the comparatively simple RIT algorithm can be shown to be *consistent* for any PCWT distribution; that is, as sample size increases, the RIT output $\widehat{M}$ converges to the set of MDE genes *M*. To prove this, note that RIT constructs a path from *Y* to each gene *X*_{
k
}in the graph whose edges (*i*, *j*) correspond to the pairwise marginal dependencies *X*_{
i
}$\overline{)\perp}$*X*_{
j
}| ∅ (i.e., the graph in figure 2B). The following theorem states that for PCWT distributions, the set of genes $\widehat{M}$ reachable through such paths coincides with the set *M* defined in equation (2).

**Theorem 1** *For any PCWT data distribution, the set of MDE genes M is identical to the set of genes* $\widehat{M}$ = {*X*_{
k
}∈ *X*} *for which there exists a sequence* ${Z}_{1}^{m}$ = {*Z*_{1},..., *Z*_{
m
}} ⊆ *X between Z*_{1} = *Y and Z*_{
m
}= *X*_{
k
}*such that Z*_{
i
}$\overline{)\perp}$*Z*_{i+1}| ∅, *i* = 1,..., *m* - 1.

The full proof of this theorem is given in Additional file 1. If the data distribution is not PCWT, the theorem may not hold; figure 1C shows a typical counter-example. Assuming that the independence tests used are consistent, consistency of the RIT algorithm immediately follows from the above. This result is a good argument in favor of RIT, since consistency is a widely accepted *necessary* condition for a sound statistical procedure [15]. To our knowledge, no other algorithm for detecting MDE genes has been proven to be consistent.

### Biological relevance of the PCWT class

Next, we will show that the PCWT class is a reasonable model for gene expression data (or, more generally, for any measurements of biological systems). Since cellular systems are believed to be well described by complex networks [18], it is reasonable to assume that the distribution of *all* variables *X'* comprising a cellular network (transcripts, proteins, metabolites, etc.) can be modelled as a *Bayesian network* [19]. The following theorem, given by [20], asserts that the PCWT class contains all data distributions associated with such networks.

**Theorem 2** *Any strictly positive distribution faithful to a Bayesian network is PCWT*.

However, we typically cannot measure all variables *X'*, but merely a subset *X*; for example, with microarrays we can perhaps measure most transcripts but no proteins or metabolites. Unfortunately, this means that in many cases *X* cannot be modelled by a Bayesian network [21]. Nevertheless, the next theorem asserts that *X* is still PCWT.

**Theorem 3** *Let X' be a random vector with a PCWT distribution and let S, T be any two disjoint subsets of the components of X'. Then with probability 1, the distribution of X* = (*X'* \ {*S*, *T*} | *T* = *t*) *is also PCWT*.

The proof is found in theorems 5 and 6 of [22]. Theorem 3 states that for PCWT distributions, we may fix some variables *T* to constant values *t* and ignore other variables *S*, and the remaining variables will still form a PCWT distribution. Thus, given that the distribution of all variables *X'* comprising the cellular network are PCWT, then any measurements *X* we make will also have a PCWT distribution, even though we fail to measure many variables of the system and perhaps fix others to constant values by experimental design. We therefore conclude that PCWT is a realistic distribution class for biological data.

### Multiplicity and FDR control

Consistency is an asymptotic result however, and is still far from satisfactory for the small sample sizes typical for microarray data. Due to the large amounts of tests made, it is necessary to properly adjust for multiplicity, or else many findings are likely to be false positives. This issue has been thoroughly investigated for univariate tests [1], but our situation is more complicated since RIT performs multiple iterations of testing, and also chooses which tests to make in each iteration depending on the outcome of the previous one.

To ensure multiplicity control, we employ an induction argument. Fix an *α* ∈ [0, 1]. Assume as the induction hypothesis that in the first **foreach** loop of the algorithm (figure 2A) we have tested the null hypotheses ${H}_{0}^{i}$ = *X*_{
i
}⊥ *Y* | ∅ for each *X*_{
i
}and obtained p-values *p*_{
i
}for each *X*_{
i
}. We then sort these to obtain the order statistics *p*_{(1)} ≤ *p*_{(2)} ≤ ... *p*_{(n)}, and apply a correction procedure to choose a gene set *S* (a "top list") with corrected p-values $\tilde{p}$_{
i
}satisfying

This requirement is slightly weaker than FWER control, and is satisfied by the FDR-controlling procedure of Benjamini and Hochberg [23] (see Additional file 1 for a proof), which we employ in this paper. Other FDR-controlling procedures could probably also be used for obtaining *S*, but we have not attempted to prove (3) in the general case.

Now consider the recursive calls RIT(*X* \ *S*, *X*_{
i
}). For each *X*_{
i
}∈ *S*, this will test the null hypotheses ${H}_{0}^{ij}$ = *X*_{
i
}⊥ *X*_{
j
}| ∅ for every *X*_{
j
}∉ *S*, producing the p-values *p*_{
ij
}. We now combine the previously obtained $\tilde{p}$_{
i
}with these *p*_{
ij
}to obtain a single p-value *p*_{
j
}for each *X*_{
j
}∉ *S*. To accomplish this, note that by theorem 1 *X*_{
j
}∉ *M* is possible at this point only if, for *every X*_{
i
}∈ *S*, either ${H}_{0}^{i}$*or* ${H}_{0}^{ij}$ holds true. Hence, the null hypothesis for *X*_{
j
}is

This situation is known in statistics as *intersection-union* testing [24, 25]. By the intersection-union method, a level *α* test for ${H}_{0}^{j}$ is reject

if E_{
i
}E *S* : max {*p*_{
i
}, *p*_{
ij
}} ≤ *α*/|*S*|,

and the corresponding p-value *p*_{
j
}is computed as

The factor |*S*| derives from a Bonferroni correction for the outer intersection in (4). This completes the induction step; as the induction hypothesis is easily satisfied in the first round of testing, it follows by induction that with these corrections, RIT always yields p-values. Finally, the Benjamini-Hochberg procedure may be applied again to control the false discovery rate. Alternatively, more stringent measures such as family-wise error rate control [26] may be used, if desired. Formal proofs of the correctness of each of the above steps can be found in Additional file 1. A detailed pseudocode of RIT implementing each step is given in Additional file 3.

### Simulated data

To illustrate the above result and also to assess the statistical power of RIT as a function of the sample size, we conducted a simulation study. To this end, we designed a distribution with multivariate differential expression, chosen so that 10% of the genes were MDE, but only half of these (5%) were UDE and thus detectable a univariate test (see methods section for details). We compared the performance of RIT against a typical univariate test, namely Student's t-test [15] with FDR correction [23], and also against the popular Recursive Feature Elimination (RFE) feature selection method [14].

Figure 3 summarizes the results of this experiment. We find that RIT does indeed control the FDR at the nominal level (*α* = 0.05), in the same way as the univariate test. The power of the univarate test converges to 0.5 as expected (since only half of the MDE genes were UDE), while the RIT converges to 1.0, in agreement with our theoretical results. Thus, when multivariate effects are present, RIT affords more power than the univariate test at the same FDR level. In contrast, the RFE method clearly does not control the FDR, choosing many genes unrelated to *Y*. RFE also displays low power, most likely because it considers some MDE genes to be "redundant" for prediction and consequently ignores these. Similar behavior is to be expected from other feature selection methods, as explained above. A second simulation study using a different distribution was also performed, with similar results (see Additional file 2). We conclude that it is feasible to apply the RIT algorithm to small-sample data while controlling the FDR at the desired level. Exact sample size requirements cannot be inferred from figure 3 however, as this depends on the data distribution, in particular the fraction of MDE genes and the amount of noise.

### Microarray data

We next tested the RIT algorithm on two publicly available microarray data sets (see methods section). The diabetes data contrasts pancreas islets expression from normal *vs.* type 2 diabetic patients [27]. The original study identified 370 genes as UDE, but this did not account for multiplicity. The q-value method declared only the top three genes of the original study to be significant: *Arnt*, *Cdc14a*, and *Ddx3y*. The RIT algorithm identified an additional 21 transcripts as MDE, of which 5 were unknown EST:s and 16 were known genes according to the NCBI Gene database [28]. We would like to stress that RIT is an hypotheses-generating method, and that the discovered MDE genes may or may not be of *functionally* related to the target variable. Nevertheless, we conducted a literature validation of the 16 known genes (table 1) to search for possible biologically important findings. Five of these (31%) were previously associated with diabetes. Among the remaining 11 novel genes, several give rise to interesting hypotheses: for example, *Dopey1* was recently shown to be active in the vesicle traffic system, the mechanism that delivers insulin receptors to the cell surface. Four genes encoded transcription factors, as do the majority of previously discovered diabetes-associated genes [27]. The *Usp9y* gene discovered by RIT is associated with male infertility and Sertoli cell-only syndrome. Interestingly, so is the UDE *Ddx3Y* gene. This is unlikely to be a coincidence as only 6 human genes were annotated with this function in NCBI Gene. This is an example of general tendency we have observed in our experiments, that additional MDE genes discovered by RIT often are functionally related to the UDE genes, which is hardly surprising given that RIT relies on pairwise independence test. The chloride channels *Clca2* and *Clcn1* are also highly interesting findings, as ion channels in pancreas islets has been shown to regulate insulin secretion [29]. The diabetes-associated potassium ion channel *Kcng1* was also discovered by RIT, strengthening this hypothesis.

The breast cancer data contrasts good-prognosis *vs.* poor-prognosis patients [30]. This set had large amounts of pairwise correlations among genes, resulting in a highly connected dependence graph. To limit the number of findings in this case, we required significant correlations to exceed a threshold 0.85 to be considered by RIT (see discussion). The original study identified a set of 70 cancer-related genes. In addition to these, the RIT algorithm identified 43 MDE genes. Literature validation revealed that 23 of these (53%) had a previously known function in cancer development, whereof 6 (14%) were specifically implicated in breast cancer (table 2). An additional 10 transcripts (23%) were cell cycle-related and may also be considered as potential cancer proliferation genes. Our literature validation thus confirmed 39 (77%) of the genes reported by RIT to be cancer-related. The higher percentage in this case as compared to the diabetes data may reflect the fact that more genes are known for cancer than for diabetes. To assessed the amount of cancer annotations among the 5,000 genes considered, we examined an additional 43 genes chosen at random. Indeed, we found that about 25% of all genes had some cancer or cell cycle-related annotation. Nevertheless, the above fraction of 77% validated genes is highly significant with a Binomial tail p-value < 10^{-20}.

Encouraged by the above result, we set out to investigate the remaining 10 genes that were not previously associated with cancer. We found three histone proteins, which may be related to chromatin remodelling. One novel gene *Gpr116* was recently identified as a G-protein with a potential role in immune response. The novel gene *Prr11* had predicted binding sites for the transcription factor E2F, which in turn is known to be crucial in the control of tumorigenesis. *Ube2s* is an essential component of the ubiquitin-protein conjugation system, which is implicated in various cancer forms. This gene is also functionally related to the known cancer gene *Ube2c*, which also was discovered by RIT. Also interesting were the novel proteins *Depdc1* and *Depdc1b*, both containing RhoGAP domains. This may implicate them in the regulation of various Rho GTPases, which are currently being investigated as cancer-therapy targets [31].

## Discussion

At a first glance, RIT might seem similar to existing algorithms for "local" network inference around "seed genes" [32, 33]. However, network inference is a much harder problem than detecting MDE, and typically requires testing for *conditional* independence (while RIT requires only marginal independence tests). Consequently, these algorithms require substantially larger samples and stronger distribution assumptions, and their time-complexity is exponential with respect to the number of genes found [33].

RIT is less useful for data with large and strongly correlated transcriptional changes, such as the breast cancer data set considered herein. For cancer data, even the (smaller) fraction of UDE genes has been estimated to be on the order of 50% of all genes [34], and the set of MDE genes is presumably much larger. Thus, the concept of MDE is simply not very useful in this case, since most genes turn out to be MDE. Thus, a principled approach for prioritizing among all these genes is urgently needed. For the cancer data, we let RIT prioritize the findings by considering stronger correlations to be more important. This seems reasonable, and we were able to confirm the end results in this case against the literature. However, this problem is ultimately unsolvable by statistical methods, and must instead be addressed by integrating other kinds of information. A possible step towards a principled solution building upon the present work would be to combine the independence tests used here with other data sources and prior beliefs (perhaps in the form of Bayesian probabilities) to guide the RIT algorithm towards more "interesting" genes.

It is important to realize that RIT does not perform feature selection in the usual machine learning-sense: feature selection aims to find the set of features (genes) optimal for constructing an accurate predictor of the target variable, while RIT aims to find the MDE genes, which need not be optimal for prediction.

These are two different problems, and they should be treated separately.

In this study we have limited ourselves to two-class data. However, it is straightforward to extend the RIT algorithm to find multivariate expression patterns with other types of target variables, such as multiple classes data or continuous target variables such as survival times. To accomplish this, only the independence tests used need to be replaced. This "modularity" is a useful property of RIT: to handle different situations, it is sufficient to "plug in" different independence tests. For example, a continuous target variable could be handled by using the Fisher z-transformation also for testing *X*_{
j
}⊥ *Y*. More complex, non-linear independence relations may be handled using nonparametric tests such as the Kolmogorov-Smirnov test [17] or kernel-based tests [35]. However, a basic limitation of the RIT algorithm is that at least one gene must be UDE for any MDE genes to be found. This is an inherent property of the PCWT class.

Dynamic (time-series) data could also be considered, although some additional assumptions may be necessary to ensure PCWT distributions in this case. For example, assuming a Markov condition, time-series data can be modelled using Dynamic Bayesian Networks (DBNs) [19]. The DBN methodology essentially transforms a dynamic model over *n* nodes into an ordinary BN over 2*n* nodes. Thus, DBNs also result in PCWT distributions as described herein (albeit of twice the dimensionality) and RIT is therefore applicable to detecting multivariate changes in dynamic as well as in static data.

## Conclusion

The RIT algorithm is a principled, general approach that increases the power of small-sample, genome-wide expression studies by considering not only univariate differential expression but also multivariate effects. In contrast to previous approaches which focus on testing gene sets [5–7, 10, 11], RIT gives a p-value for each gene and provides control over false positive findings in terms of individual genes. RIT may be very useful in situations where little univariate differential expression is observed, as exemplified by the diabetes data experiment.

## Methods

### Simulation study

In our simulations, we used a multivariate gaussian distribution with *n* = 1, 000 genes and |*M*| = 100 MDE genes, of which |*U*| = 50 were differentially expressed. We first designed a 4-dimensional gaussian distribution with a class-dependent mean vector *μ*_{
y
}= 2*y*·(0, 0, 1, 1) and covariance matrix

equal for both classes. We then constructed the full distribution for the 100 MDE genes using 25 of these 4-blocks. The remaining features had the same covariance matrix but had mean *μ* = (0, 0, 0, 0). We varied the sample size as 10, 20, 30,..., 100.

The Recursive Feature Elimination (RFE) procedure was implemented as described [14], eliminating 20% of the genes in each iteration. We used the radius-margin bound proposed by [36] as a goodness measure for choosing the optimal gene set.

### Microarray data sets

The diabetes data set is from the study by Gunton et al. [27] and is publicly available at the Diabetes Genome Anatomy Project [37]. This data set contrasts human pancreas islets expression from normal (n = 7) *vs.* type 2 diabetic (n = 5) patients. The original data comprises 44,928 probesets from the Affymetrix U133A and B chips. We used only the A chip in our experiments, since we needed to evaluate our results against literature and the A chip contains better annotated sequences. Moreover, since initial analysis using the full A chip resulted in no significant findings, we reduced multiplicity by pre-filtering genes by variance, keeping only the 5,000 most variable genes.

The breast cancer data set consist of 78 samples from patients divided into one "good prognosis" group (*n* = 44) and one a "poor prognosis" group (*n* = 34) based on the time until relapse [30]. The data set is freely available from Rosetta Inpharmatics [38]. The arrays used contains approx. 25,000 transcripts, out of which 4,918 were selected using the same quality filter as in the original publication.

## References

- 1.
Allison DB, Cui X, Page GP, Sabripour M: Microarray data analysis: from disarray to consolidation and consensus. Nat Rev Genet 2006, 7: 55–65. 10.1038/nrg1749

- 2.
Guyon I, Elisseeff A: An introduction to variable and feature selection. Journ Mach Learn Res 2003, 3: 1157–1182. 10.1162/153244303322753616

- 3.
Nilsson R, Peña JM, Björkegren J, Tegnér J: Evaluating feature selection for SVMs in high dimensions. Proceedings of the 17th european conference on machine learning 2006, 719–726.

- 4.
Ein-Dor L, Zuk O, Domany E: Thousands of samples are needed to generate a robust gene list for predicting outcome in cancer. Proc Natl Acad Sci USA 2006, 103(15):5923–5928. 10.1073/pnas.0601231103

- 5.
Subramanian A, Tamayo P, Mootha VK, Mukherjee S, Ebert BL, Gillette MA, Paulovich A, Pomeroy SL, Golub TR, Lander ES, Mesirov JP: Gene set enrichment analysis: A knowledge-based approach for interpreting genome-wide expression profiles. Proc Natl Acad Sci USA 2005, 102(43):15545–15550. 10.1073/pnas.0506580102

- 6.
Kropf S, Läuter J: Multiple Tests for Different Sets of Variables Using a Data-Driven Ordering of Hypotheses, with an Application to Gene Expression Data. Biometrical journal 2002, 44(7):789–800. Publisher Full Text 10.1002/1521-4036(200210)44:7<789::AID-BIMJ789>3.0.CO;2-#

- 7.
Lu Y, Liu PY, Xiao P, Deng HW: Hotelling's T2 multivariate profiling for detecting differential expression in microarrays. Bioinformatics 2005, 21(14):3105–3113. 10.1093/bioinformatics/bti496

- 8.
Ogata H, Goto S, Sato K, Fujibuchi W, Bono H, Kanehisa M: KEGG: Kyoto Encyclopedia of Genes and Genomes. Nucl Acids Res 1999, 27: 29–34. 10.1093/nar/27.1.29

- 9.
Ashburner M, Ball CA, Blake JA, Botstein D, Butler H, Cherry JM, Davis AP, Dolinski K, Dwight SS, Eppig JT, Harris MA, Hill DP, Issel-Tarver L, Kasarskis A, Lewis S, Matese JC, Richardson JE, Ringwald M, Rubin GM, Sherlock G: Gene Ontology: tool for the unification of biology. Nat Genet 2000, 25: 25–29. 10.1038/75556

- 10.
Szabo A, Boucher K, Jones D, Tsodikov AD: Multivariate exploratory tools for microarray data analysis. Biostatistics 2003, 4(4):555–567. 10.1093/biostatistics/4.4.555

- 11.
Xiao Y, Frisina R, Gordon A, Klebanov L, Yakovlev A: Multivariate search for differentially expressed gene combinations. BMC Bioinformatics 2004, 5: 164. 10.1186/1471-2105-5-164

- 12.
Dettling M, Gabrielson E, Parmigiani G: Searching for differentially expressed gene combinations. Genome Biol 2005., 6(R88):

- 13.
Kohavi R, John GH: Wrappers for feature subset selection. Artificial Intelligence 1997, 97: 273–324. 10.1016/S0004-3702(97)00043-X

- 14.
Guyon I, Weston J, Barnhill S, Vapnik V: Gene Selection for Cancer Classification using Support Vector Machines. Machine Learning 2002, 46: 389–422. 10.1023/A:1012487302797

- 15.
Casella G, Berger RL: Statistical Inference. 2nd edition. Duxbury advanced series, Duxbury; 2002.

- 16.
Schäfer J, Strimmer K: An empirical Bayes approach to inferring large-scale gene association networks. Bioinformatics 2005, 21(6):754–764. 10.1093/bioinformatics/bti062

- 17.
Press WH, Flannery BP, Teukolsky SA, Vetterling WT: Numerical recipes in C. 2nd edition. Cambridge University Press; 1992.

- 18.
Kitano H: Computational systems biology. Nature 2002, 420: 206–210. 10.1038/nature01254

- 19.
Friedman N: Inferring Cellular Networks Using Probabilistic Graphical Models. Science 2004, 303(5659):799–805. 10.1126/science.1094068

- 20.
Pearl J: Probabilistic reasoning in intelligent systems. Morgan Kauffman Publishers, Inc., San Fransisco, California; 1988.

- 21.
Chickering D, Meek C: Finding Optimal Bayesian Networks. In Proceedings of the 18th Annual Conference on Uncertainty in Artificial Intelligence. San Francisco, CA: Morgan Kaufmann Publishers; 2002:94–102.

- 22.
Peña JM, Nilsson R, Björkegren J, Tegnér J: Identifying the relevant nodes before learning the structure. Proceedings of the 22nd Conference on Uncertainty in Artificial Intelligence 2006, 367–374.

- 23.
Benjamini Y, Hochberg Y: Controlling the false discovery rate: a practical and powerful approach to multiple testing. J R Statist Soc B 1995, 57: 289–300.

- 24.
Berger RL: Multiparameter hypothesis testing and acceptance sampling. Technometrics 1982, 24(4):295–300. 10.2307/1267823

- 25.
Roy S: On a heuristic method of test construction and its use in multivariate analysis. Ann Math Stat 1953, 24: 220–38.

- 26.
Hochberg Y: A Sharper Bonferroni Procedure for Multiple Tests of Significance. Biometrika 1988, 75(4):800–802. 10.1093/biomet/75.4.800

- 27.
Gunton JE, Kulkarni RN, Yim S, Okada T, Hawthorne WJ, Tseng YH, Roberson RS, Ricordi C, O'Connell PJ, Gonzalez FJ, Kahn CR: Loss of ARNT/HIF1 β Mediates Altered Gene Expression and Pancreatic-Islet Dysfunction in Human Type 2 Diabetes. Cell 2005, 122: 337–349. 10.1016/j.cell.2005.05.027

- 28.
Maglott D, Ostell J, Pruitt KD, Tatusova T: Entrez Gene: gene-centered information at NCBI. Nucl Acids Res 2005, 33: D54-D58. 10.1093/nar/gki031

- 29.
Proks P, Lippiat J: Membrane ion channels and diabetes. Curr Pharm Des 2006, 12(4):485–501. 10.2174/138161206775474431

- 30.
van't Veer LJ, Dai H, Vijver MJVD, He YD, Hart AAM, Mao M, Peterse HL, Kooy KVD, Marton MJ, Witteveen AT, Schreiber GJ, Kerkhoven RM, Roberts C, Linsley PS, Bernards R, Friend SH: Gene expression profiling predicts clinical outcome of breast cancer. Nature 2002, 415: 530–536. 10.1038/415530a

- 31.
Fritz G, Kaina B: Rho GTPases: promising cellular targets for novel anticancer drugs. Curr Cancer Drug Targets 2006, 6: 1–14. 10.2174/156800906775471752

- 32.
Hashimoto RF, Kim S, Shmulevich I, Zhang W, Bittner ML, Dougherty ER: Growing genetic regulatory networks from seed genes. Bioinformatics 2004, 20(8):1241–1247. 10.1093/bioinformatics/bth074

- 33.
Peña JM, Björkegren J, Tegnér J: Growing Bayesian network models of gene networks from seed genes. Bioinformatics 2005, 21(suppl 2):ii224–229. 10.1093/bioinformatics/bti1137

- 34.
Storey JD, Tibshirani R: Statistical significance for genomewide studies. Proc Natl Acad Sci 2003, 100(16):9440–45. 10.1073/pnas.1530509100

- 35.
Gretton A, Herbrich R, Smola A, Bousquet O, Schölkopf B: Kernel Methods for Measuring Independence. Journ Mach Learn Res 2005, 6: 2075–2129.

- 36.
Vapnik VN: Statistical Learning Theory. John Wiley and Sons, Inc; 1998.

- 37.
The Diabetes Genome Anatomy Project[http://www.diabetesgenome.org]

- 38.
Rosetta Inpharmatics[http://www.rii.com/publications/2002/vantveer.html]

## Acknowledgements

This work was supported by grants from the Ph.D. Programme in Medical Bioinformatics, the Swedish Research Council (VR-621-2005-4202), Clinical Gene Networks AB and Linköping University.

## Author information

## Additional information

### Authors' contributions

RN devised the multiple comparison procedure, performed the simulation study and the application to the microarray data sets, performed most of the literature validation, and wrote most of the manuscript. JMP devised the basic idea for the RIT algorithm and the wrote the proof of theorem 1. JB assisted with the literature validation. JT participated in study design and co-wrote the manuscript.

## Electronic supplementary material

## Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

## Rights and permissions

## About this article

#### Received

#### Accepted

#### Published

#### DOI

### Keywords

- Bayesian Network
- Network Inference
- Independence Test
- Multivariate Effect
- Breast Cancer Data