- Methodology article
- Open Access

# Assessing stability of gene selection in microarray data analysis

- Xing Qiu
^{1}, - Yuanhui Xiao
^{1}, - Alexander Gordon
^{1}and - Andrei Yakovlev
^{1}Email author

**7**:50

https://doi.org/10.1186/1471-2105-7-50

© Qiu et al; licensee BioMed Central Ltd. 2006

**Received: **17 August 2005

**Accepted: **01 February 2006

**Published: **01 February 2006

## Abstract

### Background

The number of genes declared differentially expressed is a random variable and its variability can be assessed by resampling techniques. Another important stability indicator is the frequency with which a given gene is selected across subsamples. We have conducted studies to assess stability and some other properties of several gene selection procedures with biological and simulated data.

### Results

Using resampling techniques we have found that some genes are selected much less frequently (across sub-samples) than other genes with the same adjusted *p*-values. The extent to which this type of instability manifests itself can be assessed by a method introduced in this paper. The effect of correlation between gene expression levels on the performance of multiple testing procedures is studied by computer simulations.

### Conclusion

Resampling represents a tool for reducing the set of initially selected genes to those with a sufficiently high selection frequency. Using resampling techniques it is also possible to assess variability of different performance indicators. Stability properties of several multiple testing procedures are described at length in the present paper.

## Keywords

## Background

The result of every analysis of microarray data is an outcome of a random experiment. For example, the number of genes declared differentially expressed and the estimated false discovery rate (FDR) should be treated as random variables and their variability has to be assessed in the same fashion that the population variance is estimated in the usual statistical inference. The variance of the number of differentially expressed genes (as well as other outcomes of a given selection procedure) may depend on the chosen statistical test, method of multiple testing adjustment, effect sizes for different genes, and the correlation structure of the data. The latter factor deserves special attention. Although some normalization procedures may lead to a significant reduction in the correlation between gene expression levels, and thus between the associated test statistics, the remaining correlation may be strong enough to have a disastrous effect on the statistical inference from microarray data [1]. The effect correlations in microarray data on variability of the most basic performance indicators of various testing procedures calls for further investigation.

There is another facet of the problem to consider. Every specific analysis of microarray data results in a list of candidate genes that are deemed differentially expressed across the two conditions under study. The composition of this list is subject to random fluctuations and this effect also needs to be quantitatively assessed. Even if one concentrates on the selection of individual genes rather than gene combinations, the situation here is similar to that in the regression analysis aimed at selecting significant explanatory variables (covariates). When focusing on a specific variable, one can observe a certain degree of instability of this variable selection inherent in any pertinent statistical procedure. The term "stability" means "replication stability" for the selection of significant variables. This kind of stability is easy to assess and interpret in simulation studies where the "true" set of differentially expressed genes is known. When analyzing biological data, one can resort to resampling techniques for this purpose [2]. In particular, one can apply a subsampling counterpart of the "delete-*d*-jackknife" procedure to the sample at hand and estimate the frequency with which a given gene has been selected across all sub-samples. Then an additional selection criterion can be imposed by finally selecting only those genes with a frequency of selection greater than, say, 80%.

The above discussion suggests the following two ways of using resampling techniques in microarray data analysis. These techniques can be used to assess stability characteristics of a given selection procedure and compare different procedures. In this case, one is usually interested in certain characteristics of the whole set of selected genes rather than its individual members. In their very interesting paper, Pavlidis et al. [3] used leave-one-out resampling to study the stability of gene selection in conjunction of the required number of replicates in the analysis of differential expression of genes. The authors proposed two stability measures (metrics) to compare the ranked list of the genes originally selected to the ranking obtained when one replicate is removed. Then the stability measures are averaged over the subsamples. The first measure refers to the fraction of genes among the originally selected ones that are recovered in a given subsample. Stolovitzky [4] proposed a similar measure which is not conditioned on the set of genes originally selected from the data prior to resampling. However, statistical properties of the robustness index introduced by Stolovitzky remain unclear. The second measure by Pavlidis et al. [3] is more subtle; it has to do with the degree to which the ordering is preserved and can be used whenever the number of selected genes does not show strong variations among subsamples. We propose the delete-*d*-jackknife variance of the number of selected genes (which is the primary endpoint to be assessed when comparing different methodologies) across subsamples as a pertinent measure of stability of a chosen testing procedure. This measure has clear statistical properties and is easier to interpret. The distribution of the number of selected genes can also be estimated using the delete-*d*-jackknife method [5]. Another way of using resampling techniques is to assess the stability of selection for individual genes in line with the currently practiced methodology of significance testing in microarray analysis. This can be accomplished by estimating the frequency of selection of each gene given it has been selected at least in one subsample. As we show in the present paper, this measure also provides valuable information on the performance of each selection procedure when its dependence on adjusted *p*-values is included in the analysis.

We have conducted a simulation study to evaluate the effect of correlation between gene expression levels on the performance of several selection procedures in terms of the variability of such important indicators as the number of selected genes and the proportion of falsely rejected among all rejected null hypotheses. All these indicators are directly accessible in computer simulations, thereby providing an explanatory insight into the performance of different procedures. From this perspective, the Bonferroni and Westfall-Young multiple testing procedures are explored in conjunction with the Student *t*, Kolmogorov-Smirnov, and Cramér-von Mises two-sample tests. The latter two tests are distribution free. The Bonferroni and Westfall-Young step-down procedures [6] are designed to control the familywise error rate (FWER). The FWER is defined loosely as the probability of at least one Type 1 error in the context of multiple testing. The FDR-based procedures are also explored; these are represented by the empirical Bayes method [7–9] as well as the Benjamini-Hochberg and Benjamini-Yakutieli procedures [10, 11]. The FDR is defined as the expected fraction of falsely rejected among all rejected hypotheses. It should be noted that our simulation studies do not attempt to model the actual correlation structure of microarray data; their only purpose is to see which specific performance indicators may be sensitive to the presence of correlation in the data. The quantitative characteristics we report from the simulated data cannot be extrapolated to biological data and can only be viewed as proof of principle.

Another set of experiments was concerned with actual biological data. We assessed probabilistic characteristics of the number of selected genes by resampling from a large set of data on two types of childhood leukemia available from the St. Jude Children's Research Hospital Database [12]. Using this data set, we also assessed the replication stability of gene selection and its dependence on adjusted *p*-values.

## Methods

### Biological data

For the purposes of this study, use was made of the St. Jude Children's Research Hospital (SJCRH) Database on childhood leukemia which is publicly available on their website under the Supplemental Data section: [12] The whole SJCRH Database contains gene expression data on 335 subjects, each represented by a separate array (Affymetrix, Santa Clara, CA) reporting measurements on the same set of 12558 genes. We selected two groups of patients with hyperdiploid (Hyperdip) and T-cell acute lymphoblastic leukemia (TALL), respectively. The groups were balanced to include 43 patients in each group. Since the nature of our study was purely methodological, the choice of the data set was quite arbitrary; it was dictated solely by sample size considerations. The microarray data thus chosen were background corrected and normalized using the Bioconductor RMA software. This software implements the quantile normalization procedure [13, 14] carried out at the probe feature level. After the normalization, each gene is represented in the final data set by the logarithm (base 2) of its expression level.

### Simulated data

Our simulation study was designed to illustrate the effect of correlation on the performance of gene selection procedures. We simulated 2*n* independent multi-variate normal random vectors with exchangeable correlation structure, each representing log-intensities of 1255 genes of which the first 125 genes were designated to be differentially expressed. Two sets of simulations were conducted with the sample size chosen to be *n* = 15 and *n* = 43, respectively. We use the following self-explanatory notation for the four sets of simulated data: SIM15, SIM15CORR, SIM43, SIM43CORR. In total, 200 independent data sets, each consisting of 2*n* simulated vectors, were generated for each sample size. The marginal distributions of the log-intensities of "Not Different" genes were standard normal, while the log-intensities of "Different" genes expressions followed the normal distribution with mean two and unit variance.

The exchangeable pairwise correlation structure was superimposed on the normal vectors with independent components as discussed in [1]. Briefly, we first generate a 1255 × 2*n* matrix with each entry being an independent realization of a standard normal random variable. To model a set of "Different" genes, we add a value of 2 to the first 125 rows in the first group and denote the resultant matrix by *X* = {*x*_{
ij
}}, *i* = 1, ..., 1255; *j* = 1, ..., 2*n*. All the elements *xij* of this matrix are stochastically independent, but those with *i* = 1, 2, ..., 125 and *j* = 1, 2, ..., *n* are normally distributed with mean 2 and unit variance. Expression levels of the genes outside this special set of 125 genes follow the standard normal distribution. Next we generate a 2*n*-dimensional random vector with independent and identically distributed components, each component having a standard normal distribution. Denote this vector by *A* = {*a*_{
j
}}, *j* = 1, ..., 2*n*. Define ${y}_{ij}=\sqrt{\rho}{a}_{j}+\sqrt{1-\rho}{x}_{ij}$, *i* = 1, ..., 1255; *j* = 1, ..., 2*n*, so that for any *i*_{1} ≠ *i*_{2} and *j* we have corr (${y}_{{i}_{1}j}$, ${y}_{{i}_{2}j}$) = *ρ* In the present study, the correlated data were generated for a single value of the correlation coefficient *ρ* = 0.6. This high correlation coeffcient was chosen to more clearly demonstrate the effects of correlation. However, this value is not overly unrealistic because the mean (over all gene pairs) correlation coefficient estimated from the raw data referred to in Section 2.1 is equal to 0.72.

In order to see whether or not the stability of gene selection is related to the power, our explanatory simulations were conducted under two different scenarios. Under the first scenario, the sample size was small (*n* = 15) so that the power was lower than 100%. Under the second scenario the sample size was sufficiently large (*n* = 43) to attain a 100% power.

### Resampling techniques

When analyzing biological data, we used a subsampling version of the delete-*d* jackknife method [5, 15, 16], which is technically equivalent to the leave-*d*-out cross-validation. It can be proven that if $\sqrt{n}$/*d* → 0 and *n* - *d* → ∞, then the delete-*d*-jackknife is consistent for the median [15]. Therefore, the general recommendation is to leave out more than *d* = $\sqrt{n}$ but much fewer that *n* observations. A similar recommendation holds for the variance [15, 16]. We used *d* = 7 to perturb the data set, which is only slightly greater than $\sqrt{n}$, to be as close as possible to the most widely used delete-one version of jackknife. It should be noted that the delete-1-jackknife method may be inconsistent for some estimators (sample quantiles representing a typical example) and the delete-*d*-jackknife was proposed to remedy this problem [16]. When implemeting the delete-*d*-jackknife method, we resorted to sampling without replacement because the empirical Bayes method is very sensitive to ties. As far as subsampling versions of the delete-*d*-jackknife method are concerned, the schemes with and without replacement are essentially identical when the number of subsamples is large [16].

The total number of subsamples was typically equal to 200. In a separate study, we ascertained that the results for 1000 subsamples were largely similar. Let *Z* be the number of selected genes. The variance of *Z* is estimated by a resampling counterpart of the jackknife sample variance [16]:

$V=\frac{n-d}{dB}{{\displaystyle \sum _{l=1}^{B}\left({Z}_{n-d,l}-\frac{1}{B}{\displaystyle \sum _{k=1}^{B}{Z}_{n-d,k}}\right)}}^{2},$

where *B* is the total number of subsamples (*B* = 200), *Z*_{n-d, j}is the statistic *Z* evaluated at the *j* th delete-*d* jackknife subsample. The variance of the number of selected genes was used as a criterion of stability of the testing procedures under study. The corresponding distributions were also estimated. Another criterion was the selection stability for each individual gene measured by the frequency of selection conditional on the event of selection in at least one of the subsamples.

### Selection of differentially expressed genes

When resorting to the Bonferroni adjustment, one needs to compute unadjusted *p*-values from the sampling distribution of the test statistic under consideration. For the *t*-test we used quantiles of the Student distribution. Among the distribution-free methods, the Cramér-von Mises test [17] represents an appealing alternative to the Kolmogorov-Smirnov test. The reason is that the granularity of the Cramér-von Mises statistic (which causes granularity of the corresponding *p*-values) is much smaller than that for the Kolmogorov-Smirnov test. As a result, the *p*-values corresponding to the critical region increase much more steeply for the Kolmogorov-Smirnov test than for the Cramér-von Mises test, thereby making the Kolmogorov-Smirnov test less powerful.

To describe the Cramér-von Mises test, consider two independent samples *x*_{1}, *x*_{2}, ..., *x*_{
m
}and *y*_{1}, *y*_{2}, ..., *y*_{
n
}from distributions *F*(*x*) and *G*(*x*), respectively, and let *F*_{
m
}and *G*_{
n
}be their respective empirical distribution functions. We wish to test the following null hypothesis **H**_{0}:*F*(*x*) = *G*(*x*) for all *x* versus the alternative: *F* ≠ *G*. The Cramér-von Mises test-statistic for the hypothesis **H**_{0} is defined as the (squared) *L*_{2} distance between *F*_{
m
}(*x*) and *G*_{
n
}(*x*):

${W}^{2}=\frac{mn}{{(m+n)}^{2}}\left\{{\displaystyle \sum _{i=1}^{m}{[{F}_{m}({x}_{i})-{G}_{n}({x}_{i})]}^{2}+{\displaystyle \sum _{j=1}^{n}{[{F}_{m}({y}_{j})-{G}_{n}({y}_{j})]}^{2}}}\right\}.$

The asymptotic theory for the Cramér-von Mises test was developed by Anderson and Darling [18], Rosenblatt [19], Anderson [20], and Csorgo and Faraway [21]. The asymptotic approximation of the distribution of *W*^{2} under **H**_{0} is of little utility in microarray data analysis because it is very inaccurate whenever one works with extremely small *p*-values required by the FWER controlling multiple testing procedures. For example, when *n* = *m* = 43 and the exact *p*-value for the Cramér-von Mises test is equal to 3.928 × 10^{-6}, its asymptotic approximation gives 6.897 × 10^{-6}, a much larger *p*-value. The above-mentioned exact *p*-value of 3.928 × 10^{-6}corresponds to the adjusted (by the Bonferroni method) *p*-value of about 0.049 when testing 12558 hypotheses as in our application described in Section 3.1. Therefore, one needs an algorithm for computing exact quantiles of the Cramér-von Mises sampling distribution. We used the method proposed by Burr [22] for this purpose. It should be noted that the needed small *p*-values for the Cramér-von Mises test cannot be estimated with sufficient accuracy by permuting the test-statistics, because the required number of permuations is astronomical [23] and cannot be accomplished with present-day hardware.

The Westfall-Young step-down algorithm [6] bypasses the stage of computing unadjusted *p*-values and goes directly to the estimation of adjusted *p*-values at a given level of the FWER. We carried out 10,000 permutations to model a null distribution of each test statistic. We also used the multiple testing adjustment proposed by Benjamini and Hochberg [10] and its modification by Benjamini and Yekutieli [11]. The more conservative Benjamini and Yekutieli procedure is warranted with normalized data because the quantile normalization is known to induce negative correlations in microarray data [1]. The nonparametric empirical Bayes method by Efron et al. [7–9] was one more method of choice in the present paper. We used kernel smoothing (with the Gaussian kernel) for density estimation to implement the empirical Bayes method. The threshold level of the posterior probability was set at 0.95.

To distinguish between different statistical procedures, we use the following notation:

B/t – *t*-test with Bonferroni adjustment;

B/KS – Kolmogorov-Smirnov test with Bonferroni adjustment;

B/CVM – Cramér-von Mises test with Bonferroni adjustment;

WY/t – *t*-test with Westfall-Young algorithm;

WY/KS – Kolmogorov-Smirnov test with Westfall-Young algorithm;

WY/CVM – Cramér-von Mises with Westfall-Young algorithm;

BH/t – *t*-test with Benjamini-Hochberg adjustment;

BY/t – *t*-test with Benjamini-Yekutieli adjustment;

EB/t – *t*-test with gene selection by nonparametric empirical Bayes method.

### False discovery rate and power

We provide estimates of the FDR only for simulations. We do not report FDR estimates for biological data because only indirect methods [24, 25] are available in this case. Such methods introduce an additional variation in the estimates which is impossible to distinguish from that caused by a given selection procedure. In our simulation studies, the true FDR was estimated directly as the proportion of false discoveries among all discoveries. Then the sample mean (across the 200 samples) of this nonparametric estimate is reported together with the corresponding standard deviation. It happened only once (when applying the Kolmogorov-Smirnov test with Bonferroni adjustment to a sample of size *n* = 15) that we set the estimated FDR at zero (see [26] for the definition of the positive FDR). Since the expression levels of the 125 differentially expressed genes are identically distributed, the power can be defined as the expected proportion of correct discoveries among the 125 true alternative hypotheses. We provide the usual nonparametric estimates of the power thus defined and its standard deviation.

### Software

The relevant software is included in the Additional Material Files [see additional file 1].

## Results

### Analysis of biological data

Delete-*d*-jackknife subsampling for the biological data with *d* = 7.

Method | Leave-seven-out Jackknife | ||
---|---|---|---|

Mean number of selected genes | Standard deviation | Mean number of stable genes and its proportion to the mean of selected genes | |

B/KS | 622 | 71 | 504(80.99%) |

B/CVM | 1096 | 123 | 853(77.80%) |

B/t | 775 | 103 | 644(83.05%) |

WY/KS | 685 | 153 | 533(77.82%) |

WY/CVM | 889 | 124 | 711(79.98%) |

WY/t | 876 | 110 | 726(82.89%) |

EB/t | 1867 | 438 | 1481(79%) |

BH/t | 2726 | 445 | 2176(80%) |

BY/t | 1599 | 222 | 1282(80%) |

*U*-shaped so that one can distinguish two extreme groups of genes characterized by high and low stability, respectively. The proportions of genes in each "intermediate-frequency" category are relatively small. This phenomenon persists for all the statistical tests under study when the FWER is controlled either by the Bonferroni adjustment or by the Westfall-Young permutation algorithm. It is clear from Figure 1 that the population of genes selected at least once across all subsamples is heterogeneous with respect to their stability characterized by the frequency of selection.

*p*-values. To this end, we produced scatter-plots for the frequency of occurrence in the set of selected genes across the sub-samples and the original adjusted

*p*-values determined by the application of each testing procedure to the whole set of arrays. The results for the

*t*-test with Bonferroni adjustment are given in Figure 2. The leave-seven-out resampling reveals a non-linear (but still monotonic) pattern showing that the relationship in question may be quite complex. For comparison, we also present the result for the leave-one-out procedure, in which case the dependence appears to be almost linear but the scatter of points is wide because this procedure does not perturb the data sufficiently. In what follows, we will discuss only the observations resulted from the delete-7-jackknife subsampling.

*t*-test and the Cramér-von Mises test with Bonferroni adjustment are compared in Figure 3. It is clear that the genes selected by the Cramér-von Mises test are uniformly more stable than those selected by the

*t*-test. The difference is much less pronounced with the Westfall-Young algorithm as evidenced by Figure 4. Both multiple testing procedures yield similar scatter plots for the

*t*-test showing its overall poor stability in comparison to the Cramér-von Mises test (Figure 5). In contrast, the stability of the Cramér-von Mises test can be increased substantially when using the more conservative Bonferroni adjustment in place of the Westfall-Young procedure (Figure 6). These results show that the stability of gene selection provides an important additional information on each selected gene and this information can be extracted from real data by resorting to resampling techniques.

### Analysis of simulated data

*n*=

*m*= 15. Since the simulated data are normally distributed it comes as no surprise that the

*t*-test proves itself as the most powerful one among those under study. With this small sample size, however, even the

*t*-test tends to be underpowered when used in combination with the Bonferroni adjustment or Westfall-Young adjustments. The power of the

*t*-test is much higher with the Benjamini-Hochberg and nonparametric empirical Bayes procedures. The variance of the estimated power as well as the number of selected genes increases dramatically with increasing correlation between gene expression signals.

Simulating the basic characteristics of gene selection procedures, 125 differentially expressed genes, 200 simulation runs, *n* = 15. The table presents mean values over simulation runs. Standard deviations are given in parentheses.

Method | Number of Selected Genes | FDR | Power | |||
---|---|---|---|---|---|---|

S15 | S15COR | S15 | S15COR | S15 | S15COR | |

B/KS | 36(5.3) | 34(18.8) | <0.0006 | <0.0001 | 0.28(0.04) | 0.27(0.15) |

B/CVM | 80(4.9) | 80(25.0) | <0.0008 | <0.0004 | 0.64(0.04) | 0.64(0.20) |

B/t | 89(5.5) | 88(24.7) | <0.0007 | <0.0008 | 0.71(0.04) | 0.70(0.20) |

WY/KS | 36(4.7) | 53(26.3) | 0 | <0.0008 | 0.29(0.04) | 0.43(0.21) |

WY/CVM | 81(5.6) | 90(21.25) | <0.0003 | <0.0008 | 0.65(0.04) | 0.72(0.17) |

WY/t | 90(5.4) | 98(19.66) | <0.0007 | 0.0009 | 0.72(0.04) | 0.79(0.16) |

BH/t | 130(2.8) | 139(73.5) | 0.048(0.019) | 0.051(0.135) | 0.99(0.01) | 0.99(0.03) |

EB/t | 116(3.0) | 141(100.0) | 0.012(0.006) | 0.052(0.157) | 0.92(0.02) | 0.96(0.07) |

*n*=

*m*= 43). In this case, all the methods attain 100% power. For all the FWER controlling procedures, the mean number of selected genes is exactly 125 and the corresponding variance is quite small irrespective of the presence or absence of correlation between gene expression levels. The FDR estimates are also uniformly small for such procedures as indicated by Table 3. However, the effect of correlation on the standard deviation of the number of selected genes is still very strong (compare with Table 2) for the Benjamini-Hochberg and nonparametric empirical Bayes procedures, indicating the inherent instability of these procedures. It should be noted that there is also a dramatic effect of the correlation on the standard deviation of the FDR observed for the latter procedures (Table 3). The results for 1000 simulation runs were largely similar.

Simulating the basic characteristics of gene selection procedures, 125 differentially expressed genes, 200 simulation runs, *n* = 43. The table presents mean values over simulation runs. Standard deviations are given in parentheses.

Method | Number of Selected Genes | FDR | Power | |||
---|---|---|---|---|---|---|

SIM43 | SIM43CORR | SIM43 | SIM43CORR | SIM43 | SIM43CORR | |

B/KS | 125(0.3) | 125(0.5) | <0.0003 | <0.0001 | 1 | 1 |

B/CVM | 125(0.3) | 125(0.3) | <0.0005 | <0.0005 | 1 | 1 |

B/t | 125(0.2) | 125(0.4) | <0.0003 | <0.0006 | 1 | 1 |

WY/KS | 125(0.4) | 125(0.4) | <0.0002 | <0.0003 | 1 | 1 |

WY/CVM | 125(0.3) | 125(0.4) | <0.0006 | <0.0010 | 1 | 1 |

WY/t | 125(0.2) | 125(0.3) | <0.0003 | <0.0008 | 1 | 1 |

BH/t | 131(2.7) | 140(90.0) | 0.0427(0.0192) | 0.0356(0.1219) | 1 | 1 |

EB/t | 125(0.2) | 133(60.0) | 0.0082(0.0012) | 0.0246(0.1024) | 1 | 1 |

We also include the histograms for the number of selected genes resulted from our simulation studies [see additional file 2]. Note that the high variance observed for the BH/t and EB/t procedures (Figures 2 and 4 in the Additional File 2) is mainly attributable to outliers.

## Discussion

Numerous publications have considered the utility of multiple testing procedures in the context of microarray data analysis (see [27] for a review). However, little attention has been given to the replication stability of such procedures which is related to the reproducibility of scientific results. Our study shows that the variance of the number of the genes declared differentially expressed can be very high for multiple testing procedures even with reasonably large sample sizes. Whenever this is the case, the stability of membership in the list of candidate genes should be expected to be low. However, the reverse is not true. If the variance of the total number of selected genes is low, there still can be tangible variations in the stability of selection for individual genes, thereby affecting the composition of the resultant list of candidate genes. This obviously can have a strong effect on the ranking of candidate genes based on purely statistical criteria.

The present study demonstrates that the proportion of highly stable (with frequencies of more than 80%) genes appears to be almost the same for all the selection procedures under study. At the same time, the overall stability of gene selection varies among different methods. The Cramér-von Mises seems to be superior to other methods in this respect. It is difficult to control the stability of gene selection by an additional adjustment of *p*-values. Indeed, for the FWER-controlling procedures, the relationship between the original (adjusted for multiple testing) *p*-values and the selection frequency appears to be non-linear. However, resampling techniques represent a universal tool for assessing the stability in question with the data at hand. As was emphasized in Section 1, our simulation studies were designed to demonstrate the fact that the correlation between gene expression levels can affect the results of testing two-sample marginal hypotheses. The FDR-controlling procedures appear to be especially sensitive to this effect. Our recent study [28] pinpoints specific components of the empirical Bayes methodology where this effect manifests itself. The quantitative contribution of the correlation between gene expression levels to the outcomes of microarray data analysis is diffcult to estimate because no tools are available to model the actual correlation structure of such a large number of variables in computer simulations.

Tables 2 and 3 also illustrate the importance of sample size. However, the number of genes selected by the Benjamini-Hochberg and nonparametric empirical Bayes procedures is very sensitive to correlations even when the power of these methods reaches 100%. The variance of the true FDR is also quite high for such procedures. Our simulations show that the FWER-controlling procedures are more stable to the effect of correlation and this stability increases with increasing sample size. Pavlidis et al. [3] proposed to approach the sample size problem from the stability perspective and we find their idea very promising and deserving of further exploration.

The distribution-free methods are generally more stable than the *t*-test. It is our firm belief that such methods will play an increasingly important role and gradually replace the *t*-test in microarray studies. Robust versions of two-sample tests in general and of the *t*-test [29] in particular can be quite competitive with distribution-free methods [30] and this avenue invites a special investigation.

## Conclusion

As larger sets of microarray gene expression data become more readily available, the stability of gene selection is becoming easier to assess using resampling techniques. We have found that some genes are selected much less frequently (across subsamples) than other genes with the same adjusted *p*-values. The relationship between the stability of gene selection and the original (adjusted) *p*-values may be rather complex but resampling techniques can advantageously be used to select the most stable genes. Using these techniques, it is also possible to assess variability of the number of selected genes. In reference to the latter indicator, all the selection procedures studied in the present paper appear to be highly unstable. For the FWER-controlling procedures, this property correlates well with the level of random fluctuations in the estimated power of a given procedure. The more conservative FWER-controlling procedures appear to be more stable to the effect of correlation than the FDR-based procedures. The stability characteristics discussed in this paper provide an additional information that should be utilized in gene selection procedures. We suggest that resampling techniques be routinely used for selection of individual genes whenever the sample size is not prohibitively small.

## Declarations

### 7 Acknowledgements

We are grateful to the three anonymous reviewers for their insightful comments. This research is supported by NIH grant GM075299 (Yakovlev).

## Authors’ Affiliations

## References

- Qiu X, Brooks AI, Klebanov L, Yakovlev A:
**The effects of normalization on the correlation structure of microarray data.***BMC Bioinformatics*2005,**6:**120. 10.1186/1471-2105-6-120PubMed CentralView ArticlePubMedGoogle Scholar - Sauerbrei W, Schumacher M:
**A bootstrapping resampling procedure for model building: application to the Cox regression model.***Statistics in Medicine*1993,**11:**2093–2109.View ArticleGoogle Scholar - Pavlidis P, Li Q, Noble WS:
**The effect of replication on gene expression microarray experiments.***Bioinformatics*2003,**19:**1620–1627. 10.1093/bioinformatics/btg227View ArticlePubMedGoogle Scholar - Stolovitzky G:
**Gene selection in microarray data: the elephant, the blind men and our algorithms.***Current Opinion in Structural Biology*2003,**13:**370–376. 10.1016/S0959-440X(03)00078-2View ArticlePubMedGoogle Scholar - Politis DN, Romano JP:
**Large sample confidence regions based on subsamples under minimal assumptions.***The Annals of Statistics*1994,**22:**2031–2050.View ArticleGoogle Scholar - Westfall PH, Young S:
*Resampling-Based Multiple Testing.*Wiley, New York; 1993.Google Scholar - Efron B, Tibshirani R, Storey JD, Tusher V:
**Empirical Bayes analysis of a microarray experiment.***J Amer Statist Assoc*2001,**96:**1151–1160. 10.1198/016214501753382129View ArticleGoogle Scholar - Efron B:
**Robbins, empirical Bayes and microarrays.***Ann Statist*2003,**31:**366–378. 10.1214/aos/1051027871View ArticleGoogle Scholar - Efron B:
**Large-scale simultaneous hypothesis testing: The choice of a null hypothesis.***J Amer Statist Assoc*2004,**99:**96–104.View ArticleGoogle Scholar - Benjamini Y, Hochberg Y:
**Controlling the false discovery rate: A practical and powerful approach to multiple testing.***J Roy Statist Soc Ser B*1995,**57:**289–300.Google Scholar - Benjamini Y, Yekutieli D:
**The control of the false discovery rate in multiple testing under dependency.***Ann Statist*2001,**29:**1165–1188. 10.1214/aos/1013699998View ArticleGoogle Scholar **St. Jude Children's Research Hospital (SJCRH) Database on childhood leukemia**[http://www.stjuderesearch.org/data/ALL1/]- Bolstad BM, Irizarry RA, Astrand M, Speed TP:
**A comparison of normalization methods for high density oligonucleotide array data based on variance and bias.***Bioinformatics*2003,**19**(2):185–193. 10.1093/bioinformatics/19.2.185View ArticlePubMedGoogle Scholar - Irizarry RA, Gautier L, Cope LM:
**An R package for analyses of Affymetrix oligonucleotide arrays.**In*The Analysis of Gene Expression Data*. Edited by: Parmigiani G, Garrett ES, Irizarry RA, Zeger SL. Springer, New York; 2003:102–119.View ArticleGoogle Scholar - Efron B, Tibshirani R:
*An Introduction to the Bootstrap*. Chapman & Hall/CRC, New York; 1993.View ArticleGoogle Scholar - Shao J, Tu D:
*The Jackknife and Bootstrap*. Springer Series in Statistics, Springer, New York; 1995.View ArticleGoogle Scholar - Conover WJ:
*Practical Nonparametric Statistics*. 3rd edition. Wiley, New York; 1999.Google Scholar - Anderson TW, Darling DA:
**Asymptotic theory of certain "goodness of fit" criterion based on stochastic processes.***The Annals of Mathematical Statistics*1952,**23:**193–212.View ArticleGoogle Scholar - Rosenblatt M:
**Limit theorems associated with variants of the von Mises statistic.***The Annals of Mathematical Statistics*1952,**23:**617–623.View ArticleGoogle Scholar - Anderson TW:
**(1962) On the distribution of the two-sample Cramér-von Mises criterion.***The Annals of Mathematical Statistics*1962,**33:**1148–1159.View ArticleGoogle Scholar - Csorgo S, Faraway JJ:
**(1996) The exact and asymptotic distributions of Cramée-von Mises statistics.***Journal of the Royal Statistical Society. Series B (Methodological)*1996,**58:**221–234.Google Scholar - Burr EJ:
**Small-sample distribution of the two-sample Cramér-von Mises criterion for small equal samples.***The Annals of Mathematical Statistics*1963,**34:**95–101.View ArticleGoogle Scholar - Klebanov L, Gordon A, Xiao Y, Land H, Yakovlev A: A permutation test motivated by microarray data analysis. Comp Stat Data Anal, in press.Google Scholar
- Storey JD, Tibshirani R:
**Statistical significance for genomewide studies.***Proc Natl Acad Sci USA*2003,**100:**9440–9445. 10.1073/pnas.1530509100PubMed CentralView ArticlePubMedGoogle Scholar - Reiner A, Yekutieli D, Benjamini Y:
**Identifying differentially expressed genes using false discovery rate controlling procedures.***Bioinformatics*2003,**19:**368–375. 10.1093/bioinformatics/btf877View ArticlePubMedGoogle Scholar - Storey JD:
**The positive false discovery rate: a Bayesian interpretation and the**q**-value.***Ann Statist*2004,**31:**2013–2035. 10.1214/aos/1074290335View ArticleGoogle Scholar - Dudoit S, Shaffer JP, Boldrick JC:
**Multiple hypothesis testing in microarray experiments.***Statistical Science*2003,**18:**71–103. 10.1214/ss/1056397487View ArticleGoogle Scholar - Qiu X, Klebanov L, Yakovlev A:
**Correlation between gene expression levels and limitations of the empirical Bayes methodology for finding differentially expressed genes, Statistical Applications in Genetics and Molecular Biology.**2005,**4:**Article 34.Google Scholar - Smyth GK, Michaud L, Scott HS:
**Use of within-array replicate spots for assessing differential expression in microarray experiments.***Bioinformatics*2005,**21:**2067–2075. 10.1093/bioinformatics/bti270View ArticlePubMedGoogle Scholar - Wilcox RR:
*Fundamentals of Modern Statistical Methods*. Springer, New York; 2001.View ArticleGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.