- Methodology article
- Open access
- Published:

# Utilization of two sample *t*-test statistics from redundant probe sets to evaluate different probe set algorithms in GeneChip studies

*BMC Bioinformatics*
**volumeÂ 7**, ArticleÂ number:Â 12 (2006)

## Abstract

### Background

The choice of probe set algorithms for expression summary in a GeneChip study has a great impact on subsequent gene expression data analysis. Spiked-in cRNAs with known concentration are often used to assess the relative performance of probe set algorithms. Given the fact that the spiked-in cRNAs do not represent endogenously expressed genes in experiments, it becomes increasingly important to have methods to study whether a particular probe set algorithm is more appropriate for a specific dataset, without using such external reference data.

### Results

We propose the use of the probe set redundancy feature for evaluating the performance of probe set algorithms, and have presented three approaches for analyzing data variance and result bias using two sample *t*-test statistics from redundant probe sets. These approaches are as follows: 1) analyzing redundant probe set variance based on *t*-statistic rank order, 2) computing correlation of *t*-statistics between redundant probe sets, and 3) analyzing the co-occurrence of replicate redundant probe sets representing differentially expressed genes. We applied these approaches to expression summary data generated from three datasets utilizing individual probe set algorithms of *MAS5.0*, *dChip*, or *RMA*. We also utilized combinations of options from the three probe set algorithms. We found that results from the three approaches were similar within each individual expression summary dataset, and were also in good agreement with previously reported findings by others. We also demonstrate the validity of our findings by independent experimental methods.

### Conclusion

All three proposed approaches allowed us to assess the performance of probe set algorithms using the probe set redundancy feature. The analyses of redundant probe set variance based on *t*-statistic rank order and correlation of *t*-statistics between redundant probe sets provide useful tools for data variance analysis, and the co-occurrence of replicate redundant probe sets representing differentially expressed genes allows estimation of result bias. The results also suggest that individual probe set algorithms have dataset-specific performance.

## Background

One of the most promising tools available today to researchers in the life sciences is high-density oligonucleotide array technology [1]. Denoted as GeneChips^{Â®}, high-density oligonucleotide arrays allow one to monitor the relative expression of tens of thousands of genes in a single assay. Upon its introduction within the last decade, GeneChip technology, together with cDNA microarray technology [2], has been viewed as state-of-the-art methodologies that would fundamentally alter the scientific landscape. Supporting this view, the number of published GeneChip studies has exponentially expanded in the past several years to reveal differentially expressed genes (DEGs), gene expression patterns, and genetic networks associated with many experimental conditions [3â€“5]. Yet the very first step, the generation of expression summary values, on which these studies rely is still open to extensive debate.

On Affymetrix GeneChips, the expression intensity of a gene is reported by a probe set that comprises 11â€“20 individual probe pairs. Each probe pair contains a perfect match (PM) 25 mer oligonucleotide probe, which is designed to hybridize specifically to a unique gene, and a mismatch (MM) probe of the same length, which differs from the PM probe by one single base in the center of the sequence. The MM probe is intended to measure non-specific hybridization. To compare gene expression levels between GeneChips hybridized with cRNAs of biological interest, the first step is to generate an overall probe set intensity measurement ("expression summary"), which represents the relative expression level of genes from their corresponding probe pairs.

Many probe set algorithms [6â€“9] have been proposed to generate expression summaries. As each one differs in its strategies for background subtraction, signal intensity normalization between arrays, non-specific hybridization correction, and probe intensity summary, the choice of probe set algorithm for expression summary has a great impact on the subsequent expression data analysis. For example, two lists of DEGs generated from the same dataset by different probe set algorithms were more than 30% different [10]. This discrepancy was also observed in our previous GeneChip studies, in which lists of DEGs from *MAS5.0* statistical algorithm, Model Based Expression Index (*MBEI* or *dChip*), or Robust Multi-array Analysis (*RMA*) shared similarity between only 10.1% and 36.8% (unpublished observation, Z. Hu and J. F. Collins). Therefore, it is important to compare the relative performance of different probe set algorithms. Comparative studies that have been performed to date are mainly based on spiked-in cRNAs from synthetic datasets [11â€“14]. These datasets, which contain either small [15] or large [14] numbers of spiked-in cRNAs at various known concentrations, have provided useful means to estimate result bias by comparing either the true expression ratios to observed expression ratios or false discovery rate to true positives, and therefore the relative performance of various probe set algorithms. Despite the important information obtained from these comparative analyses, there is currently no consensus as to which single algorithm yields more reliable results. The performance of individual probe set algorithms might be experiment-specific [16]. Since spiked-in cRNAs do not represent endogenously expressed genes in most experiments, it would be useful to have methods to determine whether a particular probe set algorithm is more appropriate for a specific dataset, without using such external reference data.

In this study, we propose to use probe set redundancy, a quite common feature of Affymetrix GeneChips in which a particular gene is represented by two or more probe sets (denoted as redundant probe sets or RPSs), for the evaluation of the relative performance of different probe set algorithms. We present three approaches for analyzing data variance and result bias using two sample *t*-statistics from redundant probe sets, rather than analyzing them from the hybridization intensities of RPSs on each individual GeneChip. These approaches include 1) RPS variance analysis based on *t*-statistic rank order, 2) correlation analysis of *t*-statistics between RPSs, and 3) analysis of the co-occurrence of replicate RPSs in lists of DEGs. The underling assumption is that these three analysis approaches should result in consistent outcomes; that is, data with lowest RPS variance are expected to also have the strongest RPS correlation and highest RPS co-occurrence in the list of DEGs.

We applied these approaches to expression summary data generated from three datasets utilizing the individual probe set algorithms of *MAS5.0*, *dChip*, or *RMA*. We also utilized combinations of options from the three probe set algorithms [14]. The three datasets we utilized include the GeneLogic dilution dataset [17], a wholly defined control dataset [14], and a dataset from a "real' experimental study on diabetes [18]. The first two datasets have been previously used by others in benchmarking studies to evaluate the performance of probe set algorithms [11â€“14]. Thus, the comparative results from these previous analyses can be readily used as the standard by which to judge the reliability of the results from our proposed three analysis approaches. Our results demonstrate that probe set redundancy can be used as an internal reference for probe set algorithm evaluation, and therefore provide a novel approach, by which researchers are able to assess the relative performance of different probe set algorithms on their experimental data, without using exogenous controls (*e.g*. spiked-in cRNAs). Furthermore, as GeneChips often contain large numbers of RPSs, our approach allows researchers to better evaluate probe set algorithms, with direct applications to experimental datasets of interest.

## Results

For gene expression studies, the ideal scenario is that RPSs for a given gene will have identical or similar expression values under given experimental conditions. This is however often not the case due to variances introduced by biological variability and differing RPS design, such as alternative RNA splicing [19], probe location within a gene, probe base composition, and location of RPSs on the GeneChip. To reduce or avoid such variances, we propose the use of two sample *t*-statistics of RPSs across all GeneChips in an experiment instead of hybridization intensities of RPSs on each individual GeneChip for analyzing data variance and result bias.

### Data variance and result bias analyses using GeneLogic dilution dataset

#### Data filtering and RPS assignment

In GeneChip studies, a large proportion of genes are usually not expressed across all samples to be compared and, as a common practice, are filtered out before performing statistical analysis. To mimic the real experimental situation, we used *MAS5.0* "present calls" to filter out probe sets whose expression intensities were close to the background noise across all samples in a *t*-test. We applied the same filtering of at least one "present call" out of 10 samples in a *t*-test to data obtained from each probe set algorithm, since all of them used the same set of CEL files, from which "present calls" were generated. This led to a ~42% data reduction in all comparisons as shown in Table 1. To assign RPSs, we used the UniGene ID (or LocusLink ID) and designated probe sets as redundant if a UniGene ID appeared two or more times in the filtered probe set list. As shown in Table 1, the RPSs constitute ~28% of the total probe sets in all comparisons.

#### RPS variance analysis based on t-statistic rank order

We first computed RPS variances for individual genes (*e.g*. 1,021 individual genes at the dilution level 7.5) as described in "Methods". We then depicted the overall RPS variances by computing either the average RPS variances from all individual genes or the relative RPS variances for data generated by different probe set algorithms. Data from *RMA* clearly showed much lower variance than those generated by *MAS5.0*, *dChip-PM*, and *dChip-PM/PM* (Figure 1a). This result was highly reproducible at all dilution levels.

To compute relative RPS variance, we used *MAS5.0* outputs as baselines and compared RPS variance of individual genes to that of the corresponding genes in data from *dChip-PM*, *dChip-PM/MM*, or *RMA*. The relative RPS variance was represented as the percentage of genes whose RPS variance was smaller than the corresponding genes in data generated by *MAS5.0*. Thus, a number greater than 50% indicated an overall smaller relative RPS variance than *MAS5.0*. In agreement with the average RPS variance, data from *RMA* displayed smaller relative RPS variances when compared to those from other probe set algorithms (Figure 1b), in which 54.7% to 58.9% of genes had smaller RPS variance than those from *MAS5.0*. By contrast, the difference in the relative RPS variance between data generated by *MAS5.0* and those from *dChip-PM*, or *dChip-PM/MM* was not consistent at different dilution levels, in which 46.7% to 54.4% of genes had smaller RPS variance than those from *MAS5.0*. We also verified the statistical significance of the relative RPS variance between *MAS5.0* and *RMA*, *dChip-PM*, or *dChip-PM/MM* by using the Wilcoxon signed-rank test. We found that median RPS variances for data from *RMA* were significantly different from those from *MAS5.0*, with < 10^{-5} *p*-values for all dilution levels. Conversely, data from both *dChip-PM* and *dChip-PM/MM* did not display significant differences in RPS variance for all dilution levels, with the exception being dilution level 7.5 in which result from *dChip-PM* had *p*-value of 0.031. Thus, of all data generated by different probe set algorithms, those from *RMA* had the smallest RPS variance.

#### Correlation analysis of t-statistics between RPSs

The second approach we used for analyzing data variance was to investigate the correlation of *t*-statistics between RPSs. The underling hypothesis is that *t*-statistics of RPSs with similar expression patterns in two sample groups should be highly correlated. In this study, we employed Pearson correlation coefficients to estimate the correlation between paired *t*-statistics from RPSs. For genes with more than two RPSs, we generated all combinations of paired *t*-statistics to the number of \left(\begin{array}{l}N\\ 2\end{array}\right) pairs, where *N* is the number of RPSs for a given gene. As shown in Figure 2a, the paired *t*-statistics clearly show stronger correlations for data generated by *RMA* than those from other probe set algorithms at all dilution levels, indicating that data from *RMA* have more similar RPS expression patterns in two sample groups, and thus possibly lower variance.

We were concerned that the superior results from data generated by *RMA* might be associated with stronger random or biological correlation, which could potentially affect the outcomes in the comparisons described above. To investigate this possibility, we performed further analysis to estimate the correlation for lists of randomly paired *t*-statistics. These *t*-statistics were randomly sampled without replacement from *t*-statistics of all probe sets in the corresponding expression summary datasets and then used to form *t*-statistic pairs of the same size as those used in the correlation analysis for RPSs. As an example, one of the random correlation comparisons, dilution level 7.5, is shown in Figure 2b. The distributions of the correlation coefficients from 1,000 simulated lists of randomly paired *t*-statistics are centered at zero for each probe set algorithm, suggesting that no random or biological correlation exists in any of the data used for RPS correlation analysis. Similar results were also obtained from the other 5 dilution levels (data not shown).

#### Analysis of co-occurrence of replicate RPSs in lists of DEGs

We define "replicate RPSs" as those with two or more RPSs for a given gene in the final list of DEGs, and those with only one RPS as "single RPS". In the ideal scenario of no RPS variance, we expect an "all-or-none" expression pattern of RPSs for individual genes in the list of DEGs. Therefore, the degree of deviation from this ideal scenario provides a tool for researchers to estimate the result bias. The approach we used was to calculate the proportion of replicate RPSs out of both replicate and single RPSs in DEGs.

To estimate the fraction of replicate RPS in the list of DEGs, we first used the Benjamini-Hochberg [20] approach to control experiment-wise false discovery rate (FDR) to select significant genes. The adjustment for the multiple testing allowed us to determine significant *p*-values and thus establish statistical cutoffs for significant genes. The relative proportions of replicate RPSs out of all RPSs in DEG list from different probe set algorithms were consistent in all dilution levels (Figure 3aâ€“3f), when using the FDR cutoffs in a range similar to what would be utilized in actual experimental situations (0.01 to 0.1). Whereas results for *RMA* displayed the highest co-occurrence rate, results for *MAS5.0* showed the lowest co-occurrence rates. Results for *dChip-PM* and *dChip-PM/MM* had similar co-occurrence rates, which were between those from *RMA* and *MAS5.0*. These results were highly reproducible at all dilution levels, except dilution level 20.0 from *MAS5.0*.

To investigate whether data filtering would affect the outcomes of RPS variance and result bias analysis, we also utilized the entire dataset, in which the RPSs constitute 34.9% of the total probe sets, to perform the analyses. Results of RPS variance, correlation of *t*-statistics between RPSs, and the co-occurrence of replicate RPSs in DEGs all strongly agreed with results obtained from data filtered by one "present call", but with relatively smaller differences between data from *RMA* and data from other probe set algorithms (data not shown). These results suggest that data filtering may improve the performance of the three proposed approaches, although data can be directly used without filtering for RPS variance and result bias analysis.

Taken together, results from all three approaches indicated that data generated by *RMA* were superior to those generated by other probe set algorithms. This fact was reflected by RPSs with lower variance, stronger correlation, and higher co-occurrence rates in final DEG list. These results were also in agreement with previous benchmarking studies using the GeneLogic dilution dataset [11â€“13], which indicated that *RMA* had a superior performance over other probe set algorithms, as demonstrated by the lower expression variance and smaller result bias. Although the performance of *RMA* is better than that of other probe set algorithms we tested, it is worthy to note that *RMA* is not necessarily the most robust probe set algorithm in most cases, which was illustrated by the relative improvement of RPS variances and the findings from the remainder of this study using the two other datasets.

### Data variance and result bias analyses using the wholly defined control dataset

#### Summary of results from previous wholly control dataset study

This wholly defined control dataset was constructed and has been used by Choe *et al*. [14] to evaluate the methods and analysis options for Affymetrix GeneChips to identify DEGs. These options, which were derived from several probe set algorithms commonly used for GeneChip analysis, include background correction, PM correction, probe level normalization, expression summary, and probe set level normalization. These authors have applied all possible combinations of options to the data to assess whether some steps are more critical than others in maximizing the detection of DEGs. Using receiver operator characteristic (ROC) curves of false discovery rate vs. true positives, they have facilitated the assessment of the performance of various options and have reported the following findings:

**1.** A second loess normalization at the probe set level yielded a superior result.

**2.** Among various robust estimators used for expression summary, the *median polish* method performed the best.

**3.** Among different PM correction options, the method in *MAS5.0* had a superior performance over the others.

**4.** Among the background correction methods, the *MAS5.0* method generally performed better than the *RMA* method.

**5.** No clearly superior normalization method was found at the probe level.

**6.** Ten best expression summary datasets, which maximized the detection of DEGs and minimized false positives, were generated by a combination of optimal options.

These findings, which resulted from the analyses of a large number of expression summary datasets generated from the wholly control dataset by different combinations of options, provide standards to which our current assessments can be compared.

#### Comparison of results from the three proposed approaches and previous studies

We made use of RPSs from the 3,860 individual cRNAs for the study reported here, as they are known sequences, and the complete annotations can be found in the Drosophila Gene Collection release 1.0 [21]. Based on Drosophila Gene Collection clone IDs, we were able to obtain 582 RPSs, which represented 268 unique genes, from the 3,860 cRNAs. We computed data variance at 8 levels with fold change equal to or greater than 1.0, 1.2, 1.5, 1.7, 2.0, 2.5, 3.0, 3.5, or 4.0, although the majority of the RPSs (69.2%) were spiked in at identical concentrations between the two sample groups.

We employed the same analysis procedures as described for the GeneLogic dilution dataset to compute variance, correlation, and co-occurrence for RPSs. To make our results comparable to those from Choe *et al*. [14], we used 110 expression summary datasets to perform the comparison of options for probe set level normalization, and used 55 of the expression summary datasets that were generated with the loess normalization option (orange lines in Figure 4a, Figure 5a, and Figure 6a) for comparison of the remaining options.

Among our three proposed analysis approaches, results of RPS variances (Figure 4) were consistent with findings from Choe *et al*. [14]. First, a second loess normalization at the probe set level consistently decreased the RPS variance, as can be seen from the log average RPS variance across all fold change cutoffs in Figure 4a. This is also true for expression summary options, for which results from the *median polish* method displayed relatively lower RPS variance (Figure 4b). Second, Figure 4d clearly show that expression summary datasets from *MAS5.0* PM correction have lower RPS variances when all RPSs are considered (fold change >= 1.0), which is similar to previous study [14]. Similar results were also obtained for the 10 best expression summary datasets (Figure 4f), which displayed relatively lower RPS variance compared to the remaining expression summary datasets. It is noteworthy that in the higher fold change ranges (fold change >= 1.2 â€“ 4.0) the difference in RPS variance is not apparent between different PM correction options, which is also true for the 10 best expression summary datasets as compared to all other expression summary datasets. This may be due to the smaller number of RPSs used when computing RPS variance. For the remaining two comparisons, with data normalization at the probe level (Figure 4c) and background correction (Figure 4e), no significant differences was observed between options.

In agreement with the findings reported by Choe *et al*. [14], our additional analyses indicated that a second loess normalization at the probe set level substantially increased both correlation (Figure 5a) and co-occurrence (Figure 6a) of RPSs. Additionally, the *median polish* method for expression summary had a much better performance than MAS5.0, as can be seen from the higher correlations shown in Figure 5b as well as higher co-occurrence of RPSs shown in Figure 6b. For the remaining options, no single one stood out as clearly superior, but some options generally performed better than others. For example, the co-occurrences of RPSs from the *MAS5.0* option for PM correction scored higher (Figure 6d), demonstrating its relatively better performance. This was also true for the 10 best expression summary datasets (Figure 6f).

### Data variance and result bias analyses using our experimental diabetes dataset

#### Dataset features and analysis results

The third dataset came from an experimental dataset designed to study diabetes. To increase the sensitivity for detecting DEGs, we developed a statistical approach to eliminate expression outliers across biological replicates. Briefly, for each individual probe set from 5 biological replicates one of them, whose expression value had the largest deviation from the sample mean \stackrel{\xc2\xaf}{x}=\frac{1}{5}{\displaystyle {\xe2\u02c6\u2018}_{i=1}^{5}{x}_{i}}, was first selected as the putative outlier *O*_{
p
}. The sample mean {\stackrel{\xc2\xaf}{x}}_{(i)}=\frac{1}{4}{\displaystyle \underset{\begin{array}{l}{i}^{\xe2\u20ac\xb2}=1\\ {i}^{\xe2\u20ac\xb2}\xe2\u2030i\end{array}}{\overset{5}{\xe2\u02c6\u2018}}{x}_{i}} and standard deviation *s*_{(i)}(*x*_{
i
}) from the rest 4 replicates were subsequently recomputed, which was used to build a 99% confidence interval {\stackrel{\xc2\xaf}{x}}_{(i)}\xc2\pm t{s}_{{\stackrel{\xc2\xaf}{x}}_{(i)}} using the *t* distribution, where {s}_{{\stackrel{\xc2\xaf}{x}}_{(i)}}={s}_{(i)}({x}_{i})/\sqrt{4} and t = 5.84. The sample mean {\stackrel{\xc2\xaf}{x}}_{(i)} was then used to replace the outlier as follow:

{O}_{p}=\{\begin{array}{lll}{\stackrel{\xc2\xaf}{x}}_{(i)}\hfill & \text{if}{O}_{p}>{\stackrel{\xcb\u2122}{x}}_{(i)}+t{s}_{{\stackrel{\xc2\xaf}{x}}_{(i)}}\hfill & \text{or}{O}_{p}<{\stackrel{\xcb\u2122}{x}}_{(i)}\xe2\u02c6\u2019t{s}_{{\stackrel{\xc2\xaf}{x}}_{(i)}}\hfill \\ {O}_{p}\hfill & \text{otherwise}\hfill \end{array}

This approach improves the detection sensitivity only for the probe sets that have high homogeneity in expression intensity across at least four out of five biological replicates, and therefore genes identified as significant have high rate of accuracy. This dataset, in which the RPSs constitute 34.6% of the total probe sets after performing data filtering, provided an example of a typical "real" experiment, from which biological and experimental validation was readily available.

We followed the same analysis procedures as used for the GeneLogic dilution and wholly defined control datasets, and found that results from our 3 proposed approaches were generally in agreement. Data generated by *MAS5.0* and *dChip-PM/MM* were superior as compared to data from *RMA* and *dChip-PM*. This was especially apparent from the analysis of co-occurrence of replicate RPSs in list of DEGs, in which co-occurrence rates for data from *MAS5.0* and *dChip-PM/MM* were higher than data from *RMA* and *dChip-PM* (Figure 7), when controlling the FDR in the range used in actual experimental situations (0.01 to 0.1). In support of this finding, RPS variance analysisalso indicated that data from *MAS5.0* and *dChip-PM/MM* had smaller variance than data from the other 2 probe set algorithms. For example, for the relative RPS variance analysis the percentage of genes whose RPS variance was smaller than the corresponding genes from *RMA* was 52.5%, 52.4%, and 50.1% for *MAS5.0*, *dChip-PM/MM*, and *DChip-PM*, respectively.

#### Quantitative RT-PCR validation

Quantitative RT-PCR (qRT-PCR) is a common and useful method for confirming DEGs, and thus for validating results from GeneChip experiments. Ten genes distributed among different functional groups identified in a previous study were selected for qRT-PCR studies [18]. In that study data in diabetic rat group was compared to all those in the normal rat group as suggested by Affymetrix [22], which involved looking at a 5 Ã— 5 matrix for the experiment. The specific genes, primers used, and fold change values found by PCR are shown in Table 3. For each of the 10 genes, qRT-PCR average fold change from biological replicates between diabetic and control animals were first computed and then compared with the average fold change from GeneChips. As shown in Figure 8, the fold changes from *MAS5.0* and *dChip-PM/MM* are highly correlated with those from qRT-PCR, with correlation coefficients of 0.9 for both methods. Conversely, the fold changes from *RMA* and *dChip-PM* showed relatively weak correlation with those from qRT-PCR, with correlation coefficients of 0.8 for *dChip-PM* and 0.74 for *RMA*.

#### Biological validation

We also explored the level of concordance of the biological themes for DEGs to the results of data variance and result bias analyses. Taking advantage of known knowledge from previous diabetes studies, we categorized the statistically over-represented biological process categories for DEGs by DAVID 2.0 [23]. Because diabetes is a multifactorial disease [24], which leads to substantial changes in gene expression in a broad range of biological function categories, the number of over-represented biological processes is usually overwhelming, with many overlaps found using different probe set algorithms. In this study, the enriched functional categories included protein biosynthesis, macromolecule metabolism, physiological process, ribosome biogenesis, and regulation of cell proliferation. Nevertheless, we determined uniquely over-represented biological process from DEG lists generated by the chosen probe set algorithms to determine which one was superior from a physiological perspective.

Previous studies using GenChips with samples obtained from mice with streptozotocin-induced diabetes indicated that major diabetes related changes in gene expression included carbohydrate and lipid metabolism, energy metabolism, cellular transport and vesicle trafficking, intracellular signaling, and response to stress [25]. In our GeneChip studies reported here RNA was obtained from normal rats and rats with streptozotocin-induced diabetes. As shown in Table 2, biological function categories associated with diabetes [24, 25] such as intracellular signaling cascade, lipoprotein biosynthesis, steroid metabolism, negative regulation of transcription, phosphagen biosynthesis, and response to stress are enriched to a larger extent in DEGs either from *MAS5.0* or *dChip-PM/MM*. On the other hand, DEGs from *dChip-PM* have only one diabetes-related biological process, and all uniquely enriched biological processes for DEGs from *RMA* do not appear to be directly related to the diabetes phenotype. Thus, DEG lists generated by *MAS5.0* and *dChip-PM/MM* were most relevant to the disease state being studied.

## Discussion

The choice of probe set algorithms used for expression summary in GeneChip studies has a major impact on differential gene expression analysis, as important differences exist in the way the expression summary is generated using the various algorithms. Spiked-in cRNAs with known concentrations are often used to assess the relative performance of different probe set algorithms. However, this approach has apparent limitations because spiked-in cRNAs are not endogenously expressed genes in experimental systems of biological interest. In addition, if we assume that the performance of individual probe set algorithms is experiment-specific [16], the lack of spiked-in cRNA controls in most experiments prevents this approach from applying to many datasets. In contrast, our proposed new approach of using probe set redundancy addresses this issue adequately, as GeneChips often contain large numbers of RPSs, which represent experimental genes under study.

For comparison purposes, we initially employed the GeneLogic dilution dataset to directly estimate RPS variances using the expression values of RPSs by the method of Barash *et al* [10]. Notably, although RPS variances based on expression values differed significantly among data from different probe set algorithms (data not shown), the results did not agree with both ours and those previously reported by others [11â€“13], indicating that the estimation of RPS variance was less reliable using expression values than using *t*-statistic values.

For our study, rather than analyzing data variance and result bias from the hybridization intensities of RPSs on each individual GeneChip [10], we chose to use two sample *t*-test statistics to compare hybridization intensities of individual RPSs across all GeneChips in an experiment. This approach allowed us to obviate the variance introduced by differing probe set design and location of the probe within a primary or alternatively spliced transcript, which introduces RPS variance on an individual GeneChip. Other important issues are biological variability between samples, technical issues related to RNA integrity and sample preparation, and non-specific, cross-reactivity of probes [26]. If we thus assume that these introduced variances have the same impact on results obtained by different probe set algorithms (as they all use the same dataset), then we can minimize RPS variance on an individual GeneChip by comparing expression between different GeneChips and can thus effectively evaluate the performance of different probe set algorithms. The resulting *t*-statistics used to analyze RPSs across GeneChips would therefore reflect whether or not the hybridization intensity differences of RPSs within each sample group were consistent with the differences in other sample groups.

The validity of using RPS and two sample *t*-test statistics for the evaluation of probe set algorithm performance was clearly demonstrated by the correlation analysis. For example, in the GeneLogic dilution dataset, strong correlations (0.59 to 0.74) of *t*-statistics between RPSs were obtained from data for all probe set algorithms (Figure 2a), while no correlation existed for randomly paired *t*-statistics from corresponding expression summary datasets (Figure 2b). It is noteworthy that correlations of RPS *t*-statistics were less than the ideal situation of 1 (Figure 2a), which might be due to the variances introduced from such confounding factors as cross-hybridization and recognition of alternatively spliced transcripts as discussed above and suggested by others [26]. Nevertheless, the use of RPS and two sample *t*-test statistics for the evaluation of different probe set algorithm performance is clearly appropriate, as the RPSs usually constitute a large portion of the probe sets on GeneChips and the introduced variances have the same impact on results obtained by different probe set algorithms.

We utilized 3 approaches, RPS variance analysis, correlation analysis of *t*-statistics between RPSs, and analysis of the co-occurrence of replicate RPSs in DEGs, as tools to judge data variances and result bias. We also assumed that results from the three approaches would be consistent for the same expression summary dataset. Out of the three approaches, however, the correlation analysis proved to be the least useful, as correlation measures the trend of *t*-statistics instead of absolute variances. This was the case from the results of the wholly defined control dataset, in which the correlation analysis displayed the lowest level of concordance to the results from Choe *et al*. [14]. While both co-occurrence of replicate RPS analyses and RPS variance are more robust than correlation analysis, the analysis of co-occurrence of replicate RPSs in DEGs provides an approach for researchers to estimate result bias, and RPS variance analysis provides a tool for data variance analysis.

Three datasets used in this study come from different sources and had differing data quality and features, they therefore nicely represent most situations. Above all, results of data variance and result bias for expression summary datasets, whether they were generated from individual probe set algorithms or combinations of options from a few probe set algorithms, agreed well with each other in each individual test dataset and were also in good agreement with either previous findings by others or experimental evaluation from this study. First, RPS variance analyses using expression summary datasets generated from the GeneLogic dilution dataset by individual probe set algorithms indicated that data from *RMA* had an overall better quality, which was in good agreement with findings by others [11â€“13]. Second, our results from the wholly defined control dataset agreed with those conclusions from Choe *et al*. [14]. However, unlike data generated from GeneLogic dilution dataset, the expression summary datasets were generated by combinations of options from a few probe set algorithms and were in large numbers, thus allowing more sophisticated and comprehensive assessments. Finally, for our diabetes dataset, results from both experimental validation by qRT-PCR and biological validation by functional classification for DEGs agreed well with results from data variance and bias analyses, in which data from probe set algorithms using both PM and MM probes for expression summary displayed lower variance and bias.

Although our proposed analysis approaches were to evaluate the relative probe set algorithm performance within individual datasets, it was interesting to note that the fraction of replicate RPS for DEGs in the diabetes dataset was smaller than those in the other two datasets. This difference was also observed between unfiltered (data not shown) and one "present call" filtered GeneLogic dilution dataset, in which fractions of replicate RPSs for DEGs in the one "present call" filtered data were 5â€“11% higher than those obtained from data without filtering. These differences could be due to the different nature of RPSs on the different GeneChips or differing data quality.

It is noteworthy that varying conclusions for probe set algorithm performance were drawn by using different datasets. Irizarry *et al*. [12] found that *RMA* performed best among a few probe set algorithms tested, when using the GeneLogic dilution dataset. On the other hand, using the wholly defined control datasets no best single method was found by Choe *et al*. [14], who instead suggested a best-route combinations of analysis options from *MAS5.0*, *RMA*, *dChip*, and an additional loess normalization at the probe set level. Results from the diabetes dataset used in this study, when used for the evaluation of probe set algorithm performance, indicated that probe set algorithms using both PM and MM probe sets for expression summary gave better results than those methods using only PM for expression summary, such as *RMA*. Overall these results suggest that individual probe set algorithms may have experiment-specific performance. Moreover, the strong correlation between the results from previous probe set algorithm assessments and those from this study demonstrate that our proposed novel approaches based on RPS analysis are not dependent upon individual probe set algorithm performance, and are thus very likely to be reliable and reproducible.

## Conclusion

An important issue in the analysis of gene expression data from Affymetrix GeneChips is to choose a probe set algorithm for expression summary. In this study, we have proposed the use of probe set redundancy to evaluate the performance of different probe set algorithms, and have presented three approaches for assessing data variance and result bias using two sample *t*-test statistics. These methods include RPS variance analysis based on *t*-statistic rank order, correlation analysis between *t*-statistics of RPSs, and analysis of co-occurrence of replicate RPSs in DEG list. The main advantage of our approaches lies in the fact that we do not make use of external reference data, but rather investigate data variance and result bias based on RPSs, which often constitute a large portion of and have direct relevance to the genes under study. Furthermore, the use of *t*-statistics allows us to reduce or avoid RPS variances introduced from both biological variability and differing probe design. To assess the usefulness of the three proposed approaches, we have applied them to three diverse datasets using a few widely used probe set algorithms. Results from all three approaches not only agreed well with each other in each individual test dataset but they were in good agreement with either previous findings by others or experimental validation from this study. These approaches provide an alternative method to determine data variance and result bias, without the use of exogenous controls, and are thus useful for the assessment of probe set algorithm performance. The results also suggest that individual probe set algorithms have dataset-specific performance.

## Methods

### Datasets

#### GeneLogic dilution/mixture dataset

The GeneLogic dilution/mixture dataset [17] comprises 75 HG-U95A GeneChips to which two sources of RNA, human liver tissue and a central nervous system cell line (CNS), have been hybridized in various dilutions and combinations. Of these 75 GeneChips, 60 of them were hybridized with cRNA from either liver tissue or CNS at the concentrations of 1.25, 2.5, 5.0, 7.5, 10.0 or 20.0 Î¼g, with 5 replicate GeneChips for each dilution level. Data from these 60 GeneChips, called the GeneLogic dilution dataset, have been used in this study.

#### A wholly defined control dataset

Choe *et al*. [14] recently performed a study to evaluate the methods and analysis options for Affymetrix GeneChips for identifying DEGs. They have constructed a wholly defined control dataset to mimic the scenario of comparing two samples in a microarray experiment. This dataset comprises 3,860 individual *Drosophila* cRNAs of known sequences. Out of the 3,860 cRNAs, 1309 have been spiked in with different ratios of 1.2, 1.5, 1.7, 2.0, 2.5, 3.0, 3.5, or 4.0 between two test sample groups, and the remaining 2,551 cRNAs have the same relative concentration in each test sample group. Each sample has hybridized in triplicate to Affymetrix *Drosophila* arrays (DrosGenome1).

#### Diabetes dataset

The third dataset consists of 10 Affymetrix GeneChips from diabetes study [18]. Gene expression data was obtained from 2 groups of rats: normal rats and rats with streptozotocin-induced diabetes for four weeks. In each group of 5 animals, labeled cRNA was prepared from 10.0 Î¼g total RNA obtained from the leg muscle of one animal and was separately hybridized to one Affymetrix U34A Rat GeneChip. GeneChips were scanned at the Gene Expression Core Facility at Roswell Park Cancer Institute.

### Generating expression summary values

Out of many probe set algorithms, the most widely used are the *MAS5.0* statistical algorithm from Affymetrix [6], the Model Based Expression (*dChip*) Index of Li and Wong [7, 8], and the Robust Multi-Chip Analysis (*RMA*) of Irizarry *et al*. [9]. All of them are based on the statistical model of gene expression values as a function of the probe level intensities. There are, however, important differences between these probe set algorithms.

*MAS5.0* is based on the assumption of homogeneity of probe affinity and weights each probe in a probe set equally. *MAS5.0* uses both PM and MM probes for expression summary from the average value of PM â€“ MM in a probe set and employs the Tukey's Biweight weighted average of the probe level signals to avoid sensitivity to outlier probe intensity. Unlike *dChip* and *RMA*, *MAS5.0* summarizes expression individually for each GeneChip and performs a scaling normalization at the level of expression summary.

The *dChip* method assumes different probe affinity in a probe set and considers a probe value as the product of gene expression and probe-sensitivity index. The *dChip* method uses a multiplicative model and employs probe values of the same probe set from all GeneChips in an experiment to iteratively estimate gene expression-specific index as well as probe affinity-specific index for expression summary. Normalization in *dChip* is carried out at the probe level, which makes use of an invariant subset of probes that have small within-subset rank differences between GeneChips. Additionally, the *dChip* performs the expression measures using either only the PM probes ("*dChip-PM"*) or both PM and MM probes ("*dChip-PM/MM"*).

Like *dChip, RMA* makes use of data from all GeneChips in an experiment for normalization and considers the probe affinity effect. *RMA* uses a linear additive model and computes the expression summary values by the use of only PM probes. Normalization in *RMA* is performed by quantile normalization that transforms the PM distribution of each GeneChip in a dataset to a common distribution.

We used the three probe set algorithms to generate expression summary values for the GeneLogic dilution and diabetes datasets. For *MAS5.0*, we used the Affymetrix software suite *MAS5.0* with its default settings and adjusted the target intensity level to 2500 to bring the total expression intensity of each GeneChip to a fixed level. We also performed further global normalization to bring the median expression values of all GeneChips to the same scale. This was done by selecting a baseline GeneChip from the dataset, followed by scaling each GeneChip to the median of the baseline GeneChip ({\stackrel{\xcb\u0153}{m}}_{base}):

{x}_{i}^{\text{'}}=\frac{{\stackrel{\xcb\u0153}{m}}_{base}}{{\stackrel{\xcb\u0153}{m}}_{i}}{x}_{i}.

For *dChip*, we used the official *dChip* version 1.3 software [27] to obtain expression summary values for both *dChip-PM* and *dChip-PM/MM*. For *RMA*, we utilized the default *rma*() function included in the "affy" package of Bioconductor in the R statistical computing environment [28]. This default function employs *median polish* for expression summary and quantile normalization for data normalization.

The expression summary values of the wholly defined control dataset were directly obtained from Choe *et al*. [14], who created 150 multiple expression summary datasets using combinations of options from probe set algorithms *MAS* (both version 4 and 5), *dChip*, and *RMA*. These options included 3 for background correction, 4 for normalization at the probe level, 3 for PM correction, 3 for expression summary, and 2 for normalization at the probe set level. In the background correction options, the use of "subtract MM for PM correction" (*MAS* version 4) resulted in negative values when PM was less than MM, and in this case about 85% of the probe sets on the GeneChip were flagged as "not applicable" when expression summaries from *MAS* and *RMA* were used. We therefore excluded 40 expression summary datasets generated from the use of this option, and employed the remaining 110 expression summary datasets in our study.

### Welch t-test

Welch *t-* tests were performed on a probe set-by-probe set basis between two sample groups; 6 Ã— 4 for data generated from the GeneLogic dilution dataset between liver and *CNS* at the same dilution level, 1 Ã— 4 for data generated from the diabetes dataset, and 110 for expression summary datasets from the wholly defined control dataset between two sample groups generated by the use of the same combinations of options. The Welch *t*-test is specifically designed to handle the possibility of having small samples with unequal variances, and thus has been most widely used in RNA profiling data analysis. This *t*-test relies on the assumption of data normality and homoscedansticity of expression values, which may not be valid for all data. Nevertheless, the *t*-test is the best option for our study, as it has more power than non-parametric tests such as the Wilcoxon rank sum test, in addition to its popularity for gene expression data analysis.

### Computing RPS variance based on t-statistic rank order

From each Welch *t*-test, let *n* denote the number of total null hypotheses, and *T*_{
k
}the *k*-th largest *t*-statistic value, probe sets are ranked in descending order based on *t*-statistic values.

*T*_{
1
},*T*_{
2
},*T*_{
3
},*T*_{
4
}, ....., *T*_{
k
}, ...., *T*_{
n
}.

The corresponding rank order of probe sets is

*R*_{
1
},*R*_{
2
},*R*_{
3
},*R*_{
4
}, ....., *R*_{
k
}, ...., *R*_{
n
},

where *R*_{
1
}= *1*, *R*_{
n
}= n, *R*_{
k
}the *k*-th rank order of a probe set. We compute the data variance of a gene *j* with *N*_{
j
}RPSs:

{\stackrel{\xcb\u0153}{\mathrm{\xcf\u0192}}}_{j}=\frac{1}{{N}_{j}}{\displaystyle \underset{n=1}{\overset{{N}_{j}}{\xe2\u02c6\u2018}}\left|{R}_{j{k}_{n}}\xe2\u02c6\u2019\mathrm{\xce\xbc}({R}_{j})\right|}

As an alternative, the following can also be used to compute RPS variance:

{\mathrm{\xcf\u0192}}_{j}^{2}=\frac{1}{{N}_{j}}{\displaystyle \underset{n=1}{\overset{{N}_{j}}{\xe2\u02c6\u2018}}{\left[{R}_{j{k}_{n}}\xe2\u02c6\u2019\mathrm{\xce\xbc}({R}_{j})\right]}^{2}}

where *k* is the *k*-th rank order for a probe set of gene *j*, and Î¼ (*R*_{
j
}) the average distance of the *N*_{
j
}probe sets for gene *j*:

\mathrm{\xce\xbc}({R}_{j})=\frac{1}{{N}_{j}}{\displaystyle \underset{n=1}{\overset{{N}_{j}}{\xe2\u02c6\u2018}}{R}_{j{k}_{n}}}

We employed *t*-statistic values to rank probe sets instead of using the *p*-values. This is because *t*-statistic values reflect the variances of RPSs in rank order more appropriately than *p*-values do, as the *t*-statistic values distinguish the up-regulation from down-regulation for RPSs, but the *p*-values do not. Another alternative for computing the variances of RPSs is to directly use the *t*-statistic values, but this approach is less robust, as *t*-statistic values from different probe set algorithms may not be in the same scale.

### Quantitative RT-PCR

Quantitative real time RT-PCR was performed for 10 genes from the diabetes dataset (Table 3). Total RNA from either normal or diabetic animals was reverse transcribed and used to generate cDNA with Invitrogen SuperScript First-Strand Synthesis System Kit (Carlsbad, CA). Primers for selected genes were designed using Primer3 developed at Whitehead Institute and Howard Hughes Medical Institute (Cambridge, MA) and synthesized by Sigma-Genosys (The Woodlands, TX). Primer sequences were designed to flank an intron near the 3' end of the gene sequence, have 20 to 22 bp, and contain 50â€“60% GC sequences and a G or C at the 3'end. Real time PCR was performed using cDNA, gene specific primers, SYBR Green PCR Core Reagent Kit obtained from Applied Biosystems (Foster City, CA), and the iCycler IQ Real Time PCR detection system from BioRad (Hercules, CA). As a constitutive control, the Î²2 microglobuin gene was used. The PCR efficiency was 98â€“100% for all primer sets as determined using standard curves to test linearity. The fold change of the comparison of diabetic to normal rats was calculated as 2 ^{(Ct target geneD - Ct control gene)-(Ct target geneN - Ct control gene)}, where Ct is the threshold cycle.

## Abbreviations

*MAS5.0*(Microarray Suite 5.0 statistical algorithm from Affymetrix):-
*dChip*or MBEI (Model based expression index),*RMA*(Robust multi-array average), PM (Perfect match probe), MM (Mismatch probe), DEG (Differentially expressed genes), RPS (Redundant probe set), qRT-PCR (Quantitative real time â€“ polymerase chain reaction), ROC (Receiver operator characteristic).

## References

Lockhart DJ, Dong H, Byrne MC, Follettie MT, Gallo MV, Chee MS, Mittmann M, Wang C, Kobayashi M, Horton H, Brown EL:

**Expression monitoring by hybridization to high-density oligonucleotide arrays.***Nat Biotechnol*1996,**14**(13):1675â€“1680. 10.1038/nbt1296-1675Schena M, Shalon D, Davis RW, Brown PO:

**Quantitative monitoring of gene expression patterns with a complementary DNA microarray.***Science*1995,**270**(5235):467â€“470.Chesler EJ, Lu L, Shou S, Qu Y, Gu J, Wang J, Hsu HC, Mountz JD, Baldwin NE, Langston MA, Threadgill DW, Manly KF, Williams RW:

**Complex trait analysis of gene expression uncovers polygenic and pleiotropic networks that modulate nervous system function.***Nat Genetics*2005,**37**(3):233â€“242. 10.1038/ng1518Hunter BG, Beatty MK, Singletary GW, Hamaker BR, Dilkes BP, Larkins BA, Jung R:

**Maize opaque endosperm mutations create extensive changes in patterns of gene expression.***Plant Cell*2002,**14**(10):2591â€“2612. 10.1105/tpc.003905Misson J, Raghothama KG, Jain A, Jouhet J, Block MA, Bligny R, Ortet P, Creff A, Somerville S, Rolland N, Doumas P, Nacry P, Herrerra-Estrella L, Nussaume L, Thibaud MC:

**A genome-wide transcriptional analysis using Arabidopsis thaliana Affymetrix gene chips determined plant responses to phosphate deprivation.***PNAS USA*2005,**102**(33):11934 -111939. 10.1073/pnas.0505266102**Affymetrix [http://www.affymetrix.com/support/technical/byproduct.affx?product=mas]**Li C, Hung Wong W:

**Model-based analysis of oligonucleotide arrays: model validation, design issues and standard error application.***Genome Biol*2001,**2**(8):RESEARCH0032.Li C, Wong WH:

**Model-based analysis of oligonucleotide arrays: expression index computation and outlier detection.***PNAS USA*2001,**98**(1):31â€“36. 10.1073/pnas.011404098Irizarry RA, Hobbs B, Collin F, Beazer-Barclay YD, Antonellis KJ, Scherf U, Speed TP:

**Exploration, normalization, and summaries of high density oligonucleotide array probe level data.***Biostatistics*2003,**4**(2):249â€“264. 10.1093/biostatistics/4.2.249Barash Y, Dehan E, Krupsky M, Franklin W, Geraci M, Friedman N, Kaminski N:

**Comparative analysis of algorithms for signal quantitation from oligonucleotide microarrays.***Bioinformatics*2004,**20**(6):839â€“846. 10.1093/bioinformatics/btg487Bolstad BM, Irizarry RA, Astrand M, Speed TP:

**A comparison of normalization methods for high density oligonucleotide array data based on variance and bias.***Bioinformatics*2003,**19**(2):185â€“193. 10.1093/bioinformatics/19.2.185Irizarry RA, Bolstad BM, Collin F, Cope LM, Hobbs B, Speed TP:

**Summaries of Affymetrix GeneChip probe level data.***Nucleic Acids Res*2003,**31**(4):e15. 10.1093/nar/gng015Cope LM, Irizarry RA, Jaffee HA, Wu Z, Speed TP:

**A benchmark for Affymetrix GeneChip expression measures.***Bioinformatics*2004,**20**(3):323â€“331. 10.1093/bioinformatics/btg410Choe SE, Boutros M, Michelson AM, Church GM, Halfon MS:

**Preferred analysis methods for Affymetrix GeneChips revealed by a wholly defined control dataset.***Genome Biol*2005,**6**(2):R16. 10.1186/gb-2005-6-2-r16**Affymetrix [http://www.affymetrix.com/support/technical/sample_data/datasets.affx]**Seo J, Bakay M, Chen YW, Hilmer S, Shneiderman B, Hoffman EP:

**Interactively optimizing signal-to-noise ratios in expression profiling: project-specific algorithm selection and detection p-value weighting in Affymetrix microarrays.***Bioinformatics*2004,**20**(16):2534â€“2544. 10.1093/bioinformatics/bth280**GeneLogic dilution data [http://www.GeneLogic.dilution.com/]**Willsky GR, Chi LH, Liang Y, Gaile DP, Hu Z, Crans DC:

**Diabetes altered gene expression in rat skeletal muscle corrected by oral administration of vanadyl sulfate (submitted).**Modrek B, Lee C:

**A genomic view of alternative splicing.***Nat Genet*2002,**30**(1):13â€“19. 10.1038/ng0102-13Benjamini Y, Hochberg Y:

**Controlling the false discovery rate: a practical and powerful approach to multiple testing.***JRSSB*1995,**57**(1):289â€“300.**BDGP Drosophila gene collection [http://www.pubmedcentral.gov/redirect3.cgi?&&reftype=extlink&artid=551536&iid=18399&jid=7&&http://www.fruitfly.org/DGC/index.html]****Affymetrix [http://www.affymetrix.com/]**Dennis GJ, Sherman BT, Hosack DA, Yang J, Gao W, Lane HC, Lempicki RA:

**DAVID: Database for Annotation, Visualization, and Integrated Discovery.***Genome Biol*2003,**4**(5):P3. 10.1186/gb-2003-4-5-p3Saltiel AR, Kahn CR:

**Insulin signalling and the regulation of glucose and lipid metabolism.***Nature*2001,**414**(6865):799â€“806. 10.1038/414799aYechoor VK, Patti ME, Saccone R, Kahn CR:

**Coordinated patterns of gene expression for substrate and energy metabolism in skeletal muscle of diabetic mice.***PNAS USA*2002,**99**(16):10587â€“10592. 10.1073/pnas.142301999Harbig J, Sprinkle R, Enkemann SA:

**A sequence-based identification of the genes detected by probesets on the Affymetrix U133 plus 2.0 array.***Nucleic Acids Res*2005,**33**(3):e31. 10.1093/nar/gni027**dChip software [http://biosun1.harvard.edu/complab/dchip/]****Bioconductor [http://www.bioconductor.org/]**

## Acknowledgements

The authors would like to thank GeneLogic Co. and S.E. Choe for the use of their data. We thank J.F. Collins for critical reading of this manuscript. We are especially grateful for M.S. Halfon for helpful discussion and comments on the manuscript. We would like to thank L-H. Chi for doing the qRT-PCR studies. Development of the diabetes microarray dataset was funded by a grant from the National Institutes of Health to D.C. Crans and G.R.Willsky.

## Author information

### Authors and Affiliations

### Corresponding author

## Additional information

### Authors' contributions

ZH initiated, designed, and carried out the research and drafted the manuscript. GRW supervised the preparation of the diabetes data set and PCR quantification, and prepared the diabetes dataset. All authors read and approved the final manuscript.

## Authorsâ€™ original submitted files for images

Below are the links to the authorsâ€™ original submitted files for images.

## Rights and permissions

**Open Access**
This article is published under license to BioMed Central Ltd. This is an Open Access article is distributed under the terms of the Creative Commons Attribution License (
https://creativecommons.org/licenses/by/2.0
), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## About this article

### Cite this article

Hu, Z., Willsky, G.R. Utilization of two sample *t*-test statistics from redundant probe sets to evaluate different probe set algorithms in GeneChip studies.
*BMC Bioinformatics* **7**, 12 (2006). https://doi.org/10.1186/1471-2105-7-12

Received:

Accepted:

Published:

DOI: https://doi.org/10.1186/1471-2105-7-12