Volume 11 Supplement 9

## Selected Proceedings of the 2010 AMIA Summit on Translational Bioinformatics

- Proceedings
- Open Access

# Comparison of multiplex meta analysis techniques for understanding the acute rejection of solid organ transplants

- Alexander A Morgan
^{1, 2}, - Purvesh Khatri
^{1, 2, 3}, - Richard Hayden Jones
^{1, 2}, - Minnie M Sarwal
^{1, 3}and - Atul J Butte
^{1, 2, 3}

**11 (Suppl 9)**:S6

https://doi.org/10.1186/1471-2105-11-S9-S6

© Morgan et al; licensee BioMed Central Ltd. 2010

**Published:**28 October 2010

## Abstract

### Background

Combining the results of studies using highly parallelized measurements of gene expression such as microarrays and RNAseq offer unique challenges in meta analysis. Motivated by a need for a deeper understanding of organ transplant rejection, we combine the data from five separate studies to compare acute rejection versus stability after solid organ transplantation, and use this data to examine approaches to multiplex meta analysis.

### Results

We demonstrate that a commonly used parametric effect size estimate approach and a commonly used non-parametric method give very different results in prioritizing genes. The parametric method providing a meta effect estimate was superior at ranking genes based on our gold-standard of identifying immune response genes in the transplant rejection datasets.

### Conclusion

Different methods of multiplex analysis can give substantially different results. The method which is best for any given application will likely depend on the particular domain, and it remains for future work to see if any one method is consistently better at identifying important biological signal across gene expression experiments.

## Keywords

- Acute Rejection
- Meta Analysis
- Solid Organ Transplant
- Transplant Rejection
- Effect Size Estimate

## Background

Messenger RNA expression measurement with microarrays enables a very highly multiplexed examination of the relative expression of genes under different experimental conditions, and several different techniques for combining the results of different microarray experiments have been proposed [1–5]. However, there has been very little work done looking at the differences between meta analysis techniques [6] and how they may be used together to understand a concrete biomedical problem with clinical implications. By looking at a particular biomedical problem, in this case acute rejection of solid organ transplants, we can compare the results of different meta analysis approaches, both to one another and what is known about the process of transplant rejection.

In 2007, more than 28,000 solid organs (including heart, liver, lung, pancreas, kidney and intestine) were transplanted in the United States [7]. Unfortunately, recipients of allografts are subject to the problems of transplant rejection. Particularly dire is the activation of a significant T-cell mediated response against the implanted tissue, termed acute rejection. Therefore, we can expect that samples taken from patients undergoing acute rejection will have substantially increased expression of genes associated with a T-cell mediated immune response than samples taken from patients not undergoing acute rejection. However, the mechanism by which T-cells infiltrate the allografts and mediate rejection is not known. It can be hypothesized that although the prompts for tissue-specific injury may be different, there may be common mechanism of rejection (e.g., T-cell infiltration) across all solid organ transplants. Identification of shared mechanisms that lead to tissue-specific destruction can facilitate novel treatments without requiring the understanding of individual, tissue-specific mechanisms. However, the conventional experiment design, where only one type of solid organ allograft is examined in an experiment is inadequate to identify such a mechanism. In this regard, microarray data from different solid organ transplants, available from public data repositories such as NCBI Gene Expression Omnibus (GEO), can be combined to gain a greater understanding of the process of acute rejection to aid in early diagnosis and the tailoring of treatment regimens.

However, these experiments are performed on different microarray platforms by different laboratories, Several techniques for combining the results of different microarray experiments have been proposed and the theoretical differences between methods have been presented and a range of methods have been covered in review articles [1–5]. However, there has been very little work done looking at the empirical differences between meta analysis techniques. Furthermore, most previous work has used simulated data or measures of consistency [6]. We are most interested in how meta-techniques may be used together with real data to understand a concrete biomedical problem with clinical implications.

Multiplexed meta analysis differs from traditional, single measurement, approaches in the number of variables being simultaneously combined across experiments. In traditional meta analysis, a given variable is combined across several experiments which may be measuring the exactly the same feature or slightly different features that will be combined, and there is an extensive set of techniques for this form of analysis [8]. However, in multiplex meta analysis, many variables are being simultaneously combined across experiments. Some of the experiments may not have measured that particular variable, and some of the experiments may have measured that variable multiple times within a single experiment (such as multiple probes measuring expression of the same gene on a microarray).

In single measurement approaches, it is often important to identify whether or not the experiments combined are comparable in design and consistent in result. However, in multiplexed meta analysis, we are often trying to identify which variables represent a core of similar behavior across experiments and it is not essential that the other variables are homogeneous in variation.

Finally, multiplexed meta analysis is often used for data-mining purposes to identify particularly strongly associated variables across experiments, such as the development of biomarkers and other forms of hypothesis generation that will be validated experimentally. It is therefore often acceptable to sacrifice significant amounts of sensitivity for a high degree of specificity when analyzing tens of thousands of variables.

In this work, we focus on the application of some of the traditional approaches of meta analysis applied to multiplex gene expression data. We show that for the very basic task of developing a prioritized gene list, our applied parametric approach with known flaws actually outperforms non-parametric methods that should have greater accuracy at the level of any individual gene. For multiplex problems, the large error bars associated with the estimation of meta fold-change for any particular gene is less important than the overall average improvement in gene ranking that this meta analysis method provides.

## Results

Traditional meta analysis methods may be grouped into two broad categories of approach. The first is to use a non-parametric method to combine the significance results of the different experiments being amalgamated in the meta analysis. At the simplest level this can include counting methods that look at the proportion of experiments showing significance. More accurate tests include omnibus methods that combine p-values to obtain a meta p-value. The second general approach is to try to merge a series of predictions from separate experiments to develop a model of effect size. For gene expression this would amount to a meta fold-change estimate. We examined the most commonly used examples of both approaches.

*p*

_{ i }], from each of the

*k*studies and compares that to a χ

^{2}distribution:

The resulting list of meta p-values, one for each gene, can then be subject to correction for multiple hypothesis testing or used to estimate a false discovery rate and obtain a q-value [9].

As an underlying statistical test, we used two one sided t-tests, one for a hypothesis of increased expression, another for a hypothesis of decreased expression. Because Fisher's method does not take into account the sign of the variation, we examine the hypothesis of increased expression separately from decreased expression, and then for each gene, take the minimum p-value. Taking the minimum of two values introduces a very slight bias, but our primary interest is in creating a ranked list. For the combination of these five microarray studies, we get a substantially enriched selection of differentially expressed genes over random; 5.2% of the 20,113 genes studied have a meta p-value less than 0.01 for the hypothesis of increased expression (1% expected by chance), whereas 6.5% have p-values less than 0.01 for the hypothesis of decreased expression (1% also expected by chance). When the minimum p-value of the Fisher's method applied to the two one-sided t-tests was corrected for multiple hypothesis testing through the method proposed by Benjamini and Hochberg [10], 584 genes had a corrected p-value less than 0.01; this suggests a substantial enrichment for true positives.

*f*

_{ i }), weighted by the variance in the effect size within each study (

*w*

_{ i }), with the confidence intervals combined with the same weights. This means that studies with larger intra-study variation (noise) contribute less to the overall estimate of fold-change.

Unfortunately, this approach makes a strong assumption about the distribution of the data and how it may be combined. For multiplexed meta analysis, there are many variables each with possibly very different distributions, and it is very difficult, if not impossible, to identify any single best distribution to model the data, and there are generally insufficient samples to develop separate models for each variable (gene). That means that this is neither a robust nor a powerful statistical test for each variable.

The non-parametric methods of p-value combination make no assumptions about the underlying distribution of the data, and rely only on the results of the underlying statistical methods used to analyze that data. Methods such as the t-test are very robust and accurate when looking at gene expression measurements. Indeed, they are more accurate at detecting differential expression than fold-change [11]. However, the fold-change is still an important measure of effect size, and it is substantially more reproducible across studies [12]. For instance, if gene expression studies are being combined to suggest proteins for further analysis, a very small gene expression fold-change is not likely to lead to a measurable difference in protein expression, independent of the statistical significance of the difference.

The difference between significance and effect size is an important distinction in applied statistics in general. Although, statistical significance is often associated with large effect size, they are not equivalent as demonstrated in the commonly used 'volcano plot' to display gene expression data [13]. For another example, a correlation coefficient is a measure of effect size, whereas a p-value on a measure of correlation is a significance measure. When doing most types of analysis, we want both a significance estimate and an effect size estimate. In multiplexed meta analysis, we suggest combining a robust measure of significance (such as a t-test or modified t-test) using one of the non-parametric omnibus methods to identify significant difference with a parametric method of combining effect size estimates to get a reasonable estimate of effect size (i.e. fold-change).

The meta p-values obtained by Fisher's method and the meta-fold-change value obtained by the invariance method can be plotted for each gene, leading to a meta analysis version of the volcano plot [14]. A typical volcano plot graphs p-value against fold-change and shows that extremely high and extremely low fold-changes tend to be associated with highly significance p-values, and has a marked triangle shape.

## Discussion

It is important to evaluate the ability of different approaches to multiplex meta analysis to answer the questions being asked. Unfortunately, just as there is no single question asked of gene expression measurements, there is no one correct way to evaluate different gene expression analysis techniques, even with extensive biological validation.

In this pilot application, we were interested in obtaining a prioritized list of the genes involved in acute rejection across the types of transplanted organs. We have a strong prior expectation that this common process of acute rejection involves an immune response reaction. We searched for genes known to be involved in immune response (annotated as “immune response” genes), as reported by the Molecular Signatures Database [15], and observe how the different meta analysis approaches prioritize those genes. Our method of evaluation here was to count the number of immune related genes at each point of depth into the ranked list.

We find the meta fold-change derived by the inverse variance method is better at recapitulating our prior expectation of immune response genes than any single dataset, at every depth into the gene list (Figure 2). Interestingly, the list of genes obtained by ranking according to the meta p-values obtained by Fisher's method are no better than the best individual experimental results at identifying and prioritizing the genes of interest. Though we acknowledge this is only a preliminary analysis, these results suggest that the ordering of genes provided by meta effect size estimate may be more biologically relevant, even if non-parametric methods of significance may be more accurate for any individual gene.

## Conclusions

Our results show that there is value to be gained in doing meta analysis and combining results from different studies can improve the ranking of genes. However, not all meta analysis techniques are equivalent, and not all seem to provide the same level of improvement over looking at individual experiments.

Effective, useful multiplex analysis techniques for gene expression analysis must be able to provide accurate estimates of statistical significance of differential expression. However, we will continue to need general, overall estimates of effect size (fold-change). Indeed, our results suggest that for at least some biological questions these estimates of fold-change are more useful for identifying features important to the biology of the problem than meta significance estimates (p-values).

The study of gene expression measurements in acute rejection offers a particularly good opportunity to investigate methods of multiplex meta analysis. We know that a powerful immune response is involved, and the fact that the meta fold-change approach provides a greater enrichment of these genes implies that it is capturing more of the core, shared process of acute rejection across organ types. At the same time that we generally know that acute rejection involves an immune response, we are interested in the particular genes and pathways involved and can investigate the highly ranked genes more thoroughly and look for biological validation to help address the important clinical problem of organ rejection.

In this comparison of two approaches to multiplex meta analysis, we have identified some of key issues that need to be investigated further. The fact that two of the most common approaches to traditional meta analysis, a nonparametric significance test based approach and a parametric, effect size estimate approach, give very different results highlights many of the challenges. This is an opportunity for future efforts, as we can investigate methods that combine both of these approaches. It will also be important to identify in future work if one of these two methods or another approach is able to consistently identify important biologically relevant signals across gene expression experiments. The evaluation we have chosen, the prioritization of immune relevant genes in transplant rejection expression variation datasets is far from a perfect evaluation metric. Future comparisons will need to look at other biomedically important problems and gold-standards to identify if any one approach of family approaches can be shown to provide better results than another.

## Materials and methods

Overview of the datasets.

Dataset | Organ | AR | Tol | Platform | GEO Acc. |
---|---|---|---|---|---|

Gimino, et al. | Lung | 7 | 27 | hg133a | GDS999 |

Karason, et al. | Heart | 3 | 6 | hg133a | GDS1684 |

Patil, et al. | Lung | 18 | 14 | hg133a | GSE6095 |

Flechner, et al. | Kidney | 7 | 10 | hgu95a | GDS724 |

Rödder, et al. | Kidney | 18 | 8 | hg133plu2 | GSE9493 |

The individual data sets were taken from the Gene Expression Omnibus [21] and the data was quantile-quantile normalized and analyzed using Bioconductor [22]. Two one sided t-tests were performed for each probeset within each of the five constituent datasets. Fisher's method was then performed as previously described [8] by pooling all the p-values for the probesets for each gene. Probes were mapped to Entrez Gene identifiers [23] using AILUN [24]. The minimum of the two p-values (one from each of the two one-sided tests which were synthesized using Fisher's method) for each gene was taken as the meta p-value. Within each experiment, a fold change was calculated for each gene by taking the log ratio of the geometric means of the expression values between the samples showing acute rejection and those showing stable acceptance of the transplanted organ. The geometric mean of expression value is equivalent to taking the arithmetic means of the log expression values. The sample variance was also calculated for this dimensionless quantity derived from fold change (a ratio), and then the fold changes were combined for each gene using the previously described linear model, weighted by inverse variance, to provide an effect size estimate, in this case a meta fold-change. A standard deviation on this effect size estimate was calculated using a fixed effects model [8].

Gene signatures for "Immune Response" and "Defense Response" were taken from the Molecular Signatures Database [15], and mapped to our genes through their gene symbols. These signatures were compared to the rankings provided by the two meta analysis methods, and ROC curves were produced using the ROCR package [25]. The multtest package [26] was used for correction of the p-values for multiple hypothesis testing.

## Declarations

### Acknowledgements

The work was supported by the Lucile Packard Foundation for Children's Health, National Library of Medicine (R01 LM009719 and T15 LM7033), the National Institute of Allergy and Infectious Diseases (R01 AI061739 and U01 AI055795), and the Stanford MedScholars Program. We thank Alex Skrenchuk, Dr. Boris Oskotsky, and Dr. Rong Chen from Stanford University for computer and database support.

This article has been published as part of *BMC Bioinformatics* Volume 11 Supplement 9, 2010: Selected Proceedings of the 2010 AMIA Summit on Translational Bioinformatics. The full contents of the supplement are available online at http://www.biomedcentral.com/1471-2105/11?issue=S9.

## Authors’ Affiliations

## References

- Choi JK, Yu U, Kim S, Yoo OJ:
**Combining multiple microarray studies and modeling interstudy variation.***Bioinformatics*2003,**19**(Suppl 1):i84–90. 10.1093/bioinformatics/btg1010View ArticlePubMedGoogle Scholar - DeConde RP, Hawley S, Falton S, Clegg N, Knudsen B, Etzioni R: "Combining Results of Microarray Experiments: Rank Aggregation Approach". Statistical Applications in Genetics and Molecular Biology 2006., 5(1): 10.2202/1544-6115.1204Google Scholar
- English SB, Butte AJ:
**Evaluation and integration of 49 genome-wide experiments and the prediction of previously unknown obesity-related genes.***Bioinformatics*2007,**23**(21):2910–2917. 10.1093/bioinformatics/btm483PubMed CentralView ArticlePubMedGoogle Scholar - Ramasamy A, Mondry A, Holmes CC, Altman DG:
**Key issues in conducting a meta-analysis of gene expression microarray datasets.***PLoS Med*2008,**5**(9):e184. 10.1371/journal.pmed.0050184PubMed CentralView ArticlePubMedGoogle Scholar - Rhodes DR, Barrette TR, Rubin MA, Ghost D, Chinnaaiyan AM:
**Meta-Analysis of Microarrays: Interstudy Validation of Gene Expression Profiles Reveals Pathway Dysregulation in Prostate Cancer.***Cancer Res*2002,**62:**4427–4433.PubMedGoogle Scholar - Hong F, Breitling R:
**A comparison of meta-analysis methods for detecting differentially expressed genes in microarray experiments.***Bioinformatics*2008,**24**(3):374–382. 10.1093/bioinformatics/btm620View ArticlePubMedGoogle Scholar **Number of U.S. Transplants Per Year, 1988–2008**[http://www.infoplease.com/science/health/us-transplants-year-1988–2007.html]- Hedges L, Olkin I:
**Statistical methods for meta-analysis.**1985.Google Scholar - Storey JD, Tibshirani R:
**Statistical significance for genomewide studies.***Proc Natl Acad Sci U S A*2003,**100**(16):9440–9445. 10.1073/pnas.1530509100PubMed CentralView ArticlePubMedGoogle Scholar - Benjamini Y, Hochberg Y:
**Controlling the false discovery rate: a practical and powerful approach to multiple testing.***Journal of the Royal Statistical Society Series B*1985,**57:**289–300.Google Scholar - Witten D, Tibshirani R:
*A comparison of fold-change and the t-statistic for microarray data analysis.*Stanford University; 2007.Google Scholar - Shi L, al e:
**The MicroArray Quality Control (MAQC) project shows inter- and intraplatform reproducibility of gene expression measurements.***Nat Biotechnol*2006,**24**(9):1151–1161. 10.1038/nbt1239View ArticlePubMedGoogle Scholar - Cui X, Churchill G:
**Statistical tests for differential expression in cDNA microarray experiments.***Genome biology*2003,**4**(4):210. 10.1186/gb-2003-4-4-210PubMed CentralView ArticlePubMedGoogle Scholar - Cui X, Churchill GA:
**Statistical tests for differential expression in cDNA microarray experiments.***Genome Biol*2003,**4**(4):210. 10.1186/gb-2003-4-4-210PubMed CentralView ArticlePubMedGoogle Scholar - Subramanian A, Tamayo P, Mootha VK, Mukherjee S, Ebert BL, Gillette MA, Paulovich A, Pomeroy SL, Golub TR, Lander ES,
*et al*.:**Gene set enrichment analysis: a knowledge-based approach for interpreting genome-wide expression profiles.***Proc Natl Acad Sci U S A*2005,**102**(43):15545–15550. 10.1073/pnas.0506580102PubMed CentralView ArticlePubMedGoogle Scholar - Rodder S, Scherer A, Raulf F, Berthier CC, Hertig A, Couzi L, Durrbach A, Rondeau E, Marti HP:
**Renal allografts with IF/TA display distinct expression profiles of metzincins and related genes.***Am J Transplant*2009,**9**(3):517–526. 10.1111/j.1600-6143.2008.02512.xView ArticlePubMedGoogle Scholar - Flechner SM, Kurian SM, Head SR, Sharp SM, Whisenant TC, Zhang J, Chismar JD, Horvath S, Mondala T, Gilmartin T,
*et al*.:**Kidney transplant rejection and tissue injury by gene profiling of biopsies and peripheral blood lymphocytes.***Am J Transplant*2004,**4**(9):1475–1489. 10.1111/j.1600-6143.2004.00526.xPubMed CentralView ArticlePubMedGoogle Scholar - Gimino VJ, Lande JD, Berryman TR, King RA, Hertz MI:
**Gene expression profiling of bronchoalveolar lavage cells in acute lung rejection.***Am J Respir Crit Care Med*2003,**168**(10):1237–1242. 10.1164/rccm.200305-644OCView ArticlePubMedGoogle Scholar - Patil J, Lande JD, Li N, Berryman TR, King RA, Hertz MI:
**Bronchoalveolar lavage cell gene expression in acute lung rejection: development of a diagnostic classifier.***Transplantation*2008,**85**(2):224–231. 10.1097/TP.0b013e318160268aView ArticlePubMedGoogle Scholar - Karason K, Jernas M, Hagg DA, Svensson PA:
**Evaluation of CXCL9 and CXCL10 as circulating biomarkers of human cardiac allograft rejection.***BMC Cardiovasc Disord*2006,**6:**29. 10.1186/1471-2261-6-29PubMed CentralView ArticlePubMedGoogle Scholar - Barrett T, Troup DB, Wilhite SE, Ledoux P, Rudnev D, Evangelista C, Kim IF, Soboleva A, Tomashevsky M, Edgar R:
**NCBI GEO: mining tens of millions of expression profiles--database and tools update.***Nucleic Acids Res*2007,**35**(Database issue):D760–765. 10.1093/nar/gkl887PubMed CentralView ArticlePubMedGoogle Scholar - Gentleman R, Carey VJ, Huber W, Irizarry RA, Dudoit S:
*Bioinformatics and Computational Biology Solutions Using R and Bioconductor.*New York: Springer-Verlag; 2005.View ArticleGoogle Scholar - Wheeler DL, Barrett T, Benson DA, Bryant SH, Canese K, Chetvernin V, Church DM, DiCuccio M, Edgar R, Federhen S,
*et al*.:**Database resources of the National Center for Biotechnology Information.***Nucleic Acids Res*2007,**35**(Database issue):D5–12. 10.1093/nar/gkl1031PubMed CentralView ArticlePubMedGoogle Scholar - Chen R, Li L, Butte AJ:
**AILUN: reannotating gene expression data automatically.***Nat Methods*2007,**4**(11):879. 10.1038/nmeth1107-879PubMed CentralView ArticlePubMedGoogle Scholar - Sing T, Sander O, Beerenwinkel N, Lengauer T:
**ROCR: visualizing classifier performance in R.***Bioinformatics*2005,**21**(20):3940–3941. 10.1093/bioinformatics/bti623View ArticlePubMedGoogle Scholar - Pollard KS, Dudoit S, van der Laan MJ:
**Multiple Testing Procedures: the multtest Package and Applications to Genomics.**In*Bioinformatics and Computational Biology Solutions Using R and Bioconductor*. Edited by: Gentleman R, Carey VJ, Huber W, Irizarry RA, Dudoit S. New York: Springer; 2005.Google Scholar - Carr D, Littlefield R, Nicholson W, Littlefield J:
**Scatterplot matrix techniques for large N.***Journal of the American Statistical Association*1987, 424–436. 10.2307/2289444Google Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.