 Software
 Open Access
 Published:
StAR: a simple tool for the statistical comparison of ROC curves
BMC Bioinformatics volume 9, Article number: 265 (2008)
Abstract
Background
As in many different areas of science and technology, most important problems in bioinformatics rely on the proper development and assessment of binary classifiers. A generalized assessment of the performance of binary classifiers is typically carried out through the analysis of their receiver operating characteristic (ROC) curves. The area under the ROC curve (AUC) constitutes a popular indicator of the performance of a binary classifier. However, the assessment of the statistical significance of the difference between any two classifiers based on this measure is not a straightforward task, since not many freely available tools exist. Most existing software is either not free, difficult to use or not easy to automate when a comparative assessment of the performance of many binary classifiers is intended. This constitutes the typical scenario for the optimization of parameters when developing new classifiers and also for their performance validation through the comparison to previous art.
Results
In this work we describe and release new software to assess the statistical significance of the observed difference between the AUCs of any two classifiers for a common task estimated from paired data or unpaired balanced data. The software is able to perform a pairwise comparison of many classifiers in a single run, without requiring any expert or advanced knowledge to use it. The software relies on a nonparametric test for the difference of the AUCs that accounts for the correlation of the ROC curves. The results are displayed graphically and can be easily customized by the user. A humanreadable report is generated and the complete data resulting from the analysis are also available for download, which can be used for further analysis with other software. The software is released as a web server that can be used in any client platform and also as a standalone application for the Linux operating system.
Conclusion
A new software for the statistical comparison of ROC curves is released here as a web server and also as standalone software for the LINUX operating system.
Background
The prediction of discrete states or categories for any event or for any object requires a classification process. In order to be useful, many realworld applications rely on an optimized classification process. These include some important problems such as diagnosis of diseases, definition of medical treatments, economical and security risk assessment, weather forecast, air traffic regulation and quality control of industrial processes [1].
Classification tasks in bioinformatics are also common and can be found in many different and relevant applications, such as the prediction of genome and protein structure [2, 3], the prediction of the cellular location [4], the prediction of molecular function [5] and the prediction of molecular interactions [6]. In general, a classification process is always involved in the prediction of a pattern that can be related to some response in living systems.
A popular approach for assessing binary classifiers is analysis of their ROC curves on a set of representative data [7, 8]. A ROC curve corresponds to a bidimensional plot of the sensitivity versus 1specificity for a given classifier with continuous or ordinal output score. Two main factors have to be considered by the user when estimating the ROC curves: 1) The design of the study. Three types of dataset can be generated when pursuing a classification task: (i) paired data, where all classifiers are applied to each individual, (ii) unpaired data, where only one of the classifiers is applied to each individual, and (iii) partiallypaired data [9], where the dataset is composed of both paired and unpaired data. In the case of paired and partiallypaired datasets, correlation between ROC curves has to be taken into consideration. Our software is primarily designed for paired data. However, it can also analyze balanced unpaired data where the number of units is the same for each classifier. It cannot be used with partiallypaired data. 2) The outcome distribution. Depending on this factor, three types of approaches can be more or less appropriate for fitting ROC curves and estimating the corresponding AUCs: (i) A parametric approach [10, 11], where we assume a parametric distribution for the outcomes of the positive and negative individuals. (ii) A semiparametric approach, where we assume that discrete ordinal outcomes correspond to classification of an unobserved latent decision variable into ordinal categories defined by unknown cutpoints or threshold values, or that continuous outcomes can be expressed as an unknown monotonic transformation of the latent distribution [12], with positive and negative individuals having different latent decision variables. In this case, parametric distributions (e.g. normal, logistic, lognormal) are assumed for the latent decision variables. (iii) A nonparametric approach [13, 14], where no distributional assumptions are made about the outcomes for the positive and negative individuals.
For each type of approach, different methods for estimating the AUC after the ROC curve is generated have been described [11, 15, 16]. The advantages and disadvantages of using one or another approach under different scenarios have been previously assessed [17, 18]. Among the advantages of using a parametric or semiparametric method are that these methods generate a smooth ROC curve, and the assumption of a distribution provides a natural means by which statistical inference such as hypothesis testing and confidence intervals can be achieved. When data deviate from the assumed distribution (e.g. normal or lognormal) or simply the outcome distribution for the positive or negative individuals is uncertain, nonparametric methods for estimating the ROC curve become a useful and robust alternative. Even though the ROC curve generated by nonparametric methods is jagged, that problem has been tackled in a nonparametric manner by means of kernel density estimation of the empirical distributions in a previous study [13].
Although many examples in the existing literature about the development of new classifiers describe the use of ROC curves and their corresponding AUCs to assess their performances, the statistical significance of their differences is often not reported. This is mostly due to the lack of freely available software that is easy to use or to automate for the pairwise comparison of many binary classifiers. Albeit there are several software for performing statistical ROC analysis [19], to the best of our knowledge, the only free and readily available software for statistical ROC analysis that assesses the significance of the difference of the AUC for a pair of classifiers is ROCKIT [20, 21]. This software uses maximum likelihood to fit a binormal ROC curve to the data and the statistical significance of the differences of a variety of indexes are assessed on the basis of a bivariate binormal model. In terms of usability, it has some drawbacks: 1) the input data format is rather cumbersome; 2) the output file contains many relevant data embedded in a humanreadable text and thus needs to be parsed for further analysis; 3) the number of classifiers that can be simultaneously assessed is quite limited; 4) additional software is needed for plotting the ROC curves; 5) in case of errors, the program does not provide any feedback to the user about the causes of the abnormal interruption; and 6) it cannot be easily automated when a fast comparison of several classifiers is required.
In this work we describe new software that is freely available as a web server tool and also as a standalone application for the Linux operating system that allows the simultaneous pairwise comparison and statistical assessment of many binary classifiers. The approach chosen is the nonparametric method for comparing AUCs based on the MannWhitney Ustatistic for comparing distributions of values from two samples [14]. It has been shown that the AUC calculated by the trapezoidal rule is equal to the MannWhitney Ustatistic applied to the outcomes for the negative and positive individuals. Thus, two or more AUCs for paired data can be statistically compared by estimating the covariance matrix for the AUCs, based on the general theory of Ustatistics, and then constructing a largesample test in the usual way. The implementation of this method, not freely available until now, provides the advantages and robustness of using a nonparametric approach for estimating AUCs in the case of paired datasets, accounting for the inherent correlation of this type of data. One limitation of this method that may be considered, as stated by DeLong et al. [14], is that the trapezoidal rule underestimates the true AUC when the variables take a small number of discrete values. Among the main features of our software are: 1) it is based on a nonparametric approach for the analysis of AUCs [14], 2) it uses a simple input format, 3) it can plot multiple ROC curves simultaneously, 4) the output data is compact, simple and can be exported for further analysis with other statistical tools; and 5) it generates a humanreadable report in PDF format, which is useful for a fast initial inspection of the results.
It is worth noting that a freely available computer program for Windows, though still in its beta version, is DBM MRMC 2.1 [22]. This software is an extended version of a previous package, LABMRMC [23–28], which allows users to compare AUCs using the jackknife method. DBM MRMC 2.1 provides, among other functionalities, statistical analysis of the AUC computed by the trapezoidal method, which is equivalent to the AUC computed with the MannWhitney Ustatistic. The program uses ANOVA methods together with jackknifing [23, 25, 26] (instead of the Delong method used by our program) to assess the statistical significance of the observed difference between two classifiers. Even though this software provides a wide range of alternatives in ROC fitting, measurement of ROC indexes and assessment of statistical significance for those indexes, DBM MRMC is still in its beta version at this moment and has the same drawbacks in terms of usability found in ROCKIT (these drawbacks are also present in LABMRMC package), such as the input/output handling, lack of automated options for fast analysis of many classifiers and lack of straightforward plotting of results.
Software Implementation
The optimal threshold (OT) for each classifier is defined after the ROC analysis is performed and consists in the score value that leads to the maximal accuracy of classification. The assessment of the statistical significance of the differences of the AUCs between two classifiers is implemented as previously described [14]. Briefly, suppose that R tests are applied on the same N individuals, which can be classified as positive or negative. Suppose that m of these individuals are actually positive and n are actually negative (m+n = N), and that positive individuals tend to have greater values than negative individuals. If we let $\left\{{X}_{i}^{r}\right\}$ and $\left\{{Y}_{j}^{r}\right\}$ be the sets of outcome values on the rth test that correspond to the positive and negative individuals, respectively (i = 1,...,m; j = 1,...,n; 1 ≤ r ≤ R), the AUC for each classifier is computed with the MannWhitney Ustatistic for comparing distributions of values from two samples, as follows:
The theory on generalized Ustatistics allows us to obtain an estimated covariance matrix for two or more AUC estimates of correlated ROC curves; this R × R matrix is computed as follows:
where the (r_{1}, r_{2})th element of S_{10} is given by
and similarly
For further explanation on how to get these estimates, please refer to reference [14]. Accounting for correlation between ROC curves is a necessary step for paired data where two classifiers are used on the same subjects. For balanced unpaired data the offdiagonal elements of S are set to zero since the AUCs are not correlated. Consequently, the covariance matrix S is used to compute the following chisquared statistic for testing if there is a difference between two or more classifiers:
Here, $\widehat{\theta}$ is the vector of AUC estimates and L is a suitable contrast matrix (ie. H0: Lθ = 0, where 0 is the zero matrix). The statistic follows a chisquared distribution with rank(L) degrees of freedom under the null hypothesis of no difference between classifier AUCs. For a pair of classifiers the statistic reduces to
and a (1  α)100% confidence interval is given by
This particular software implementation and its successful application have been recently validated by us through the comparison to other software [29–31]. In these studies hundreds of classifiers for the prediction of errors in protein structures were assessed, and the results of statistical significance obtained with StAR software were consistent with those from ROCKIT.
Software Description
The input of the software simply consists of two data files containing the positive and negative subjects, as defined by the user. It is important to be aware of the definition used for the negative and positive data, since the meaning or interpretation of false positives and true positives reported will depend on this definition. Each input file must have a multicolumn format, where a given column contains the obtained scores from a specific classifier for all subjects tested. Additional input parameters that do not affect the results of the calculations are optional and include a job name and the possibility of getting the display of the classifiers sorted by decreasing AUC values, among others. Detailed online help about the required format for the input files is provided.
An initial summarized report with the results of the calculations for each classifier is given in a table that contains a variable number of columns, which correspond to the following in the most extended output case: 1) Sequential number of each classifier; 2) selection option of each classifier to perform further analysis (by default, all classifiers are selected); 3) description name or identification code of each classifier; 4) AUC of each classifier; 5) a plus sign ('+') indicating if the classifier score has been inverted in order to force the AUC greater or equal than 0.5, 6) maximal accuracy of each classifier (ie. obtained at an optimal classification threshold that is estimated after the ROC analysis); 7) optimal classification threshold (ie. the score value that, when used as a classification threshold, leads to the maximal accuracy); 8) false positive rate obtained at the optimal classification threshold; 9) true positive rate obtained at the optimal classification threshold; 10) total number of negative subjects evaluated; and 11) total number of positive subjects evaluated. Online description is provided for each field in this table.
Additionally, six additional actions on the provided data are available for further analysis. First, the user can plot the ROC curves for the selected classifiers. Some graphic display options or changes to the plots are available. Second, the estimated covariance matrix and the pvalue of the global test for a difference between any of the classifiers is displayed. Third, the difference of any two classifiers provided can be assessed at a given significance level, which by default is set to 0.05, but it can be modified by the user. Fourth, for each pairwise comparison of the classifiers, the software reports the confidence intervals at a given confidence coefficient. Fifth, a humanreadable report in PDF format that summarizes the results of the analysis can be generated. Finally, several files containing the detailed results from the analysis performed by the user at the selected significance level can be downloaded. These include the ROC plot points for each classifier, the estimated covariance matrix, a table containing the pvalue and confidence interval of the AUC difference observed for each pairwise comparison of classifiers with color coding of the pvalue used to indicate if the difference was significant.
The standalone version of this software is also released for the Linux operating system. The Linux version offers the same capabilities of the web server, but without the graphic display and the interactive options. A detailed tutorial that describes how to use the software is available at the server web site.
Availability and Requirements
Project name: StAR: Statistical Comparison of ROC Curves.
Project homepage: http://protein.bio.puc.cl/star.html
Operating system(s): any (web server version), Linux (standalone version).
Programming language: C++, PHP, PERL.
Other requirements: none
License: none
Any restrictions to use by nonacademics: none
Abbreviations
 ROC:

Receiver operating characteristic
 AUC:

Area under the ROC curve
 OT:

Optimal threshold
 PDF:

Portable document format
 StAR:

Statistical analysis of ROC curves.
References
 1.
Swets JA, Dawes RM, Monahan J: Better decisions through science. Sci Am 2000, 283(4):82–87.
 2.
Usuka J, Brendel V: Gene structure prediction next term by spliced alignment of genomic DNA with protein sequences: increased accuracy by differential splice site scoring. Journal of Molecular Biology 2000, 297: 1075–1085. 10.1006/jmbi.2000.3641
 3.
Orengo CA, Jones DT, Thornton JM: Protein superfamilies and domain superfolds. Nature 1994, 372(6507):631–634. 10.1038/372631a0
 4.
Chou KC, Elrod DW: Protein subcellular location prediction. Protein Engineering 1999, 12: 107–118. 10.1093/protein/12.2.107
 5.
Ashburner M, Ball CA, Blake JA, Botstein D, Butler H, Cherry JM, Davis AP, Dolinski K, Dwight SS, Eppig JT, Harris MA, Hill DP, IsselTarver L, Kasarskis A, Lewis S, Matese JC, Richardson JE, Ringwald M, Rubin GM, Sherlock G: Gene ontology: tool for the unification of biology. The Gene Ontology Consortium. Nat Genet 2000, 25(1):25–29. 10.1038/75556
 6.
Vazquez A, Flammini A, Maritan A, Vespignani A: Global protein function prediction from proteinprotein interaction networks. Nature Biotechnology 2003, 21: 697–700. 10.1038/nbt825
 7.
Fawcett T: ROC Graphs: Notes and Practical Considerations for Researchers. Tech Report HPL2003–4, HP Laboratories 2004.
 8.
Swets JA: Measuring the accuracy of diagnostic systems. Science 1988, 240(4857):1285–1293. 10.1126/science.3287615
 9.
Metz CE, Herman BA, Roe CA: Statistical Comparison of Two ROCcurve Estimates Obtained from Partiallypaired Datasets. Medical Decision Making 1998, 18(1):110–121. 10.1177/0272989X9801800118
 10.
Hanley JA: The use of the binormal model for parametric ROC analysis of quantitative diagnostic tests. Medical Decision Making 1988, 8: 197–203. 10.1177/0272989X8800800308
 11.
Metz CE: Basic Principles of ROC analysis. Semin nucl med 1978, 8: 283–298. 10.1016/S00012998(78)800142
 12.
Metz CE, Herman BA, Shen JH: Maximum likelihood estimation of receiver operating characteristic (ROC) curves from continuouslydistributed data. Statistics in Medicine 1998, 17(9):1033–1053. 10.1002/(SICI)10970258(19980515)17:9<1033::AIDSIM784>3.0.CO;2Z
 13.
Zou KH, Hall WJ, Shapiro DE: Smooth nonparametric receiver operating characteristic (ROC) curves for continuous diagnostic tests. Statistics in Medicine 1996, 16(19):2143–2156. Publisher Full Text 10.1002/(SICI)10970258(19971015)16:19%3C;2143::AIDSIM655%3E;3.0.CO;23
 14.
Delong ER, Delong DM, ClarkePearson DL: Comparing the Areas Under Two or More Correlated Receiver Operating Characteristic Curves: A Nonparametric Approach. Biometrics 1988, 44(3):837–845. 10.2307/2531595
 15.
Dorfman DD, Alf E: Maximum likelihood estimation of parameters of signal detection theory and determination of confidence intervals  rating  method data. Journal of Mathematical Psychology 1969, 6: 487–496. 10.1016/00222496(69)900194
 16.
Bamber D: The area above the ordinal dominance graph and the area below the receiver operating characteristic graph. Journal of Mathematical Psychology 1975, 12: 387–415. 10.1016/00222496(75)900012
 17.
HajianTilaki KO, Hanley JA, Joseph L, Collet JP: A Comparison of Parametric and Nonparametric Approaches to ROC Analysis of Quantitative Diagnostic Tests. Medical Decision Making 1997, 17(1):94–102. 10.1177/0272989X9701700111
 18.
Goddard MJ, Hinberg I: Receiver operator characteristic (ROC) curves and nonnormal data: An empirical study. Statistics in Medicine 1989, 9(3):325–337. 10.1002/sim.4780090315
 19.
Stephan C, Wesseling S, Schink T, Jung K: Comparison of eight computer programs for receiveroperating characteristic analysis. Clin Chem 2003, 49(3):433–439. 10.1373/49.3.433
 20.
Metz CE: Statistical analysis of ROC data in evaluating diagnostic performance. In Multiple regression analysis: applications in the health sciences (D Herbert and R Myers, eds). New York: American Institute of Physics ; 1986:365.
 21.
Metz CE: A new approach for testing the significance of differences between ROC curves measured from correlated data. In Information processing in medical imaging (Ed F Deconinck). Nijhoff, The Hague. ; 1984:432–445.
 22.
DBM MRMC 2.1[http://perception.radiology.uiowa.edu]
 23.
Dorfman DD, Berbaum KS, Metz CE: Receiver operating characteristic rating analysis. Generalization to the population of readers and patients with the jacknife method. Invest Radiol 1992, 27: 723–731. 10.1097/0000442419920900000015
 24.
Dorfman DD, Metz CE: Multireader multicase ROC analysis: comments on Begg’s commentary. Academic Radiol 1995, 2(Supplement 1):S76.
 25.
Hillis SL, Berbaum KS: Montecarlo validation of the DorfmanBerbaumMetz method using normalized pseudovalues and less databased model simplification. Academic radiology 2005, 12: 1534–1541. 10.1016/j.acra.2005.07.012
 26.
Hillis SL, Obuchowski NA, Schartz KM, Berbaum KS: A comparison of the DorfmanBerbaumMetz and ObuchowskiRockette methods for receiver operating characteristic (ROC) data. Statistics in Medicine 2005, 24: 1579–1607. 10.1002/sim.2024
 27.
Roe CA, Metz CE: DorfmanBerbaumMetz method for statistical analysis of multireader, multimodality receiver operating characteristic data: validation with computer simulation. Academic radiology 1997, 4(4):298–303. 10.1016/S10766332(97)800323
 28.
Roe CA, Metz CE: Variancecomponent modeling in the analysis of receiver operating characteristic index estimates. Academic radiology 1997, 4(8):587–600. 10.1016/S10766332(97)802103
 29.
Ferrada E, Melo F: Nonbonded terms extrapolated from nonlocal knowledge based energy functions improve error detection in near native protein structure models. Protein Science 2007, 16: 1410–1421. 10.1110/ps.062735907
 30.
Ferrada E, Vergara IA, Melo F: A knowledgebased potential with an accurate description of local interactions improves discrimination between native and nearnative protein conformations. Cell Biochemistry and Biophysics 2007, 49: 111–124. 10.1007/s1201300700505
 31.
Melo F, Sali A: Fold assessment for comparative protein structure modeling. Protein Science 2007, 16: 2412–2426. 10.1110/ps.072895107
Acknowledgements
We are truly grateful of the helpful comments and suggestions made by the anonymous reviewers of this manuscript, which in our opinion have substantially improved its clarity of presentation. This work was funded by grant 1051112 from FONDECYT.
Author information
Affiliations
Corresponding author
Additional information
Authors' contributions
IAV and TN developed and implemented the core computer programs. EF and AWS wrote some additional scripts and tested the software. FM supervised this project and wrote the manuscript with the help of IAV. All authors read and approved the final version.
Ismael A Vergara, Tomás Norambuena contributed equally to this work.
Rights and permissions
About this article
Cite this article
Vergara, I.A., Norambuena, T., Ferrada, E. et al. StAR: a simple tool for the statistical comparison of ROC curves. BMC Bioinformatics 9, 265 (2008). https://doi.org/10.1186/147121059265
Received:
Accepted:
Published:
Keywords
 Receiver Operating Characteristic
 Receiver Operating Characteristic Curve
 Binary Classifier
 Estimate Covariance Matrix
 Negative Individual