 Research article
 Open Access
 Published:
A framework for significance analysis of gene expression data using dimension reduction methods
BMC Bioinformatics volume 8, Article number: 346 (2007)
Abstract
Background
The most popular methods for significance analysis on microarray data are well suited to find genes differentially expressed across predefined categories. However, identification of features that correlate with continuous dependent variables is more difficult using these methods, and long lists of significant genes returned are not easily probed for coregulations and dependencies. Dimension reduction methods are much used in the microarray literature for classification or for obtaining lowdimensional representations of data sets. These methods have an additional interpretation strength that is often not fully exploited when expression data are analysed. In addition, significance analysis may be performed directly on the model parameters to find genes that are important for any number of categorical or continuous responses. We introduce a general scheme for analysis of expression data that combines significance testing with the interpretative advantages of the dimension reduction methods. This approach is applicable both for explorative analysis and for classification and regression problems.
Results
Three public data sets are analysed. One is used for classification, one contains spikedin transcripts of known concentrations, and one represents a regression problem with several measured responses. Modelbased significance analysis is performed using a modified version of Hotelling's T^{2}test, and a false discovery rate significance level is estimated by resampling. Our results show that underlying biological phenomena and unknown relationships in the data can be detected by a simple visual interpretation of the model parameters. It is also found that measured phenotypic responses may model the expression data more accurately than if the designparameters are used as input. For the classification data, our method finds much the same genes as the standard methods, in addition to some extra which are shown to be biologically relevant. The list of spikedin genes is also reproduced with high accuracy.
Conclusion
The dimension reduction methods are versatile tools that may also be used for significance testing. Visual inspection of model components is useful for interpretation, and the methodology is the same whether the goal is classification, prediction of responses, feature selection or exploration of a data set. The presented framework is conceptually and algorithmically simple, and a Matlab toolbox (Mathworks Inc, USA) is supplemented.
Background
An important part of expression data analysis is the significance analysis of genes, in which relevant features are found among thousands of irrelevant ones. Two widely used algorithms for significance analysis are SAM [1] and Limma [2]. These methods are well suited for assessing differences of expression across classes or levels in an experimental design. Dimension reduction methods such as principal component analysis (PCA) and partial least squares regression (PLSR) [3–5] is another group of popular tools for expression data analysis. These methods attempt to describe the underlying variation of the data in a lowdimensional space, enabling the use of statistical methods that cannot deal with the high number of correlated geneexpressions in the original data. The real signal is separated from random noise, and the components of the dimension reduction can often be interpreted biologically.
A much used dimension reduction method is PCA [3, 4]. This is an unsupervised method that describes the main features of a data set using a small number of orthogonal components. The computational method of singular value decomposition (SVD) is in many aspects the same as PCA, and often it constitutes the main part of the PCA algorithm. When background knowledge is available about the objects, supervised methods may be used to identify genes that associate with this information. For instance, ttests may be used to discriminate between classes in an experimental design. Also, correlation tests such as Spearman's correlation may be used to probe expression profiles for similarity with continuous responses. PLSR [3, 5] is a very powerful dimension reduction method which can utilise any sort of background information directly in the decomposition. PLSR components are designed to optimally describe the dependent variables in terms of the expression data. The main focus of PLSR in the microarray literature has been on classification, where a regression model is trained to fit a binary array representing class labels. Many methods are available to convert the prediction model to a classification rule, for instance by discriminant analysis of the PLSR components [6, 7]. We use the intuitively pleasing property that the predicted, validated responses may be discretised directly in order to predict classassignments. When PLSR is used for classification it is referred to as discriminant PLSR (DPLSR) or PLS discriminant analysis (PLSDA). Nguyen and Rocke were the first to demonstrate how DPLSR may be applied to classify cancers based on expression data [6, 7]. In that work, they used an additional discriminant analysis for the final classification. PLSR has also successfully been used to model patient survival times [8] or cell cycle data [9], which are applications where a continuous response is used directly as input. Overviews of dimension reduction methods for analysis of microarray data are given in [10, 11].
The dimension reduction methods offer a new way of doing significance analysis on expression data. The underlying variation of interest is described by a limited number of components, also known as latent variables (LVs). The LVs are axes in a coordinate system that span, depending on the dimension reduction method used, the most relevant structures in the data set. Each object (array) is associated with a limited number of scores, which represent the projection of the measurements onto the latent axes. There is also a loading vector for each component, providing a mapping between the latent and the original variables (features). The score vectors thus span interesting directions in the original feature space, and the accompanying loading vectors hold the contribution of the features in those directions. Jackknife is a method which can be used to assess the significance of variables from the stability of crossvalidated model parameters [3, 12]. Genes important for class separation may be found by jackknife of the components that best span the different classes. We use a modified version of Hotelling's T^{2} test statistic [13] that is specially tailored for crossvalidated data. The framework presented is a very general tool for significance testing, and it is applicable in principle to any dimension reduction method from which crossvalidated loadings can be obtained. Once a dimension reduction method is chosen based on the data available or the experimental setup, the same methodology for significance analysis may be used whether one, several or no responses are included in the modelling. The dependent variables may describe continuous phenotypic information, categorical information or a combination of both.
We apply our framework for significance analysis on three publicly available expression data sets, including one with known spikedin transcripts. Differences between PCA and PLSRmethods will be explored, and the results will be compared to SAM and Limma where possible. For ordinary PLSR or DPLSR, some of the weaker correlation patterns between the expression data and the responses may be overlooked in the modelling. We have therefore chosen to use a variant of PLSR called BridgePLSR [14]. This method is able to extract all information relevant to the responses in the first components, and is therefore better suited for this kind of significance analysis. The false discovery rate (FDR) is estimated by resampling to provide a threshold for significance [9, 15]. The set of parameters obtained from either of the dimension reduction methods in this work is referred to as a bilinear model. This reflects that the components representing the data are linear in both the rows and the columns. The same term is used also about the mathematical formulation of how the data are represented by scores and loadings, but the exact meaning in each case should be clear from the context. Bold uppercase letters denote matrices, and bold lowercase letters describe columnvectors. An expression data set is arranged in an array X with experiments along the rows and with a column for each feature.
A toolbox written in Matlab (Mathworks Inc. USA) with the code used in this work is provided under the GNU general public license (see additional file 1).
Results
Smoker transcriptome
We first look at a data set that investigates the effects of smoking on human airway transcripts [15]. Data are available from three groups of subjects according to their smoking history. The original article reports significant differences both between a group of smokers compared to subjects that have never been smoking, and between former smokers and the group of never smokers. A PCA of the expression data X reveals that there is a large source of variation in the data that does not correspond to the available group information. In figure 1, a PCA score plot of the two first components is given, where the objects are coloured according to the classes "Never", "Former", and "Current". Unreported experimental factors, for instance a change in laboratory procedures, may be responsible for the separation of objects into two clusters along the first component. In the corresponding loading plot in figure 2, the 725 most significant genes from the PCA analysis are given in green. These genes span mainly the unreported Xvariation and are not important for the smokinginduced differences of expression. The Venn diagram in figure 3 confirms that the genes detected by PCA are to a large degree irrelevant for classification of these data.
The PCAanalysis indicates that the group information should be allowed to guide the decomposition if any genes related to smoking are to be found. We know that a minimum of K  1 independent variables are needed for a linear separation of K classes. For instance, one gene measured for several subjects may in theory be used to differentiate between two classes, while a minimum of two genes are needed for assignments into three classes. Due to coregulation and interdependencies between genes, such single, descriptive features are usually both uninformative and hard to find. However, dimension reduction offers independent linear combinations of genes that may be used instead. If Y is a binary matrix holding the class information for each object, a good linear separation based on the expression data may be obtained with components that span all the Yrelated variance in X. In BridgePLSR, this variance may be completely described in a minimum number of components. The score plot from a twocomponent BridgePLSR model is given in figure 4. Here, three groups corresponding to the predefined classes are revealed. This model explains 54% of Y, whereas an ordinary PLSR for comparison explains 45% in two components. Jackknife is performed based on a full leaveoneout crossvalidation, and the significant features are given in the loading plot in figure 5. The green spots correspond to features called significant by both SAM and BridgePLSR, whereas the red spots indicate significant genes not found by SAM. Neighbouring genes in the loading plot have similar profiles, whereas genes in opposite quadrants are negatively correlated. Also, the arrays in a specific area of the score plot have a high expression value for the genes in the corresponding area of the loading plot. As the significant genes span mainly the first component, they discriminate well between smokers and neversmokers, according to the score plot. Along the second component, which separates former smokers from the rest, fewer significant genes are found. This indicates that this correlation is weaker and more susceptible to random variations in the data. It is seen, however, that BridgePLSR finds more genes than SAM along this direction. Any significant genes along the second component describes smokinginduced transcriptomic changes among the former smokers.
A Venn diagram which compares the significant genes found with BridgePLSR, SAM and Limma is given in figure 6. All methods agree on the significance calls for the majority of the genes, however, BridgePLSR finds an additional 185 genes which are undetected by the other methods. To verify that these genes are not primarily false positives, the genes are annotated and their biological significance with regard to smoking is evaluated. A comprehensive list of our findings is given as supplementary information (see additional file 2). Of the genes identified as differentially expressed by BridgePLSR, only 66 have unknown or poorly known biological function. An additional 14 have well understood biological function often related to the natural system, but with no readily apparent link to smoking or lung damage. Of the remaining genes, 49 are known to be involved in regulation of cell growth, cell cycle, or apoptosis, processes known to be affected by smoking [16, 17]. Seven of these genes as well as another 14 genes on the list not related to cell growth, have been shown to be involved in various forms of lung cancer. As the majority of lung cancer sufferers are smokers or previous smokers, these genes can be linked to smoking directly. Another interpretation is that these genes change early in the cancer development process. Seven genes are identified that have been reported to take part in lung development, this is not surprising as smoking can cause dramatic changes in the airways. In addition to cancer, we identify 17 genes that are related to other diseases in the lungs and bronchia, such as asthma, inflammation, fibrosis, or response to metallic toxins. Another large group of genes belongs to the protein life cycle, particularly the ribosome, or the ubiquitin cycle. This could point to an overall change in the rate of protein turnover caused by smoking. This is consistent with the results of Tomfohr et al. [18], who analysed the same data set with gene set enrichment analysis. Five genes are also found that are known to be expressed in lungs, but for which we find no link to cancer development or lung damage in the literature. Overall it seems that of the genes for which biological background information is available, the majority relate to biological processes that are influenced by smoking. The additional genes found by BridgePLSR are therefore unlikely to be false discoveries.
For prediction purposes, ordinary PLSR seems to be superior to BridgePLSR. The best predictive PLSR model is obtained for two components, with an explained Yvariance of 32%. The best predictive BridgePLSR model is a onecomponent solution with a crossvalidated explained variance in Y of 20%. The residual mean squared error of prediction even increases slightly when two components are used. Caution must therefore be used when interpreting the second component. They are both included here because the main goal of the analysis is significance testing. Hotelling's T^{2}test using only one component corresponds to an ordinary onesample ttest, and is, in this case, only able to separate between the groups "Never" and "Current". In order to find genes that might be relevant for former smokers, the second component must be investigated. The jackknife is believed to reject most of the genes that cause the residual error to increase, as those genes have a high variance across crossvalidation segments. For prediction purposes, the model with the lowest residual error might be used instead. It should be noted, however, that even if the BridgePLSR or PLSRmodel with one or two components, respectively, is optimal for prediction, neither will be able to predict former smokers.
A ridge parameter γ controls the influence of Y in the decomposition. A small value of this parameter ensures that the first BridgePLSR components span the directions of maximum covariance between X and Y. This is ideal from the point of view of significance analysis, as the major components will span all the Yrelevant variance. However, if γ = 0, the maximum number of components is limited by the rank of Y, which for these particular data is two. For explorative reasons, γ is set to a default value of 0.01 throughout this work in order to enable interpretation also of later components. For the first two components, this has no visible effect compared to a γvalue of zero. Components three and four, however, are almost identical to the first two PCA components in figure 1 (not shown). This is a further verification that the BridgePLSR is able to separate the Yrelevant structures in X from technical and other irrelevant variation. It is possible to optimise both the ridge parameter and the number of components to minimise the residual error of the bilinear model. This would require an extra layer of validation, for instance by cross model validation [19], to obtain unbiased error estimates. As the primary goal in this work is significance analysis, such an optimisation is out of scope. However, a supplementary response surface is provided for reference, along with the ordinary PLSR results (see additional file 3). It can be seen that the crossvalidated error variance is robust with regard to γ for small values of the parameter.
Analysis by ordinary PLSR is performed for the sake of comparison. This model uses two components to separate never and current smokers, and additional two or three components are needed to describe the former smokers. The optimal number of components according to the crossvalidation is two, but the majority of the significant genes found by this model span only the irrelevant Xinformation (not shown). Because of the irrelevant, dominating structure in X, BridgePLSR is a much better choice for significance analysis in this case.
Spikedin data
A further analysis is performed with Affymetrix spikein data in a constant human background [20]. The spikedin transcripts hybridise to 42 known probe sets, however, several authors report that also other probe sets show differential expression due to sequence or function similarity. Sheffler et al. [21] remove 56 such genes as outliers prior to analysis. Two independent groups [22, 23] suggest that additional 22 or 23 probe sets, respectively, should be added to the original list of known spikeins. These lists differ in a single probe set, which is known to have identical sequence to one of the spikedin genes for 5 of 11 probe pairs. We include the 22 extra genes which are suggested by both groups in our analysis, yielding a total of 64 spikein probe sets. BridgePLSR with jackknife finds all these genes among the 66 most significant ones. An estimated FDR of 0.01 results in 69 significant genes, corresponding to a true FDR of 7%. The five false positives are the probe sets "203173_s_at", "204890_s_at", "204891_s_at", "217019_at", and "200822_x_at". The maximum squared correlation coefficients between the expression values of these probe sets and the concentration profiles are, respectively, r^{2} ∈ {0.92, 0.88, 0.86, 0.33, 0.29}. For comparison, r^{2} for the list of true positives range from 0.39 to 0.74. In passing, it is noted that the three false positives with high correlation are included also in the outlierlist of Sheffer et al. [21].
Rat liver study
A data set investigating liver damage in rats after exposure to toxic doses of paracetamol is also analysed [24]. In addition to design information of dose and time of exposure, several pathological variables based on blood and tissue chemistry are available for the rats examined. There is an abundance of possible comparisons in these data, and analysis by Limma or SAM is therefore difficult. We perform an analysis by SAM in which two groups of "high" and "low" dosage are defined for each of the four timepoints of the design. The drastic treatment some of these rats receive induces profound genetic responses, leading to a long list of differentially expressed features. Testing for all 28 contrasts defined by the 8 groups, SAM assigns significance to 50% of the genes at a significance level of 0.001. This result verifies something that might be expected – that poisoning of an organism can affect several biological processes that in turn involve many genes. Investigation of more complex relationships based on these methods is not straightforward.
PLSR is well suited for handling both multiple comparisons and continuous responses. The score plot from a BridgePLSR model with all available response information in Y is given in figure 7. The spots are coloured according to the administered drugdoses. The low doses are clearly separated from the toxic doses along the first component, but the classes within those two groups are mixed together. The dashed lines envelope three groups that are related to the time from drug exposure to necropsy. The lower group consists of the 6hour design points, the upper group corresponds to the 48hour points, while in the middle group the 18 and 24hour points are mingled together. The second component is thus related to time since exposure. A leaveoneout crossvalidation is performed also for these data. The optimal model for prediction is obtained after four components, with a residual variance of 47%. However, we are interested mainly in the genes which are differentially expressed across dose and time. We therefore use a twocomponent model in the significance analysis, with a prediction error of 58%. For these data with much structure in Y, the differences between the results from BridgePLSR and ordinary PLSR are small. All predictive components describe the same phenomena, and the residual errors are similar, both in a calibrated and validated sense. For reference, the full response surface including several values γ of is given as supplementary information (see additional file 3).
A helpful plot is the Yloading plot in figure 8. Here, each spot corresponds to a responsevariable, and the predictabilities are given by the associated colourbar. The colour of a spot reflects the crossvalidated, explained variance for that variable, given in percent of the true response's variation. The variables with red spots are very well predicted while the blue spots represent variables that cannot reliably be modelled based on the expression data. Two variables which have high predictive abilities and are correlated with the dose are ALT (alanine aminotransferase) and AST (aspartate aminotransferase). Increased levels of these enzymes in the blood stream are diagnostic markers for liver damage. The large degree of explained variance for these variables indicates that the first component actually describes liver injury directly, rather than just the concentration of paracetamol in the rats. Because the individual responses to the drug vary, these markers are better suited to model the genetic response to overdose than the design is. This information should be included also in the interpretation of the scores in figure 7. It is seen that the response to overdose peaks between 18 and 24 hours after exposure, as these groups span the first component better than the 6 and 48hour groups. After 48 hours, metabolism may have disposed of some of the paracetamol, the livercells may have started to repair, and the levels of ALT and AST in the blood have started to decrease. Other correlations seen in figure 8 may also be of interest. Liver damage seems to be associated with increased concentrations of total bile acids and blood urea nitrogen, and there is also an indication that the level of cholesterol decreases. The reduced glutathione level in the liver tissue increases with time after poisoning, and the level of blood albumin seems to decrease accordingly.
An estimated FDR significance cutoff of 0.1% yields 25% significant features for the BridgePLSR model. The Xloading plot is given in figure 9, with the significant genes marked by green spots. The majority of these features span the component describing liver injury, but some are also found that vary with time. The significant genes found by SAM are scattered more uniformly over the same loadings, making any systematic trends hard to detect (not shown).
Discussion
Dimension reduction methods are explorative in nature, and the decomposition attempts to find components that describe as much structure as possible using a small number of components. This is a very effective way to detect the sources of variation in large data tables, and visual inspection of score and loading plots enables discovery of unknown phenomena. The information carried by the components often corresponds to underlying biological processes, greatly simplifying interpretation of the results. Comparison of score and loading plots forms a central interpretative tool for mapping measured genetic responses to different subjects or levels of an experimental design. The dimension reduction methods may be vulnerable to outliers and systematic noise, and it is important that the results are properly validated and understood. The inspection of scores and loadings can be used as an aid in the discovery of bad arrays or bad spots, respectively.
Most genes are known to interact with other genes in potentially complex cellular networks. Investigation of one gene at a time is therefore not optimal for a good understanding of a biological system. The dimension reduction methods attempt to find linear combinations of genes that describe the data well, and are therefore better equipped to model coregulated genes. The representation of expression data by a small number of components also enables use of statistical methods that assume independence. The significance testing itself is also multivariate, as features are tested for significance in multiple dimensions simultaneously. Genes that are not significant along any of the individual components may still be important for the full model.
The dimension reduction methods are applicable to a wide range of problems. PCA is widely used for clustering and for exploration of the main structure of a data set. In this work, PCA revealed a large systematic variance in the smoker data which was irrelevant for classification. A visual inspection of the scores clearly showed that a supervised decomposition was needed. PLSR has been much used in the classification of expression data, and other uses have been shown where the Yinformation is noncategorical. It is hard to control all levels of variability in living organisms, and the responses to stimuli may differ between test subjects. Similarly, biopsies used in cancer classification are heterogeneous material and may vary even among patients with the same diagnosis. The direct use of clinical and pathological parameters in the analysis may enable a more accurate modelling than if just the classes or levels of a design are used.
While significance analysis by jackknife is much used in other fields [3], at least two new concepts are introduced in this work. One is to use BridgePLSR to describe more of the response variation in fewer components. This may be needed when there is a large number of noisy or irrelevant variables in X, which is typical for expression data analysis. For the ratliver data, no discernible difference was found between BridgePLSR and standard PLSR. For the smoker data, PLSR was more predictive for future objects, however, BridgePLSR was able to describe the classes better and is preferred for significance analysis. It is acknowledged also in [25] that the best genes for prediction are not always the most interesting ones. The second novel concept is the modified Hotelling's T^{2}test for use with jackknifed models. This test is believed to be superior to the standard alternative of performing a ttest for each component individually.
There are several possible extensions to the presented framework that may be included to handle more complex problems. For instance, dimension reduction algorithms exist that inherently handle missing data [4]. Often background information about the features are available, for instance in the form of gene ontology or metabolic networks [26, 27]. There are various ways to include a priori information in the analysis, either in the form of a preprocessing step or in a specialised algorithm [28]. The dimension reduction methods are not even restricted to twoway data. It is possible that a threeway organisation of the ratliver data into an array of timebydose for each gene would result in a more stable decomposition. Several possible dimension reduction methods exist also for arrays of higher order [29, 30]. Finally, the presented methodology is not restricted to expression data. Any bioinformatics application producing many measured variables per available object can be expected to benefit from these methods.
Conclusion
The dimension reduction methods are multivariate and well suited to deal with the intricate network of dependencies between genes. They offer a versatility that allows any continuous or discrete background information available to be incorporated in the analysis. The results are intuitively understood based on simple scatter plots, which also enable the discovery of unknown phenomena in the data. The identification of differentially expressed genes is an important part of expression data analysis. We have presented a general scheme for significance testing that can be applied to a wide range of dimension reduction methods. The validity of the procedure has been established by comparison with other, well known algorithms for significance analysis, by biological interpretation, and by analysis of spikein data.
Methods
Data sets
The smoker data [15] contain expression levels for M = 7939 probe sets measured for N = 75 subjects on Affymetrix HGU133A GeneChips (Affymetrix, Inc., USA). The subjects are divided into 34 current smokers, 18 former smokers and 23 individuals that have never been smoking. The expression data are arranged in a matrix X of size N rows by M columns, and the class information is contained in a binary array Y. Each class is represented by a column in Y, where a value of one assigns an object to that class. X is logarithmically transformed prior to the analysis. No other preprocessing or scaling is performed, that is, we use the data directly as provided by Spira et al. [15].
The spikein data [20] are obtained from a designed experiment where 14 groups of three transcripts each were spiked in at 14 different concentration levels in three replicates. There is thus a total of N = 42 objects in the experiment, corresponding to specific concentration mixtures. Expression levels were measured on a modified Affymetrix HGU133A array, and concentrations ranging from zero to 512 pM were arranged in a Latin square setup. The 42 spikein genes were measured in a constant human background, 22300 probe sets in total. The data are preprocessed using the gcrmapackage [31], which is part of the Bioconductor project [32]. As groups of three transcripts are spiked in at the same amounts, the response matrix Y contains 14 concentration profiles, one for each group. Several authors have remarked that some of the background genes show consistent differential expression across the experiments [21–23]. Some of those genes contain sequences which match the spikeins exactly. We therefore adopt the extended spikein genelist given by McGee and Chen [22], yielding a total of 64 entries. The list is identical to the one suggested by Lo and Gottardo [23], except the probe set "205397_x_at" is removed. This probe set has identical sequence to the originally included "205398_at" for 5 of 11 probes. Due to memory limitations, there is a need to filter away some genes before the analysis. We therefore include only those normalised probe sets with a standard deviation above 0.4, which results in a total of M = 14.490 variables in X. Potential alternatives to filtering would have been to apply a leavetwoout crossvalidation, or to use fewer Yvariables or components in the analysis.
In the liver toxicity study [24], male rats of the inbred strain Fisher 344 were administered nontoxic (50 or 150 mg/kg), moderately toxic (1500 mg/kg) or severely toxic (2000 mg/kg) doses of paracetamol in a controlled experiment. Necropsies were performed at 6, 18, 24 and 48 hours after exposure. RNA analysis was performed on an Agilent platform (Agilent Technologies, USA), and four biological replicates were obtained for each of the 16 designpoints. All spots flagged as bad or not found by the image analysis program are removed. In addition, spots are removed if they have total intensity less than 200, foreground to background intensity less than 1.5, spot diameter less then 25 pixels, or more than 50% of pixels saturated. As the background estimates available in the Agilent image analysis report are local backgrounds measured in pixels near the spots, it is doubtful if this is a useful estimate of the background in the spot [33]. Therefore, no background correction is done. The filtered data are transformed to log_{2}ratios and normalised using global loess to improve colour balance and remove global intensity effects [34]. Ratios for spots that have been removed in the filtering are imputed using knearest neighbours with k = 10 [35]. Two histological and ten clinical variables containing markers for liver injury are available for each object. The expression data are arranged in a matrix X of N = 64 objects and M = 7445 expression levels. For analysis by SAM, the two nontoxic levels are pooled in a group of low dose, and the two toxic levels are pooled in a group of high dose. The data are further divided into groups for each of the four timepoints, giving a total of 8 groups to be tested for differential expression. In the bilinear analysis, a matrix Y of 14 columns is constructed instead that contains the design information of dose and time as well as the phenotypic responses available. A description of all the variables in Y are given in figure 8. The Yvariables SDH, BILE, AST and ALT are logtransformed in order to get responses which are more closely bellshaped.
BridgePLS regression
PLSR finds linear combinations (scores) of the original features spanning directions of high covariance between X and Y. In the most common PLSR implementation, the data are deflated each time a component is found, by subtraction of that component from the remaining Xdata. This ensures that each score vector is orthogonal to all previous components. It also enables better model fit, as several components are allowed in the modelling of each individual response. Crossvalidation is usually applied to find the number of components which yields the best predictions [3]. However, the best prediction model does not necessarily contain the most informative components. If a single response is modelled (i.e. if Y has only one column), qualitatively useful information is expected from the vector of covariances between the variables in X and the response [36, 37]. The covariances are given by the socalled loading weights corresponding to the first component, and this vector is therefore well suited for significance analysis. Subsequent loading weights are less interpretable, as they, due to the deflation step, may be influenced by information in X unrelated to Y. In order to obtain informative weights for multiple responses, it is possible to make one PLSRmodel for each individual response. Such an approach was taken e.g. in [9], where cell cycle regulated genes were found using the first loading weight from two singleresponse PLSRmodels.
BridgePLSR has no deflation step, and the loading weights are used directly in the model rather than calculating new loadings [14]. If the ridgeparameter γ is set to zero, BridgePLSR is simply a singular value decomposition of the covariance matrix X ' Y, with a subsequent prediction relation between X and Y. At the other extreme, when γ = 1, BridgePLSR becomes an ordinary principal component regression [3]. Any intermediate value of γ corresponds to some weighted average between those two methods. For significance analysis, and for interpretation of components in relation to Y, the optimal value of γ is zero. The loadings will then collectively explain the same Yinformation as the set of loading weights from all onecomponent, individual response PLSRmodels. For multiple correlated responses, Kvalheim [38] recommends orthogonalisation of Y before calculating such singleresponse models. This ensures that the models will be close to orthogonal to each other as long as the predictive relation between X and Y is good. These steps are all circumvented with BridgePLSR, as the BridgePLSR loadings are orthogonal by default. In addition, a single, predictive BridgePLSR model is more useful and appealing to work with than multiple singleresponse PLSRmodels.
When γ = 0, the maximum number of components equals the sum of independent responses in Y. If all these components are used, the residuals will be orthogonal to Y, and the loadings (= loading weights) will together account for all the Yrelated information in X. Because only one component is available for each orthogonal response when γ = 0, the predictive ability of BridgePLSR is often less than that of an ordinary PLSR. However, a tuning of the γparameter may increase the predictive ability of the model. This is expected for instance in cases where Y is large or noisy compared to X. It is also seen from the response surfaces provided in supplementary information that a larger γvalue may stabilise the decomposition such that early local minima are avoided. However, the response surfaces provided are not validated by an external validation and should not be used for optimisation. Any such tuning of parameters should be accompanied for instance by cross model validation to avoid optimistically biased results [19, 39].
For the analysis in this work, γ is set to the small default value of 0.01. It is seen from the response surfaces in supplementary information that this yields the same predictive ability (down to two decimals) as if γ were zero. Also, the components obtained using the two values of γ are qualitatively equal, such that the influence on the significance analysis is negligible. The γparameter is given this value only to enable interpretation of more than two components for the smoker data, although these components are not used in the jackknife. BridgePLSR is algorithmically simple, and the models are validated by crossvalidation or test sets in a straight forward manner.
Pseudocode BridgePLSR

1.
Start with columncentred X and Y

2.
Find by SVD Y = U_{ y }S_{ y }${V}_{\text{y}}^{\text{T}}$

3.
Construct a representation of Y that has the rank of X: G = (1  γ) U_{ y }S_{ y }${U}_{y}^{T}$ + γ I

4.
Find by SVD X^{T}G = USV^{T}

5.
Loadings P_{ A }given by the A largest components in U

6.
Scores T_{ A }= XP_{ A }

7.
Yloadings Q_{ A }= Y^{T}T_{ A }(${T}_{A}^{T}$T_{ A })^{1}

8.
Regression coefficients B = P_{ A }${Q}_{A}^{T}$

9.
Predictive relation $\widehat{Y}=1{b}_{0}^{T}+XB$, where the offset ${b}_{0}=\overline{y}{B}^{T}\overline{x}$
Optional extensions
Even though all the Yrelated information is accounted for in BridgePLSR, the model may also include variance which is orthogonal to Y. Such orthogonal variance is the reason the smokerclasses are not better separated in figure 4. In a perfectly fitted model, all the samples of either class would appear as a single spot in the score plot, such that the three classes formed a triangle. In general, perfect fit is not a goal, as such a model would describe also measurement noise and other artifacts. However, in cases with much irrelevant Xvariation, removal of some unwanted structures might improve the analysis. At least three approaches can be taken to this end. Variable selection attempts to remove the least relevant variables from X, improving the subsequent dataanalysis [19]. Target rotation of PLSR loading weights toward orthogonalised responses will improve the fit provided the number of initial components is sufficiently high [37, 38]. Target rotation of two BridgePLSR components toward two independent responses will give no improvement, as the BridgePLSR loadings already span the directions of interest (when γ = 0). A third approach is to use some sort of orthogonal signal correction, a family of methods which separate components from X which are orthogonal to Y [40, 41]. The method of 02PLS [40] is especially interesting, as this model formulation is very similar to BridgePLSR. In a test using the smoker data, target rotation based on five PLSR components gave similar results as an O2PLS model where three orthogonal components had been removed (not shown).
Removal of some unwanted Xinformation may increase the predictive ability and find even more relevant genes. However, all the above procedures for improving the modelling are based on the data themselves, which means that degrees of freedom are in effect removed from X. It follows that the risk of overfit increases the more the data are manipulated. As an example, a perfect fit will always result when a target rotation is based on all data, or when all orthogonal components are removed from the model. Much caution and good knowledge of the data are therefore advised when any of the above procedures are implemented. As these methods are useful only for data with much irrelevant Xvariation, we have chosen not to include them further in this work. As was demonstrated with the smokerdata, BridgePLSR will find the important directions in the data no matter how much irrelevant Xstructure is present.
Significance analysis
A crossvalidation using an appropriate dimension reduction method is performed, and loadings from each submodel with leftout objects is obtained. The perturbed loadings are used to estimate the stability of each feature. The optimal number of components A for prediction, and the overall predictability of the model may be estimated as described in [3]. Only components that lead to a significant decrease in the crossvalidated prediction error should then be included. In this work, we concentrate more on significance analysis than on optimisation of prediction models, and interpretation of score plots is therefore at least as important when A is found.
Many of the dimension reduction methods yield orthonormal loadings, in which cases the relative importance of the components are reflected by the length of the scores. For the significance analysis, it is important that the loadings themselves are weighted by the variance of their components. All perturbed loadings from the crossvalidation also have to be aligned properly. A bilinear model can be written on the form $\widehat{X}={T}_{A}{P}_{A}^{T}={T}_{A}C{C}^{1}{P}_{A}^{T}$, where C is any invertible matrix of rank A. This means that the components have full rotational freedom, and it is often observed that a component is flipped 180° in one submodel relative to another. Components of similar eigenvalues may also be sorted differently in different submodels. A Procrustes rotation of the perturbed models toward the calibration model will correct for these inconsistencies [3]. We perform uncentred and unscaled rotations of all loadings prior to significance testing.
Each variable (gene) is tested under the null hypothesis that its corresponding loadingparameter is zero in the Adimensional space spanned by the components. It follows that the choice of dimension reduction is very important for the outcome of the test. Figures 1 and 2 show that components spanning only irrelevant Xvariation will yield irrelevant genes. On the other hand, when the components describe some Yinformation of interest, we directly find the features important for the different responses (e.g. figures 4 and 5). Hotelling's T^{2}test is able to assess the significance of parameters in several dimensions simultaneously. We have modified the test to make it more suited for jackknife of bilinear model parameters. Some of these modifications are described in [3, 12] for the regular onesample ttest. The variation of the perturbed submodels around an estimated true model describes the stability for each gene. Traditionally, the mean of the submodels represents the centre points, but the median or even the full calibration model may also be used. Because the perturbed models are not independent of the full model, the deviations are summed instead of averaged. The latter point makes a difference if the actual value of the T^{2}score is interpreted, not if the T^{2}score is used simply as a ranking of the genes. Finally, shrinkage of the genespecific variances toward a total variance across genes is included [25]. This is to avoid poorly estimated variances due to the small number of objects in microarray experiments. A shrinkage factor of zero means that every gene is evaluated individually, while a factor of one treats all genes as having equal variances. The actual value of the parameter is a choice that must be made on the basis of the number of replicates and the overall data quality. Poor quality data and few replicates may justify increased shrinkage.
The T^{2}score for gene i ∈ {1,..., M} is given as
The vector p_{ i }is of length A and represents the true loadings, β ∈ [0, 1] is the shrinkage factor, ${P}_{i}^{CV}$ is a matrix of size KbyA holding the loadings from the K submodels and 1 is a vector of ones. S^{tot}is the mean or median of the scaled, genespecific variancecovariance matrices ${S}_{i}^{gene}$ across all genes. The scaling $g=\frac{K1}{K}$ is taken from [3] and is close to one in most cases. In this work, p_{ i }is the mean value over submodels, the coefficient β = 0.1 and the median of covariances is used to find S^{tot}.
FDR estimation
We adapt the definition of the positive false discovery rate [42], which states that
FP and TP are the numbers of false and true positives, respectively. A qvalue ∈ [0, 1] assigned to each gene gives the minimum FDR that results in a rejected hypothesis for that feature. An FDR significance level cutoff is applied to the qvalues to decide which features to call significant. The conditional statement that the sum of positive outcomes must be larger than zero can often be disregarded in practice. It is then possible to estimate the qvalues by resampling, using the relation
Every feature i is associated with a score ${T}_{i}^{2}$. The denominator holds the total number of genes at least as significant as this based on the calibration model. The number of false positives at the same significance level is not known, and must be estimated. The estimated count for gene i is found by reshuffling and denoted $\widehat{FP}$_{ i }. The rows of Y are shuffled R times to give several matrices ${Y}_{r}^{\ast}$, r ∈ {1,..., R}, with random objectinformation only. A jackknife based on one of the randomised matrices is expected to give no true positives, especially if the number of significant features is small compared to the total number of features tested. Here, $\widehat{FP}$ is found as the median number of significant outcomes over R = 1000 shuffled responses, and all features with ${\widehat{q}}_{i}$ smaller than a preset significance level α are called significant. Estimation of $\widehat{FP}$ when no Y is available is difficult, because it is not possible to shuffle X for each feature separately without simultaneously disrupting all other systematic structure in the data. For the smokerdata, we have therefore used the same number of significant features for PCA as we found for BridgePLSR. The FDR significance level is set to α = 0.05 for the smokerdata and to α = 0.01 for the spikein data. Because the relevant signal in the paracetamol data set is very strong, a significance level of α = 0.001 is enough to get a large number of significant features for these data.
SAM and Limma analysis
SAM and Limma are used to identify differentially expressed genes in the smoker data set. Both methods are part of the Bioconductor software (Version 1.9) [32], and we use the packages limma and siggenes for Limma and SAM, respectively. Limma is performed with the functions lmFit and eBayes with default settings. We control Benjamini and Hochberg's FDR at a level of α using the function decideTests. SAM is run with the function sam.dstat with 1000 permutations, the parameter delta = 0.8 and the remaining parameters at default setting.
References
 1.
Tusher VG, Tibshirani R, Chu G: Significance analysis of microarrays applied to the ionizing radiation response. P Natl Acad Sci USA 2001, 98: 5116–5121.
 2.
Smyth GK: Limma: linear models for microarray data. In Bioinformatics and computational biology solutions using R and Bioconductor. Edited by: Gentleman R, Carey V, Dudoit S, Irizarry R, Huber W. New York, USA: Springer; 2005:397–420.
 3.
Martens H, Martens M: Multivariate analysis of quality: An introduction. Chichester, UK: Wiley; 2001.
 4.
Wold S, Esbensen K, Geladi P: Principal component analysis. Chemometr Intell Lab 1987, 2: 37–52.
 5.
Boulesteix AL, Strimmer K: Partial least squares: a versatile tool for the analysis of highdimensional genomic data. Brief Bioinformatics 2007, 8: 32–44.
 6.
Nguyen DV, Rocke DM: Tumor classification by partial least squares using microarray gene expression data. Bioinformatics 2002, 18: 39–50.
 7.
Nguyen DV, Rocke DM: Multiclass cancer classification via partial least squares with gene expression profiles. Bioinformatics 2002, 18: 1216–1226.
 8.
Park PJ, Tian L, Kohane IS: Linking gene expression data with patient survival times using partial least squares. Bioinformatics 2002, 18: S120S127.
 9.
Johansson D, Lindgren P, Berglund A: A multivariate approach applied to microarray data for identification of genes with cell cyclecoupled transcription. Bioinformatics 2003, 19: 467–473.
 10.
Dai JJ, Lieu L, Rocke D: Dimension reduction for classification with gene expression microarray data. Stat Appl Genet Mo B 2006., 5: Article 6. Article 6.
 11.
Boulesteix AL: Reader's Reaction to "Dimension Reduction for Classification with Gene Expression Microarray Data" by Dai et al (2006). Stat Appl Genet Mo B 2006., 5: Article 16. Article 16.
 12.
Martens H, Martens M: Modified jackknife estimation of parameter uncertainty in bilinear modelling by partial least squares regression (PLSR). Food Qual Prefer 2000, 11: 5–16.
 13.
Xiong M, Zhao J, Boerwinkle E: Generalised T^{2}test for genome association studies. Am J Hum Genet 2002, 70: 1257–1268.
 14.
Gidskehaug L, StødkildeJørgensen H, Martens M, Martens H: BridgePLS regression: twoblock bilinear regression without deflation. J Chemometr 2004, 18: 208–215.
 15.
Spira A, Beane J, Shah V, Liu G, Schembri F, Yang X, Palma J, Brody JS: Effects of cigarette smoke on the human airway epithelial cell transcriptome. P Natl Acad Sci USA 2004, 101: 10143–10148.
 16.
Kopiński P, Przybylsk G, BalickaSlusarczyk B, Jarzemska A, Dyczek A, Pinis G, Bartczak K, Plato M, Jankowski M, Szczeklik J: [Apoptosis of alveolar lymphocytes in sarcoidosis and in control group is more frequent in smokers than in nonsmoking persons]. Przegl Lek 2006, 63: 841–847.
 17.
Dasgupta P, Chellappan S: Nicotinemediated cell proliferation and angiogenesis: new twists to an old story. Cell Cycle 2006, 5: 2324–2328.
 18.
Tomfohr J, Lu J, Kepler TB: Pathway level analysis of gene expression using singular value decomposition. BMC Bioinformatics 2005, 6: 225.
 19.
Anderssen E, Dyrstad K, Westad F, Martens H: Reducing overoptimism in variable selection by crossmodel validation. Chemometr Intell Lab 2006, 84: 69–74.
 20.
Affymetrix – Latin Square Data[http://www.affymetrix.com/support/technical/sample_data/datasets.affx]
 21.
Sheffler W, Upfal E, Sedivy J, Noble WS: A learned comparative expression measure for Affymetrix GeneChip DNA microarrays. Proc IEEE Comput Syst Bioinform Conf 2005, 00: 144–154.
 22.
McGee M, Chen Z: New SpikedIn Probe Sets for the Affymetrix HGU133A Latin Square Experiment. COBRA Preprint Series 2006. Article 5. [http://biostats.bepress.com/cobra/ps/art5] Article 5.
 23.
Lo K, Gottardo R: Flexible empirical Bayes models for differential gene expression. Bioinformatics 2007, 23: 328–335.
 24.
Irwin RD, Parker JS, Lobenhofer EK, Burka LT, Blackshear PE, Vallant MK, Lebetkin EH, Gerken DF, Boorman GA: Transcriptional profiling of the left and median liver lobes of male F344/N rats following exposure to acetaminophen. Toxicol Pathol 2005, 33: 111–117.
 25.
Allison DB, Cui X, Page GP, Sabripour M: Microarray data analysis: from disarray to consolidation and consensus. Nat Rev Genet 2006, 7: 55–65.
 26.
Subramanian A, Tamayo P, Mootha VK, Mukherjee S, Ebert BL, Gillette MA, Paulovich A, Pomeroy SL, Golub TR, Lander ES, Mesirov JP: Gene set enrichment analysis: A knowledgebased approach for interpreting genomewide expression profiles. P Natl Acad Sci USA 2005, 102: 15545–15550.
 27.
National Center for Biotechnology Information[http://www.ncbi.nlm.nih.gov/]
 28.
Martens H, Anderssen E, Flatberg A, Gidskehaug LH, Høy M, Westad F, Thybo A, Martens M: Regression of a data matrix on descriptors of both its rows and of its columns via latent variables: LPLSR. Comput Stat Data An 2005, 48: 103–123.
 29.
Bro R, Geladi P, Smilde AK: Multiway analysis with applications in the chemical sciences. Chichester, UK: Wiley; 2004.
 30.
Bro R: Multiway calibration. Multilinear PLS. J Chemometr 1996, 10: 47–61.
 31.
Wu Z, Irizarry RA, Gentleman R, MartinezMurillo F, Spencer F: A modelbased background adjustment for oligonucleotide expression arrays. J Am Stat Assoc 2004, 99: 909–917.
 32.
Gentleman RC, Carey VJ, Bates DM, Bolstad B, Dettling M, Dudoit S, Ellis B, Gautier L, Ge Y, Gentry J, Hornik K, Hothorn T, Huber W, Iacus S, Irizarry R, Li FLC, Maechler M, Rossini AJ, Sawitzki G, Smith C, Smyth G, Tierney L, Yang JYH, Zhang J: Bioconductor: Open software development for computational biology and bioinformatics. Genome Biol 2004, 5: R80.
 33.
Yang YH, Buckley MJ, Speed TP: Analysis of cDNA microarray images. Brief Bioinform 2001, 2: 341–349.
 34.
Smyth GK, Speed T: Normalization of cDNA microarray data. Methods 2003, 31: 265–273.
 35.
Troyanskaya O, Cantor M, Sherlock G, Brown P, Hastie T, Tibshirani R, Botstein D, Altman RB: Missing value estimation methods for DNA microarrays. Bioinformatics 2001, 17: 520–525.
 36.
Haaland DM, Thomas EV: Partial LeastSquares Methods for Spectral Analyses. 1. Relation to Other Quantitative Calibration Methods and the Extraction of Qualitative Information. Anal Chem 1988, 60: 1193–1202.
 37.
Kvalheim OM, Karstang TV: Interpretation of LatentVariable Regression Models. Chemometr Intell Lab 1989, 7: 39–51.
 38.
Kvalheim OM: LatentVariable Regression Models with HigherOrder Terms: An Extension of Response Modelling by Orthogonal Design and Multiple Linear Regression. Chemometr Intell Lab 1990, 8: 59–67.
 39.
Ambroise C, McLachlan GJ: Selection bias in gene extraction on the basis of geneexpression data. P Natl Acad Sci USA 2002, 99: 6562–6566.
 40.
Trygg J, Wold S: O2PLS, a twoblock (XY) latent variable regression (LVR) method with an integral OSC filter. J Chemometr 2003, 17: 53–64.
 41.
Westerhuis JA, de Jong S, Smilde AK: Direct orthogonal signal correction. Chemometr Intell Lab 2001, 56: 13–25.
 42.
Storey JD, Tibshirani R: Statistical significance for genomewide studies. P Natl Acad Sci USA 2003, 100: 9440–9445.
Acknowledgements
We thank the reviewers for constructive and thorough feedback. The Norwegian Microarray Consortium and the Functional Genomics (FUGE) initiative of the Norwegian Research Council (NFR) are gratefully acknowledged for financial support to our group.
Author information
Additional information
Competing interests
The author(s) declares that there are no competing interests.
Authors' contributions
LG wrote the manuscript. LG, EA and AF contributed equally to method development, programming and data analysis. BKA (group leader) helped to draft the manuscript. All authors read and approved the final version.
Electronic supplementary material
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Gidskehaug, L., Anderssen, E., Flatberg, A. et al. A framework for significance analysis of gene expression data using dimension reduction methods. BMC Bioinformatics 8, 346 (2007). https://doi.org/10.1186/147121058346
Received:
Accepted:
Published:
Keywords
 Significance Analysis
 Singular Value Decomposition
 Partial Little Square Regression
 Score Plot
 Partial Little Square Regression Model