 Research
 Open access
 Published:
So you think you can PLSDA?
BMC Bioinformatics volume 21, Article number: 2 (2020)
Abstract
Background
Partial LeastSquares Discriminant Analysis (PLSDA) is a popular machine learning tool that is gaining increasing attention as a useful feature selector and classifier. In an effort to understand its strengths and weaknesses, we performed a series of experiments with synthetic data and compared its performance to its close relative from which it was initially invented, namely Principal Component Analysis (PCA).
Results
We demonstrate that even though PCA ignores the information regarding the class labels of the samples, this unsupervised tool can be remarkably effective as a feature selector. In some cases, it outperforms PLSDA, which is made aware of the class labels in its input. Our experiments range from looking at the signaltonoise ratio in the feature selection task, to considering many practical distributions and models encountered when analyzing bioinformatics and clinical data. Other methods were also evaluated. Finally, we analyzed an interesting data set from 396 vaginal microbiome samples where the ground truth for the feature selection was available. All the 3D figures shown in this paper as well as the supplementary ones can be viewed interactively at http://biorg.cs.fiu.edu/plsda
Conclusions
Our results highlighted the strengths and weaknesses of PLSDA in comparison with PCA for different underlying data models.
Background
Partial LeastSquares Discriminant Analysis (PLSDA) is a multivariate dimensionalityreduction tool [1, 2] that has been popular in the field of chemometrics for well over two decades [3], and has been recommended for use in omics data analyses. PLSDA is gaining popularity in metabolomics and in other integrative omics analyses [4–6]. Both chemometrics and omics data sets are characterized by large volume, large number of features, noise and missing data [2, 7]. These data sets also often have lot fewer samples than features.
PLSDA can be thought of as a “supervised” version of Principal Component Analysis (PCA) in the sense that it achieves dimensionality reduction but with full awareness of the class labels. Besides its use for dimensionalityreduction, it can be adapted to be used for feature selection [8] as well as for classification [9–11].
As its popularity grows, it is important to note that its role in discriminant analysis can be easily misused and misinterpreted [2, 12]. Since it is prone to overfitting, crossvalidation (CV) is an important step in using PLSDA as a feature selector, classifier or even just for visualization [13, 14].
Furthermore, precious little is known about the performance of PLSDA for different kinds of data. We use a series of experiments to shed light on the strengths and weaknesses of PLSDA visàvis PCA, as well as the kinds of distributions where PLSDA could be useful and where it fares poorly.
The objective of dimensionalityreduction methods such as PCA and PLSDA is to arrive at a linear transformation that converts the data to a lower dimensional space with as small an error as possible. If we think of the original data matrix to be a collection of nmdimensional vectors (i.e., X is a n×m matrix), then the above objective can be thought of as that of finding a m×d transformation matrix A that optimally transforms the data matrix X into a collection of nddimensional vectors S. Thus, S=XA+E, where E is the error matrix. The matrix S, whose rows correspond to the transformed vectors, gives ddimensional scores for each of the n vectors in X.
The new features representing the reduced dimensions are referred to as principal components (PC). In PCA, the transformation preserves in its first PC as much variance in the original data as possible. On the other hand PLSDA preserves in its first PC as much covariance as possible between the original data and its labeling. Both can be described as iterative processes where the error term is used to define the next PC. Figure 1 highlights the differences showing an example of a synthetic data set for which the PC chosen by PCA points to the bottom right, while the one chosen by PLSDA is roughly orthogonal to it pointing to the bottom left.
It is also important to note that a higher explained variance or higher correlation for both PCA and PLSDA doesn’t always mean a better model, even though they are many times linked [14]. The following paragraphs give a more thorough description of the methods and their differences:
PCA Informally, the PCA algorithm calculates the first PC along the first eigenvector by minimizing the projection error and then iteratively projects all the points to a subspace orthogonal to the last PC and repeats the process on the projected points. An alternative formulation is that the principal component vectors are given by the eigenvectors of the nonsingular portion of the covariance matrix C given by:
where C_{n} is the n×n centering matrix. The loading vectors, denoted by L_{1},…,L_{n}, are given in terms of the eigenvectors, e_{1},…,e_{n} and the eigenvalues, λ_{1},…,λ_{n}, of C as follows:
PLSDA In its standard variant the components are required to be orthogonal to each other. In a manner similar to Eq. (1), the first PC of PLSDA can be formulated as the eigenvectors of the nonsingular portion of C, given by:
where y is the class label vector.
The iterative process computes the loading vectors a_{1},…,a_{d}, which give the importance of each feature in that component. In iteration h, it has the following objective:
where b_{h} is the loading for the label vector y_{h},X_{1}=X, and X_{h} and y_{h} are the residual (error) matrices after transforming with the previous h−1 components.
sPLSDA Variant of PLSDA that makes a sparsity assumption, i.e., that only a small number of features are responsible for driving a biological event or effect under study has been devised [15, 16] and shown to be successful with applications where the number of features far outnumber the number of samples [17]. Using lasso penalization, these methods add penalties (L_{1} or L_{0} norm) to better guide the feature selection and model fitting process and achieve improved classification by allowing to select a subset of the covariates instead of using all of them.
Methods
In this section, we discuss the aim, design and settings of the experiments.
Synthetic data for the experiments
The following describes a standard experimental setup. Clarifications are provided wherever the experiments differed from this norm. For each of the experiments, labeled synthetic data were generated as follows. The basic input parameters for each experiment are the number of samples n and the number of features of each sample m. Every data set assumed that there was a rule (e.g., a linear inequality), which was a function of some subset of the m features (i.e., signal features), while the rest were considered as noise features. The input parameter also included the rule and consequently the set of signal features. This rule will be considered as the ground truth. PLSDA was then applied to the data set to see how well it performed feature selection or how well it classified. All experiments were executed using PCA and sPLSDA, where the loading vector is only nonzero for the selected features. Both are available in the mixOmics R package [18], which was chosen because it is the implementation most used by biologists and chemists. The noise features of all points are generated from a random distribution that is specified as input to the data generator. The default is assumed to be the uniform distribution. The satisfied rule dictates the generation of the signal features.
Performance metrics for the experiments
As is standard with experiments in machine learning, we evaluated the experiments by computing the following measures: true positives (tp), true negatives (tn), false positives (fp), false negatives (fn), precision (tp÷(tp+fp)), and recall (tp÷(tp+fn)). Note that in our case precision and recall are identical. This is because of their formula is the same if fp=fn. The data is created with s signal features and s features are selected. Because s is the number of signal features, regardless of whether they were selected or not, s=tp+fn. Also, because only s features are selected, s=tp+fp. Making both equations equal, we get that fp=fn.
Since tn is large in all our feature extraction experiments, some of the more sophisticated measures are skewed and therefore not useful. For example, the F1 score will be necessarily low, while accuracy and specificity will be extremely high. When the number of noise features is low, precision could be artificially inflated. However, this is not likely in real experiments.
Graphs are shown as 3D plots where the z axis represents the performance measure (percentage of signal features in the features marked as important by the tools), and the x and y axes show relevant parameters of the experiment.
Experiments varying n/m
We first show how the ratio of the number of samples, n, to the number of features, m affects the apparent performance of PLSDA and the number of spurious relationships found.
As described earlier, we generated n random data points in mdimensional space (from a uniform distribution) and labeled them randomly. The ratio n/m was reduced from 2:1 to 1:2 to 1:20 to 1:200. Given the data set, it is clear that any separation of the data found by any method is merely occurring fortuitously. When we have at least twice as many features as samples, PLSDA readily finds a hyperplane that perfectly separates both merely by chance. As shown in Fig. 2, the two randomly labeled groups of points become increasingly separable. This is explained by the curse of dimensionality, that predicts the sparsity of the data to grow increasingly faster with the number of dimensions.These executions only range in ratios from 2:1 to 1:200. In many current omics data sets, ratios can even exceed 1:1000 (i.e., data sets with 50 samples and 50,000 genes are common). This is one of the reasons of the need of sample size determination when designing an experiment [19].
If any separating hyperplane is used as a rule to discriminate blue points from orange points, then even though the apparent error rate (AE) decreases for this set, its ability to discriminate any new random points will remain dismal [20]. In fact, the CV error rates using 1000 repetitions for the first PC in the four experiments shown in Fig. 2 were 0.53, 0.53, 0.5 and 0.48 respectively, showing that even though separability increased, the errors remain unreasonably large. CV errors vary with the seed used to initialize the matrix but the trend remains.
Results
In this section, we discuss a variety of experiments with synthetic and real data that will help us explain the strenghts and weaknesses of PLSDA visávis PCA and other tools.
Experiments using PLSDA as a feature selector
We used 3 sets of methods for generating the synthetic points. In the first set, we consider point sets that are linearly separable. In the second data set we assume that the membership of the points in a class is determined by whether selected signal features lie in prespecified ranges. Finally, we perform experiments with clustered points.
Experiments with Linearly Separable Points
For these experiments we assume that the data consist of a collection of n random points with s signal features and m−s noise features. They are labeled as belonging to one of two classes using a randomly selected linear separator given as a function of only the signal features. The experiments were meant to test the ability of PLSDA (used as a feature extractor) to correctly identify the signal features. The performance scores shown in Fig. 3 were averaged over 100 repeats. Note that the linear model used implements the following rule \({\mathcal R}_{1}\), where C is a constant set to 0.5:
Two sets of experiments were performed. In the first set, s was fixed at 10, but n and m were varied (see Fig. 3). In the second set n was fixed at 200, but s and m were varied (see Additional file 1). PCA consistently outperformed PLSDA in all these experiments with linear relationships. Also, when the number of samples was increased, the performance of PCA improved, because there is more data from which to learn the relationship. However, it did not help PLSDA, because the model is not designed to capture this kind of relations. Note that PCA is successful only because the features that are the signal are the only ones correlated.
The loading vector is a reflection of what PCA or PLSDA guessed as the linear relationship between the features. We, therefore, set out to verify how far was the linear relationship that was guessed by the tools used. Even if the tools picked many noise features, we wanted to see how they weighted the noise and signal features they picked. Toward this goal, we ran an extra set of experiments with the model shown above to see if the loading vector from PLSDA indicated a better performance than what might be suggested by Fig. 3. Note that ideally the loading vector should have zeros for the noise features and ones for the signal features. We computed the cosine distance between the loading vector computed in the experiment and the vector reflected by the true relationship. As shown in Additional file 2, we see that the loading vectors of both PCA and PLSDA failed to reflect the true relationship. These experiments were performed using n=200 averaged over 100 repetitions. Even though PCA successfully selected many of the signal features during feature selection, it was unable to get sufficiently close to the underlying linear relationship, perhaps due to the compositional nature of the signal variables, which gives rise to correlations.
Other experiments carried out with the same results include changing the magnitude of constant in the inequality and changing the relationship from a linear inequality to two linear equalities, i.e., the points lie on two hyperplanes.
Cluster model
For these experiments, the signal features of the points were generated from a clustered distribution with two clusters separated by a prespecified amount. All noise features were generated from a uniform distribution. The R package clusterGeneration was used for this purpose, which also allows control over the separation of the clusters. Cluster separation between the clusters was varied in the range [−0.9,0.9]. Thus when the points are viewed only with the noise features, they appear like a uniform cloud, and when viewed only with the signal features, the members of the two classes are clustered. Note that cluster separation of 0.9 will appear as indistinguishable clusters, while a separation of 0.9 will appear as wellseparated clusters. The experiments were executed with s= 10, n= 200, averaged over 100 repetitions.
The executions with clustered data showed PLSDA to be clearly superior to PCA. As shown in Fig. 4, while it is true that the difference narrows when the number of samples is made very large or the clusters are widely separated (i.e., cleanly separated data),it still remains significant. PLSDA is able to select the correct hyperplane even with few samples and even when the separation between the clusters is low (values close to 0). PCA needs both an unreasonably large number of samples and very well separated clusters to perform respectably in comparison. However, data with high separation values are embarrassingly simple to analyze with a number of competing methods.
Interval model
In this set of experiments the rules that determined class membership are often encountered in biological data sets. We used two different methods to generate data from this model. In the first one, we constrained the signal features and in the second we constrained the noise ones. To generate such data sets, members of one class had the constrained features selected uniformly at random from prespecified intervals, while all other features were generated from a uniform distribution in the range [0,1].
We divided the range [0,1] into subintervals of width 1/p. Experiments were carried out with p= 3, 5 and 10. Depending on the experiment, signal and noise feature were assigned to either a subinterval of width 1/p or the entire interval of [0,1].
The results are shown in Additional file 3. When the signal features are constrained, PLSDA consistently outperforms PCA. This due to the strong correlation between the signal features for class members that PLSDA is able to detect. On the other hand, when the noise features are constrained, PCA consistently outperforms PLSDA. The latter performs poorly when the number of signal features is 1 and p=3, because the distribution of values for the single signal is not very different from the distribution of the noise.
Experiments as a classifier
Our final set of experiments with synthetic data was to see how PLSDA fared as a classifier. The following experiments were executed 100 times each, with 10 signal features. For the crossvalidation error calculation, 5 folds and 10 repetitions were used. In all of the experiments there is a correspondence between a high performance as feature selector and a low CV error.
As shown in Additional file 4a for the linear relationship model, its performance is no better than chance for a 2class experiment. This corroborates the poor performance of PLSDA as a feature selector for this model.
For the results with the cluster model shown in Additional file 4b, the CV error is almost 0 in every case, except when the number of samples is low, which is again consistent with what we saw in the feature selection experiments. The performance gets noticeably worse when, in addition to a low number of samples, the number of noise features is large. This is because the signal is hidden among many irrelevant features, something that one has come to expect with all machine learning algorithms. Additional file 4c and d show the results for the interval model. As in the case of the feature selection experiments, both versions performed roughly the same, classifying much better than chance and having their best performance when the number of samples was large and the number of noise features was low, as expected.
Comparisons with other methods
To compare the PLSDA with other known feature selectors, we applied 3 more methods to the previous data models: Independent Component Analysis (ICA), as a feature extraction method that transforms the input signals into the independent sources [21]. Sparse Principal Component Analysis (SPCA) via regularized Singular Value Decomposition (SVD) [22] was built by adding sparsity constraint. Regularized Linear Discriminant Analysis (RLDA) was computed by using L_{2} regularization to stabilize eigendecomposition in LDA [23].
We found that PCAbased algorithms (PCA and SPCA) have similar overall performance among the three experiments. The same happens with LDAbased models (RLDA and sPLSDA).
As Additional files 5 and 6 show, PLSDA, ICA and RLDA are not able to detect linear relationships, while SPCA and PCA are. For the interval model with p=3, either to constrain signal or noise doesn’t seem to change the behavior of the LDAbased models, being outperformed by PCA when noise is constrained as shown Additional files 7 and 8. The performance of every method except for ICA goes down as s becomes small. The performance of ICA depends on the number of noise features for both the interval and linear models. In the cluster model experiment as shown in Additional file 9, SPCA performs better than PCA as the separation between the cluster gets higher. The separation between the cluster does not affects the performance of ICA, that stays near 0. RLDA’s and PLSDA’s performance excel, with similar behavior (Fig. 5).
Novel analysis of a real dataset
Bacterial Vaginosis (BV) is the most common form of vaginitis affecting a large number of women across the world [24]. BV is associated with an imbalance of the vaginal flora and damage to the epithelial and mucus layer compromising the body’s intrinsic defense mechanisms. This can result in adverse sequelae and increasing the risk of many STIs [25].
In a landmark paper, human vaginal microbial communities were classified into five community state types (CSTs) [26]. CSTs I, II, III, and V are dominated by different Lactobacillus species, whereas CST IV has no specific dominant species and is regarded as the heterogeneous group. While this CST classification has enhanced our understanding of bacterial vaginosis [26–28], a quantitative method to reliably distinguish the CSTs was not available until the development of the specificity aggregation index [29] based on the species specificity [30]. The values of this index range from 0, indicating that the species is absent in that CST to 1, indicating that that OTU is always detected and only detected in that CST.
We used the abundance matrix from [26] (394 samples, 247 OTUs), and with a one vs all approach we devised a simple scheme to differentiate each CST from all of the others using the abundance of each taxon. The importance of each feature given by the specificity index computed in [29] was used as the ground truth. Only the top 10 OTUs for each CST were considered and their importance values were normalized.
Results are summarized in Fig. 6. As PLSDA and PCA return a ranked list of features, a varying threshold on the percentage of features selected is shown on the X axis of Fig. 6. The Y axis represents the sum of the specificity indices achieved by the best features at that cutoff. Note that by just selecting half of the features, a cumulative specificity of 0.9 is achieved by both methods. PLSDA reaches specificity values over 0.8 with less than 5 features selected, which means that in all of the cases, PLSDA’s top features are indeed the right set of features. In contrast, PCA’s specificity has a slower growth at the beginning (selects the wrong features), but when half of them are selected both methods achieve the same specificity.
Discussion
Our work sheds light on the kind of relationships and data models with which PLSDA can be effective both as a feature selector as well as a classifier. In particular, we claim that when classes are determined by linear or nonlinear relationships, PLSDA provides almost no insight into the data. But it is effective when the classes have a clustered distribution on the signal features, even when these features are hidden among a large number of noise attributes. PLSDA retains a strong performance also when the classes are contained in northotopes (i.e., rectangular boxes in the subspace of the signal features).
In all of the experiments carried out there was a correspondence between performance of the tools as feature selector and CV error. This reinforces the argument that the CV error is an excellent way to differentiate a good model from a bad one and every paper using PLSDA must report it to have any validity. Moreover, justbychance good behaviors are commonplace when using this tool because the sparsity of the data grows increasingly faster with the number of dimensions and it becomes easier for PLSDA to find a perfectly separating hyperplane.
Also even though PCA ignores the information regarding the class labels of the samples, it can be remarkably effective as a feature selector for classification problems. In some cases, it outperforms PLSDA which is made aware of the class labels in its input.
Conclusions
The obvious conclusion from our experiments is that it is a terrible idea to use PLSDA blindly with all data sets. In spite of its attractive ability to identify features that can separate the classes, it is clear that any data set with sufficiently large number of features is separable and that most of the separating hyperplanes are just “noise”. Thus using it indiscriminately would turn into a “golden hammer”, i.e., an oftused, but inappropriate tool. Fortunately, the use of CV would readily point to when it is being used ineffectively.
Our work sheds light on the kind of relationships and data models with which PLSDA can be effective and should be used both as a feature selector as well as a classifier in the case that the underlying model of the data is known or can be guessed. When it is not possible, one should rely on the CV error and use extreme care when making conclusions and extrapolations.
Also, one should take advantage of the multitude of tools available and use different methods depending on the dataset, as the simple PCA was able to outperform PLSDA depending on the conditions.
Availability of data and materials
All the code that generated and analyzed the datasets can be downloaded from https://github.com/DaniRuizPerez/SoYouThinkYouCanPLSDA_Public. All the 3D figures shown in this paper as well as the supplementary ones can be viewed interactively at http://biorg.cs.fiu.edu/plsda. The dataset analyzed are derived from the following published article: [26]
Abbreviations
 CV:

Crossvalidation
 fn:

False negatives
 fp:

False positives
 ICA:

Independent component analysis
 PC:

Principal components
 PCA:

Principal component analysis
 PLSDA:

Partial leastsquares discriminant analysis
 RLDA:

Regularized linear discriminant analysis
 SPCA:

Sparse principal component analysis
 sPLSDA:

Sparse partial leastsquares discriminant analysis
 SVD:

Singular value decomposition
 tn:

True negatives
 tp:

True positives
References
Ståhle L, Wold S. Partial least squares analysis with crossvalidation for the twoclass problem: A monte carlo study. J Chemometrics. 1987; 1(3):185–96.
Barker M, Rayens W. Partial least squares for discrimination. J Chemometrics. 2003; 17(3):166–73.
Gottfries J, Blennow K, Wallin A, Gottfries C. Diagnosis of dementias using partial least squares discriminant analysis. Dementia Geriatric Cognit Disorders. 1995; 6(2):83–8.
Worley B, Powers R. Multivariate analysis in metabolomics. Curr Metabol. 2013; 1(1):92–107.
Worley B, Halouska S, Powers R. Utilities for quantifying separation in PCA/PLSDA scores plots. Anal Biochem. 2013; 433(2):102–4.
Rohart F, Gautier B, Singh A, Le Cao KA. mixOmics: An R package for ’omics feature selection and multiple data integration. PLoS computational biology. 2017; 13(11):108597.
Eriksson L, Antti H, Gottfries J, Holmes E, Johansson E, Lindgren F, Long I, Lundstedt T, Trygg J, Wold S. Using chemometrics for navigating in the large data sets of genomics, proteomics, and metabonomics (gpm). Anal Bioanalyt Chem. 2004; 380(3):419–29.
Christin C, Hoefsloot HC, Smilde AK, Hoekman B, Suits F, Bischoff R, Horvatovich P. A critical assessment of feature selection methods for biomarker discovery in clinical proteomics. Mole Cell Proteom. 2013; 12(1):263–76.
Nguyen DV, Rocke DM. Classification of Acute Leukemia Based on DNA Microarray Gene Expressions Using Partial Least Squares. Linux Journal. 2002;:109–24. https://doi.org/10.1007/9781461508731_9.
Tan Y, Shi L, Tong W, Hwang GG, Wang C. Multiclass tumor classification by discriminant partial least squares using microarray gene expression data and assessment of classification models. Comput Biol Chem. 2004; 28(3):235–43.
Botella C, Ferré J, Boqué R. Classification from microarray data using probabilistic discriminant partial least squares with reject option. Talanta. 2009; 80(1):321–8.
Brereton RG, Lloyd GR. Partial least squares discriminant analysis: taking the magic away. J Chemometrics. 2014; 28(4):213–25.
Westerhuis JA, Hoefsloot HC, Smit S, Vis DJ, Smilde AK, van Velzen EJ, van Duijnhoven JP, van Dorsten FA. Assessment of PLSDA cross validation. Metabolomics. 2008; 4(1):81–9.
Kjeldahl K, Bro R. Some common misunderstandings in chemometrics. J Chemometrics. 2010; 24(78):558–64.
Chun H, Keleş S. Sparse partial least squares regression for simultaneous dimension reduction and variable selection. J Royal Stat Soc: Ser B (Stat Methodol). 2010; 72(1):3–25.
Lê Cao KA, Boitard S, Besse P. Sparse pls discriminant analysis: biologically relevant feature selection and graphical displays for multiclass problems. BMC Bioinformatics. 2011; 12(1):253.
Chung D, Keles S. Sparse partial least squares classification for high dimensional data. Stat Appl Genet Mole Biol. 2010; 9(1).
Le Cao KA, Rohart F, Gonzalez I, Dejean S, Gautier B, Bartolo F. mixOmics: Omics data integration project. R package, version. 2017.
Saccenti E, Timmerman ME. Approaches to sample size determination for multivariate data: Applications to pca and plsda of omics data. J Proteome Res. 2016; 15(8):2379–93. https://doi.org/10.1021/acs.jproteome.5b01029. PMID: 27322847.
Ambroise C, McLachlan GJ. Selection bias in gene extraction on the basis of microarray geneexpression data. Proc Nat Acad Sci. 2002; 99(10):6562–6.
Hyvarinen A KJ, E O. Independent Component Analysis: Wiley; 2001. ISBN 9780471405405.
Shen H, Huang JZ. Sparse principal component analysis via regularized low rank matrix approximation. J Multivariate Anal. 2008; 99(6):1015–34. https://doi.org/10.1016/j.jmva.2007.06.007.
Friedman JH. Regularized discriminant analysis. J Am Stat Assoc. 1989; 84(405):165–75. https://doi.org/10.1080/01621459.1989.10478752. http://arxiv.org/abs/https://www.tandfonline.com/doi/pdf/10.1080/01621459.1989.10478752.
Allsworth JE, Peipert JF. Prevalence of bacterial vaginosis: 2001–2004 national health and nutrition examination survey data. Obstetrics Gynecol. 2007; 109(1):114–20.
Cherpes TL, Meyn LA, Krohn MA, Lurie JG, Hillier SL. Association between acquisition of herpes simplex virus type 2 in women and bacterial vaginosis. Clin Infect Dis. 2003; 37(3):319–25.
Ravel J, Gajer P, Abdo Z, Schneider GM, Koenig SSK, McCulle SL, Karlebach S, Gorle R, Russell J, Tacket CO, Brotman RM, Davis CC, Ault K, Peralta L, Forney LJ. Vaginal microbiome of reproductiveage women. Proc Nat Acad Sci. 2011; 108(Supplement 1):4680–7. https://doi.org/10.1073/pnas.1002611107.
Gajer P, Brotman RM, Bai G, Sakamoto J, Schütte UME, Zhong X, Koenig SSK, Fu L, Ma ZS, Zhou X, Abdo Z, Forney LJ, Ravel J. Temporal dynamics of the human vaginal microbiota. Sci Trans Med. 2012; 4(132):132–5213252. https://doi.org/10.1126/scitranslmed.3003605. http://arxiv.org/abs/http://stm.sciencemag.org/content/4/132/132ra52.full.pdf.
Ma B, Forney LJ, Ravel J. Vaginal microbiome: Rethinking health and disease. Ann Rev Microbiol. 2012; 66(1):371–89. https://doi.org/10.1146/annurevmicro092611150157. PMID: 22746335.
Ma ZS, Li L. Quantifying the human vaginal community state types (csts) with the species specificity index. PeerJ. 2017; 5:3366.
Mahendra M, Samuel P, Dieter E. Microbial ecosystems are dominated by specialist taxa. Ecol Lett; 18(9):974–82. https://doi.org/10.1111/ele.12478. http://arxiv.org/abs/https://onlinelibrary.wiley.com/doi/pdf/10.1111/ele.12478.
Acknowledgments
The authors thank the members of the Bioinformatics Reseach Group (BioRG) for their useful comments during the course of this research.
About this supplement
This article has been published as part of BMC Bioinformatics Volume 21 Supplement 1, 2020: Selected articles from the 8th IEEE International Conference on Computational Advances in Bio and medical Sciences (ICCABS 2018): bioinformatics. The full contents of the supplement are available online at https://bmcbioinformatics.biomedcentral.com/articles/supplements/volume21supplement1.
Funding
This work was partially supported by grants from the Department of Defense Contract W911NF1610494, NIH grant 1R15AI12871401, and NIJ grant 2017NEBX0001. Publication costs were funded by personal funds.
Author information
Authors and Affiliations
Contributions
DRP was the major contributor of this work. DRP and GN conceived and designed the experiments. DRP implemented the experiments. DRP and HG executed the experiments. PM and KM provided the vaginal dataset and helped with the biological analysis. DRP, GN and HG contributed in writing the manuscript, All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Additional file 1
Figure S1. Performance for linearly separable points model, varying signal and noise.
Additional file 2
Figure S2. Performance for linearly separable points model with the cosine model, varying signal and noise.
Additional file 3
Figure S3. Performance table for different configurations of the interval model.
Additional file 4
Figure S4. Classification accuracy for the different data models.
Additional file 5
Figure S5. Performance for linearly separable points model, varying signal and noise.
Additional file 6
Figure S6. Performance for linearly separable points model, varying samples and noise.
Additional file 7
Figure S7. Performance of other methods, signal constrained interval with p=3.
Additional file 8
Figure S8. Performance of other methods, noise constrained interval with p=3.
Additional file 9
Figure S9. Performance of other methods for the cluster model, High number of samples.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
RuizPerez, D., Guan, H., Madhivanan, P. et al. So you think you can PLSDA?. BMC Bioinformatics 21 (Suppl 1), 2 (2020). https://doi.org/10.1186/s1285901933107
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1285901933107