- Research
- Open Access
- Published:
Learning a confidence score and the latent space of a new supervised autoencoder for diagnosis and prognosis in clinical metabolomic studies
BMC Bioinformatics volume 23, Article number: 361 (2022)
Abstract
Background
Presently, there is a wide variety of classification methods and deep neural network approaches in bioinformatics. Deep neural networks have proven their effectiveness for classification tasks, and have outperformed classical methods, but they suffer from a lack of interpretability. Therefore, these innovative methods are not appropriate for decision support systems in healthcare. Indeed, to allow clinicians to make informed and well thought out decisions, the algorithm should provide the main pieces of information used to compute the predicted diagnosis and/or prognosis, as well as a confidence score for this prediction.
Methods
Herein, we used a new supervised autoencoder (SAE) approach for classification of clinical metabolomic data. This new method has the advantage of providing a confidence score for each prediction thanks to a softmax classifier and a meaningful latent space visualization and to include a new efficient feature selection method, with a structured constraint, which allows for biologically interpretable results.
Results
Experimental results on three metabolomics datasets of clinical samples illustrate the effectiveness of our SAE and its confidence score. The supervised autoencoder provides an accurate localization of the patients in the latent space, and an efficient confidence score. Experiments show that the SAE outperforms classical methods (PLS-DA, Random Forests, SVM, and neural networks (NN)). Furthermore, the metabolites selected by the SAE were found to be biologically relevant.
Conclusion
In this paper, we describe a new efficient SAE method to support diagnostic or prognostic evaluation based on metabolomics analyses.
Background
Deep neural networks have proven their effectiveness in bioinformatics for classification and feature selection [1,2,3,4,5]. They have also been recently used in metabolomic studies [6,7,8,9,10]. Classical stacked autoencoders [11] were used recently in metabolomic studies [12].
Autoencoders were introduced within the field of neural networks decades ago, their most efficient application at the time being dimensionality reduction [13, 14]. Autoencoders have also been used for denoising different types of data [11] to extract relevant features. One of the main advantages of the autoencoder is the projection of the data in the low dimensional latent space.
These autoencoder models include variational autoencoders (VAE) [15]. VAE networks encourage the latent space to fit a prior distribution, like a Gaussian. This can alter the accuracy of the model. In order to cope with this issue, some recent papers have proposed latent spaces with more complex distributions (e.g. mixtures of Gaussians [16]) on the latent vectors, but they are non-adaptive and unfortunately may not match the specific data distribution.
In this work, we relaxed the parametric distribution assumption on the latent space to learn a non-parametric data distribution of clusters [17]. Our network encourages the latent space to fit a distribution learned with the clustering labels rather than a parametric prior.
Recent untargeted metabolomic methods using liquid chromatography coupled with high resolution mass spectrometry (LC-MS/MS) allow for fast and high-resolution detection of massive amounts of metabolites. Metabolomics is a very promising omics field for fundamental research in biology as well as for clinical research applications. Indeed, metabolomics can be used to reveal new biomarkers of physiological or pathological states [18,19,20,21], and could be particularly useful for personalized medicine [22, 23].
In this study, we described a new SAE method using structured constraints and compare its performances to classical machine learning and Neural Network methods, when applied to three clinical metabolomic databases.
Methods
A New supervised Autoencoder (SAE) framework
Projecting the samples in the lower dimension latent space is crucial to separate them accurately. Herein we propose to use a neural network autoencoder framework.
Let us recall that the encoder part of the autoencoder maps feature-points from a high dimensional space to a low dimensional latent space, and that the decoder maps feature points from that low dimensional space to a high dimensional space.
Figure 1 depicts the main constitutive blocks of our proposed approach. We have added to our SAE a “soft max” block to compute the classification loss.
Let X be the dataset, as an \(m \times d\) data matrix made of m line samples \(x_1,\dots ,x_m\). Let \(y_{i}=j, j \in [1...k]\) be the label, indicating that the sample \(x_i\) belongs to the j-th cluster. Let Z, be the latent space, \(\widehat{X}\) the reconstructed data (Fig. 1) and W the weights of the neural network.
The goal is to compute the weights W minimizing the total loss, which depends on both the classification loss and the reconstruction loss. Thus, we propose to minimize the following criterion to compute the weights W of the autoencoder (see [17] for details).
Where \(\phi ( Z,Y)\) is the classification loss in the latent space and \(\psi (\widehat{X}-X)\) is the reconstruction loss.
The parameter \(\lambda\) controls the weight of the reconstruction loss in the criterion. We used the Cross Entropy Loss for the classification loss function \(\phi\). We used the robust Smooth \(\ell _1\) (Huber) Loss [24] as the reconstruction loss function \(\psi\), as it is more robust to outliers than the classical Mean Squared Error (MSE) loss. The dimension of the latent space is defined by the number of clusters.
Structured constraints, sparsity and feature selection
The basic idea for feature selection is to use a sparse regularizer that forces some coefficients to be zero. To achieve feature selection, classically, the Least Absolute Shrinkage and Selection Operator (LASSO) formulation [25,26,27,28,29] is used to add an \(\ell _1\) penalty term to the classification loss. However the LASSO is computationally expensive [26, 27]. Thus, we used a feature selection method by optimizing a criterion under constraints [30].
Let us recall that the classical \(\ell _{2}\) norm constraint does not induce any sparsity. Moreover the “group Lasso \(\ell _{2,1}\) constraint” induces small sparsity [31] and the \(\ell _{1}\) constraint induces unstructured sparsity [32, 33]. Thus we used \(\ell _{1,1}\) constrained regularization penalty \(\Vert W\Vert _1^1 \le \eta\) for feature selection [17].
Algorithm
We compute the \(\ell _{1,1}\) constraint with the following algorithm: we first compute the radius \(t_i\) and then project the rows using the \(\ell _1\) adaptive constraint \(t_i\).
Following the work developed by [34], which proposed a double descent algorithm, we replaced the thresholding by our \(\ell _{1,1}\) projection and devised a new double descent algorithm (See Barlaud and Guyard [35]) as follows :

Implementation
Pytorch implementation of our supervised autoencoder
We implemented our sparse supervised autoencoder model in the Pytorch framework. The losses are averaged across observations for each batch. We chose the ADAM optimizer [36], as the standard optimizer in PyTorch. We used the Netbio SAE, a linear fully connected network (LFC), which has an input layer of d neurons, 1 hidden layer of 96 neurons followed by a ReLU activation function, and a latent layer of dimension 2 (the number of classes). The parameter \(\eta\) is determined by the maximum accuracy after cross-validation.
We compared the Netbio SAE with a classical linear fully connected Neural Network (NN) with the same structure.
We used the captum package [37] to compute the feature weights of the SAE.
We provide comparisons with a PLS-DA using 4 components, with Random Forests using 400 estimators and a maximum depth of 3 (using the Gini importance (GI) for feature ranking), and with a support vector classifier (SVM) with a linear kernel. For the SVM, we perform a cross-validation grid search to find the best regularization parameter C.
We provide the statistical evaluation (Accuracy, AUC, and F1 score) using a 4-fold cross validation process: the dataset is randomly divided into four parts, and trained on three of the four splits. The metrics are computed on the remaining test split, which wasn’t used during training. We then repeat this process three more times, leaving a different split as the test set each time. The final metrics given in this paper are averages over the four cross-validation steps, over three different random seeds (12 different testing/training splits in total).
We compare the performances of the different methods using the F1 Score. The F1 Score is the weighted average of Precision and Recall. Therefore, this score takes both false positives and false negatives into account. The F1 score is more relevant than accuracy, especially for unbalanced datasets.
The computation of the statistical metrics, the classifiers, the cross-validation function and the grid search were all provided by the scikit-learn machine learning python package. The python code is available on github: https://github.com/CyprienGille/Supervised-Autoencoder.
Diagnosis with confidence score
One of the main advantages of an autoencoder is the projection of the data in the latent space, which can easily be visualized if the latent space is of dimension 2.Footnote 1 Thanks to this, we propose a clinical diagnosis simulation: having trained a network on a database of patients, we can predict a diagnosis with a confidence score for new patients. To perform this simulation, we removed a patient from each of the k classes from the databases. We then trained the SAE on (n-k) patients and we fed the k “test” patients through the best net. We thus obtained a visualization of the projections of these new “test” patients in the latent space as well as their classification with a confidence score (see Figs. 4, 10 and 7).
The clinician then has an accurate and reliable system to help with the diagnosis. Indeed, in addition to obtaining the confidence score for the diagnosis, the clinician can see where the patient is located among the others in the database and have a critical evaluation of the prediction (the clinician can easily see if a patient stands out).
Evaluation on 3 clinical metabolomics databases
The SAE was tested on three different metabolomic datasets : the “LUNG” , “BREAST”, and “BRAIN” datasets.
The LUNG dataset was published by Mathe et al. [38] and is available at MetaboLights (study identifer MTBLS28). It includes metabolomics data concerning urine samples from 469 Non-Small Cell Lung Cancer (NSCLC) patients prior to treatment and 536 controls collected from 1998 to 2007 in seven hospitals and in the Department of Motor Vehicles (DMV) from the greater Baltimore, Maryland area. Urine samples were analyzed using an unbiased metabolomics LC-MS/MS approach. Mathe et al. used Random Forests to classify patients as lung cancer patients or controls [38]. The aim was to create a new screening test for lung cancer, based on metabolomics data from urine. Lung cancer is one of the most common cancers and it is well established that early diagnosis is crucial for treatment. An efficient screening method based on urinary metabolomics could be of great benefit.
The BREAST dataset was kindly provided by Dr. Jan Budczies and can be found in the supplementary material of Budczies et al. [39]. It includes metabolomics data concerning 271 breast tumor samples: 204 tumors with over-expression of estrogen receptors (ER) and 67 tumors without over-expression of ER. Metabolomics analysis was performed using Gas chromatography followed by time of flight mass spectrometry as described in [40].
The BRAIN dataset was obtained through a study performed in our lab\(^{1*}\). It includes metabolomic data obtained on 88 frozen samples of glial tumors. The samples were retrospectively collected from two declared biobanks from the Central Pathology Laboratory of the Hospital of Nice and from the Center of Biological Resources of Montpellier (Plateforme CRB-CHUM). Consent or non-opposition was verified for every participant. Tumors were analyzed using Liquid Chromatography coupled to tandem Mass Spectrometry (LC-MS/MS) in an unbiased metabolomics approach. The details of the analysis are available in Additional file 1.
With this dataset, the goal was to create a model that accurately discriminated between mutated isocitrate dehydrogenase (IDH) and IDH wild-type glial tumors. The dataset includes (38 IDH wild-type tumors and 50 IDH-mutant tumors). This mutation is a key component of the World Health Organization classifcation of glial tumors [29]. The mutational status is usually assessed by IDH1 (R132H)-specifc (H09) immunohistochemistry. Yet this technique can lead to False-Negative results, which can only be identified by sequencing. Thus an accurate metabolomic based test, able to assess the IDH mutational status, could be a promising additional diagnostic tool.
The characteristics of the three metabolomic datasets are presented in Table 1. We chose to study these databases for their diversity both in terms of the number of features and number of patients, to test the robustness of our method on different types of databases.
The LUNG dataset includes a very large number of patients (1,005), with an equivalently large number of features (2,944), and 2 classes. The BREAST dataset includes a midsize number of patients (271), with a small number of features (161), and 2 classes. The BRAIN dataset includes a limited number of patients (88), with a much higher number of features (7,022), and 2 classes.
Results
LUNG dataset
Statistical performances
As shown in Table 2 our SAE outperformed PLS-DA, Random Forests, SVM and NN by 4.58, 9.58, 9.63 and \(2.74\%\) respectively for the F1 score. Note that we checked that increasing the number of trees for Random forests from 100 to 400 resulted in a small improvement in accuracy of only \(1 \%\) while the computational cost increased by a factor of 3. The performances of the SAE were a little better when using an \(\ell _1\) loss than when using an \(\ell _2\) loss.
Feature selection using the \(\ell _{1,1}\) structured constraint
Figure 2 shows the matrix (\(d \times n\)) of the network connections between the input layer (d feature neurons) and the hidden layer (n neurons).
It shows the benefit of using the \(\ell _{1,1}\) constraint: The \(\ell _{1,1}\) constraint selects features while the constraint \(\ell _{1}\) selects only weights of features. All the following results are given with the \(\ell _{1,1}\) constraint.
As shown in Table 3, all methods selected metabolite “MZ 264.121”, which most likely corresponds to creatine riboside (expected m/z value in the positive mode: 264.1190). Note that the SVM selected metabolite “MZ 264.121” at rank 3. Metabolite “MZ 308.098”, which most likely corresponds to N-acetylneuraminic acid, was only selected by the SAE and the NN at rank 2 and 3, respectively. These metabolites were described by Mathé et al. [38] as the most important metabolites to discriminate between lung cancer patients and healthy individuals. Note that the author of RF proposes two measures for feature ranking, the variable importance (VI) and Gini importance (GI): a recent study showed that if predictors are categorical, or real with multimodal Gaussian distributions, both measures are biased [41].
As shown in Fig. 3, selected features were not significantly correlated. The highest correlation found was between MZ 308.09 and MZ 332.09, with a Pearson coefficient of 0.67. Both features correspond to adducts of N-acetylneuraminic acid (MZ 308.09 being the [M+H]+ adduct and MZ 332.09 the [M+Na]+ adduct).
Diagnosis in the latent space with a confidence score
As shown in Fig. 4, the two classes are well separated in the latent space of the SAE. Furthermore, the red and green squares show the location of the two random “test” patients in the SAE’s latent space. The red patient is at the heart of the class distribution and the green patient is close to the edge of the other class. This is important for a clinician’s assessment of the result. Moreover, the distribution plot shows the nearly perfect separability of the distributions calculated with the SAE, which means most of the patients were diagnosed with a high degree of confidence. The patient represented by the red square was classified in class 0 with a confidence score of 0.94 and the patient represented by the green square was labeled class 1 with a confidence score of 0.70. Both predicted labels were correct.
BREAST dataset
Statistical performances
As shown in Table 4 our SAE outperformed PLS-DA, Random Forests, SVM and NN by 9.16, 14.1, 9.11 and \(2.23\%\) respectively for the F1 score. The performances of the SAE were a little better when using an \(\ell _1\) loss than when using an \(\ell _2\) loss.
Feature selection using the \(\ell _{1,1}\) structured constraint
Figure 5 shows the matrix ( \(d \times n\)) of the network connections between the input layer (d feature-neurons) and the hidden layer (n neurons). It shows the benefit of using the \(\ell _{1,1}\) constraint: The \(\ell _{1,1}\) constraint selects features, while the constraint \(\ell _{1}\) selects only weights of features.
As shown in Table 5, the SAE and the NN selected the same top five metabolites (beta-alanine, xanthine, uracil, glutamic acid). These metabolites have already been shown to have significantly different concentrations in ER− breast tumors compared to ER+ breast tumors in the original paper by Budczies et al. [39]. Increased concentrations of glutamic acid and 2-hydroxyglutaric acid indicate higher glutaminolysis, a key feature of metabolic changes in cancer cells. As shown in Budczies et al. [39], increased concentrations of uracil, xanthine and beta-alanine levels are related to higher hexokinase 3, xanthine dehydrogenase and 4-aminobutyrate aminotransferase expressions, respectively.
As shown in Fig. 6, selected features were highly correlated.
Prognosis in the latent space with confidence score
Figure 7 (left), shows the accurate separation of the two classes in the latent space of the SAE. The red and green squares show the location of the two random “test” patients in the SAE’s latent space. The patient represented by the red square was classified in class 0 with a confidence score of 0.55 and the patient represented by the green square was labeled class 1 with a confidence score of 0.80. Both predictions are correct. Figure 7 (right) shows the separability of the distributions calculated with the SAE.
BRAIN dataset
Statistical performances
Table 6 shows that, despite the small number of patients, the supervised autoencoder outperformed PLS-DA, Random Forest, SVM and NN by 8.78, 4.61, 6.15 and \(18.47 \%\) respectively for the F1 score. For this base with few patients the performance of NNs collapses as reported in the literature. As for the other databases, the performances of the SAE were a little better when using an \(\ell _1\) loss than when using an \(\ell _2\) loss.
Feature selection using the \(\ell _{1,1}\) structured constraint
Figure 8 shows the matrix ( \(d \times n\)) of the network connections between the input layer (d feature-neurons) and the hidden layer (n neurons). It shows the benefit of using the \(\ell _{1,1}\) constraint: The \(\ell _{1,1}\) constraint selects features, while the constraint \(\ell _{1}\) selects only weights of features.
As expected, the top features selected by each method (shown in Table 7) correspond mainly to different isotopes and adducts of 2-hydroxyglutarate (marked in bold). The features selected using the SAE were all different adducts of this specific product of IDH-mutated cells. Indeed, POS_MZ132.03 and POS_MZ131.03 correspond to the [M+H-H2O]+ adduct of 2-hydroxyglutarate with one 13C isotope for the first ion. POS_MZ171.02 is the [M+Na]+ adduct, NEG_MZ147.02 is the [M-H]- and POS_MZ86.03 is the [M+Na+H]2+ adduct. NEG_MZ148.03 is the [M-H]- adduct of 2-hydroxyglutarate with one 13C isotope. POS_MZ173.03 is the [M+Na]+ adduct with two 13C isotope. Finally, POS_MZ149.04 is the [M+H]+ adduct ion of 2-hydroxyglutarate. As expected, and shown in Fig. 9, these features, all corresponding to adducts of 2-hydroxyglutarate, were highly correlated.
Diagnosis in the latent space with confidence score
Figure 10 (left), shows the nearly perfect separation of the two classes in the latent space of the SAE. Furthermore, the red and green squares show the location of the two random “test” patients in the SAE’s latent space. The patient represented by the red square was classified in class 0 with a confidence score of 0.75 and the patient represented by the green square was labeled class 1 with a confidence score of 0.62. Both predictions were correct. Figure 10 (right) shows the peak separability of the distributions calculated with the SAE. It shows that most patients will have a good prediction with a high degree of confidence.
Discussion
Thus, we have shown that our SAE outperformed classical machine learning methods and NN for classification of metabolomics data, while providing reliable confidence score for the predictions and performing relevant feature selection.
The real distributions of many datasets, including metabolomics datasets, are far more complex than multi-gaussian mixtures.Thus we chose to use a non-parametric supervised autoencoder (SAE) rather than a classical autoencoder that assumes a latent space modeling [42, 43] and force a multi-gaussian distribution upon the data.
Regardless of data size and feature space dimensions, the SAE outperforms all other methods (PLS-DA, Random Forests, SVM and NN). As expected, the NN also outperformed classical methods (PLS-DA, Random Forests and SVM), except on small databases. Indeed, NN are known to be less accurate when trained on small numbers of samples [44, 45]. Furthermore, as anticipated, the SAE’s performances were a little better when using the Huber loss than when using the MSE. This is most likely due to the fact that the Huber loss is more robust to outliers.
The SAE provides high-level distribution visualization of the samples in the latent space, as well as their classification confidence score. This is crucial for any diagnostic tool. Indeed, these two features enable clinicians to gauge how reliable each prediction is and if a sample corresponds to a potential outlier, for which predictions should be considered with particular care.
Metabolomics is a very promising approach, particularly adapted to routine clinical practice, because metabolomics analyses are fast and relatively inexpensive. However, human metabolomics are complex data, influenced by many external and internal factors. The high number of features included in metabolomics analyses require high performance statistical methods such as our SAE to be exploited. However, no statistical method can replace the critical reasoning of a researcher to make conclusions on the statistical results and to identify potential confounding factors. To make such conclusions, the statistical method needs to have some degree of interpretability.
Interestingly, the SAE combined with a structured projection provides efficient feature selection (Tables 3, 5 and 7). This feature selection step is crucial for interpretability. Better yet, we have verified that the selected features in the LUNG, BREAST and BRAIN datasets were known to be biologically relevant metabolites. Efficient feature selection adds interpretability to the model which is crucial for metabolomic studies in biological research or clinical trials.
We have observed that selected features can have a low to very high degree of correlation. In our case, the correlated features were isotopes and adducts of metabolites with high weights for the classification. Even though multivariate methods, such as the one we have used, account for correlation, correlated features do have an impact on feature selection and the performances of the trained models. When studying metabolomics one must adapt the level of filtering. Indeed, filtering removes isotopes and adducts but can also remove important features. This must be taken into consideration when using our SAE or any other classification method for metabolomics analyses.
Conclusion
In this paper we have proposed a new and efficient classification method for metabolomics datasets, based on the representation of data on the latent space of a new supervised autoencoder (SAE). In clinical applications, our method provides a diagnosis score for each patient’s predicted class. Moreover, from a statistical point of view (Accuracy, AUC, F1 score) our SAE outperformed PLS-DA, Random Forest, SVM, and NN while selecting biologically relevant features.
Availability of data and materials
We implemented the code with python. Functions and scripts are freely available at https://github.com/CyprienGille/Supervised-Autoencoder. Furthermore the metabolomic databases are also available.
Notes
If the latent space is of dimension \(k>2\), we can project the latent space on a 2D plot using a PCA.
References
Song Y, Zheng S, Li L, Zhang X, Zhang X, Huang Z, Chen J, Wang R, Zhao H, Zha Y, Shen J, Chong Y, Yang Y. Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images. IEEE/ACM Trans Comput Biol Bioinform. 2021;18:2775–80.
Ravì D, Wong C, Deligianni F, Berthelot M, Andreu-Perez J, Lo B, Yang G-Z. Deep learning for health informatics. IEEE J Biomed Health Inform. 2017;21(1):4–21.
Min S, Lee B, Yoon S. Deep learning in bioinformatics. Brief Bioinform. 2016;18(5):851–69.
Zhang L, Lv C, Jin Y, Cheng G, Fu Y, Yuan D, Tao Y, Guo Y, Ni X, Shi T. Deep learning-based multi-omics data integration reveals two prognostic subtypes in high-risk neuroblastoma. Front Genet. 2018;9:477.
Sen P, Lamichhane S, Mathema VB, McGlinchey A, Dickens AM, Khoomrung S, Orešič M. Deep learning meets metabolomics: a methodological perspective. Brief Bioinform. 2021;22:1531–42.
Alakwaa F, Chaudhary K, Garmire L. Deep learning accurately predicts estrogen receptor status in breast cancer metabolomics data. J Proteome Res. 2018;17:337–47.
Bradley W, Robert P. Multivariate analysis in metabolomics. Curr Metabol. 2013;1:92–107.
Asakura P, Date Y, Kikuchi J. Application of ensemble deep neural network to metabolomics studies. Anal Chim Acta. 2018;1037:92–107.
Mendez K, Broadhurst D, Reinke S. Application of artificial neural networks in metabolomics: a historical perspective. Metabolomics. 2019;15:1–14.
Sen P, Lamichhane S, Mathema VB, McGlinchey A, Dickens AM, Khoomrung S, Orešič M. Deep learning meets metabolomics: a methodological perspective. Brief Bioinform. 2020;22(2):1531–42.
Vincent P, Larochelle H, Lajoie I, Bengio Y, Manzagol P-A. Stacked denoising autoencoders: learning useful representations in a deep network with a local denoising criterion. J Mach Learn Res. 2010;11:3371–408.
Xiaojing F, Xiye W, Mingyang J, Zhili P, Shicheng Q. An improved stacked autoencoder for metabolomic data classification. Comput Intell Neurosci. 2021;2021:1051172.
Hinton GE, Zemel R. Autoencoders, minimum description length and Helmholtz free energy. Advances in Neural Information Processing Systems; 1994, pp. 3–10.
Goodfellow I, Bengio Y, Courville A. Deep learning, vol. 1. Cambridge: MIT press; 2016.
Kingma D, Welling M. Auto-encoding variational Bayes. In: International conference on learning representation; 2014.
Dilokthanakul N, Mediano PAM, Garnelo M, Lee MCH, Salimbeni H, Arulkumaran K, Shanahan M. Deep unsupervised clustering with Gaussian mixture variational autoencoders 2016. arXiv:1611.02648
Barlaud M, Guyard F. Learning a sparse generative non-parametric supervised autoencoder. In: Proceedings of the international conference on acoustics, speech and signal processing, TORONTO, Canada; 2021.
Yazdani H, Cheng LL, Christiani DC, Yazdani A. Bounded fuzzy possibilistic method reveals information about lung cancer through analysis of metabolomics. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2020;17(2):526–35.
Liu Y, Xu X, Deng L, Cheng K-K, Xu J, Raftery D, Dong J. A novel network modelling for metabolite set analysis: a case study on crc metabolomics. IEEE Access. 2020;8:106425–36.
Banimustafa A, Hardy N. A scientific knowledge discovery and data mining process model for metabolomics. IEEE Access. 2020;8:209964–10005.
Qi Z, Voit EO. Strategies for comparing metabolic profiles: implications for the inference of biochemical mechanisms from metabolomics data. IEEE/ACM Trans Comput Biol Bioinform. 2017;14(6):1434–45.
Long NP, Nghi TD, Kang YP, Anh NH, Kim HM, Park SK, Kwon SW. Toward a standardized strategy of clinical metabolomics for the advancement of precision medicine. Metabolites. 2020;10(2):51. https://doi.org/10.3390/metabo10020051.
Cakmak A, Celik MH. Personalized metabolic analysis of diseases. IEEE/ACM Trans Comput Biol Bioinform. 2021;18(3):1014–25.
Huber PJ. Robust statistics; 1981.
Tibshirani R. Regression shrinkage and selection via the Lasso. J R Stat Soc Ser B Methodol. 1996;58:267–88.
Hastie T, Rosset S, Tibshirani R, Zhu J. The entire regularization path for the support vector machine. J Mach Learn Res. 2004;5:1391–415.
Friedman J, Hastie T, Tibshirani R. Regularization path for generalized linear models via coordinate descent. J Stat Softw. 2010;33:1–122.
Hastie T, Tibshirani R, Wainwright M. Statistcal learning with sparsity: the lasso and generalizations. Boca Raton: CRC Press; 2015.
Li J, Cheng K, Wang S, Morstatter F, Trevino RP, Tang J, Liu H. Feature selection: a data perspective. ACM Comput Surv. 2016;50:1–45. https://doi.org/10.1145/3136625.
Barlaud M, Belhajali W, Combettes P, Fillatre L. Classification and regression using an outer approximation projection-gradient method. IEEE Trans Signal Process. 2017;65:4635–43.
Barlaud M, Chambolle A, Caillau J.-B. Classification and feature selection using a primal-dual method and projection on structured constraints. In: International conference on pattern recognition, Milan; 2020.
Condat L. Fast projection onto the simplex and the l1 ball. Math Program Ser A. 2016;158(1):575–85.
Perez G, Barlaud M, Fillatre L, Régin J-CA. filtered bucket-clustering method for projection onto the simplex and the l1-ball. Math Program. 2019;182:445–64.
Zhou H, Lan J, Liu R, Yosinski J. Deconstructing lottery tickets: zeros, signs, and the supermask. In: Wallach H, Larochelle H, Beygelzimer A, d’ Alché-Buc F, Fox E, Garnett R, editors. Advances in neural information processing systems, vol. 32. Red Hook: Curran Associates, Inc.; 2019. p. 3597–607.
Barlaud M, Guyard F. Learning sparse deep neural networks using efficient structured projections on convex constraints for green AI. In: International conference on pattern recognition, Milan; 2020.
Kingma D, Ba J. A method for stochastic optimization. In: International conference on learning representations; 2015, pp. 1–13.
Lundberg SM, Lee S.-I. A unified approach to interpreting model predictions. In: Neural information processing systems, Barcelone, Spain; 2017, p. 30
Mathé E, et al. Noninvasive urinary metabolomic profiling identifies diagnostic and prognostic markers in lung cancer. Cancer Res. 2014;74(12):3259–70.
Budczies J, Brockmöller S, Müller B, Barupal D, Richter-Ehrenstein C, Kleine-Tebbe A, Griffin J, Orešič M, Dietel M, Denkert C, Fiehn O. Comparative metabolomics of estrogen receptor positive and estrogen receptor negative breast cancer: alterations in glutamine and beta-alanine metabolism. J Proteom. 2013;94:279–88.
Budczies J, Denkert C, Müller BM, Brockmöller SF, Klauschen F, Györffy B, Dietel M, Richter-Ehrenstein C, Marten U, Salek RM, Griffin JL, Hilvo M, Orešič M, Wohlgemuth G, Fiehn O. Remodeling of central metabolism in invasive breast cancer compared to normal breast tissue—a GC-TOFMS based metabolomics study. BMC Genomics. 2012;13(1):334. https://doi.org/10.1186/1471-2164-13-334.
Altmann A, Toloşi L, Sander O, Lengauer T. Permutation importance: a corrected feature importance measure. Bioinformatics. 2010;26(10):1340–7.
Emdadi A, Eslahchi C. Auto-HMM-LMF: feature selection based method for prediction of drug response via autoencoder and hidden Markov model. BMC Bioinformatics. 2021;22:33.
Liu D, Huang Y, Nie W, Zhang J, Deng L. SMALF: miRNA-disease associations prediction based on stacked autoencoder and XGBoost. BMC Bioinformatics. 2021;22:219.
Markham IS, Rakes TR. The effect of sample size and variability of data on the comparative performance of artificial neural networks and regression. Comput Oper Res. 1998;25(4):251–63. https://doi.org/10.1016/S0305-0548(97)00074-9.
Hush, DR. Classification with neural networks: a performance analysis; 1989, p. 277–80. https://doi.org/10.1109/ICSYSE.1989.48672
Acknowledgements
The authors thank Dr Jean-Marie Guigonis (Bernard Rossi facility) for the LC-MS analyses. The authors thank Pr Fanny Burel-Vandenbos and Valerie Rigau for providing the samples of the Brain dataset.
Funding
This work has been supported by the French government, through UCA-JEDI Investment in the Future project managed by the National Research Agency (ANR) with the reference number ANR-15-IDEX-01. Equipment for this study was purchased through grants from the Recherche en matières de Sûreté Nucléaire et Radioprotection program from the French National Research Agency and the Conseil Départemental 06.
Author information
Authors and Affiliations
Contributions
DC,CG and MB implemented and performed the experiment and analyzed the obtained results. TP and OH managed resources and funding acquisition. DC,CG,TP,OH and MB wrote the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
LUNG dataset The LUNG dataset was published by Mathe et al. [38] and available at MetaboLights database (study identifer MTBLS28). BREAST dataset The BREAST dataset was kindly provided by Dr. Jan Budczies and can be found in the supplementary material of Badczies et al. [39]. BRAIN dataset The samples from the BRAIN dataset were retrospectively collected from two declared biobanks. The authors confirm that all methods were carried out in accordance with relevant guidelines and regulations. The authors confirm that informed consent was obtained from all subjects. The authors confirm that this retrospective study has been approved by the institutional ethics committees of the University Hospital of Nice and the University Hospital of Montpellier.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Chardin, D., Gille, C., Pourcher, T. et al. Learning a confidence score and the latent space of a new supervised autoencoder for diagnosis and prognosis in clinical metabolomic studies. BMC Bioinformatics 23, 361 (2022). https://doi.org/10.1186/s12859-022-04900-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s12859-022-04900-x