 Research
 Open Access
 Published:
Evaluation of treebased statistical learning methods for constructing genetic risk scores
BMC Bioinformatics volume 23, Article number: 97 (2022)
Abstract
Background
Genetic risk scores (GRS) summarize genetic features such as single nucleotide polymorphisms (SNPs) in a single statistic with respect to a given trait. So far, GRS are typically built using generalized linear models or regularized extensions. However, these linear methods are usually not able to incorporate genegene interactions or nonlinear SNPresponse relationships. Treebased statistical learning methods such as random forests and logic regression may be an alternative to such regularizedregressionbased methods and are investigated in this article. Moreover, we consider modifications of random forests and logic regression for the construction of GRS.
Results
In an extensive simulation study and an application to a real data set from a German cohort study, we show that both treebased approaches can outperform elastic net when constructing GRS for binary traits. Especially a modification of logic regression called logic bagging could induce comparatively high predictive power as measured by the area under the curve and the statistical power. Even when considering no epistatic interaction effects but only marginal genetic effects, the regularized regression method lead in most cases to inferior results.
Conclusions
When constructing GRS, we recommend taking random forests and logic bagging into account, in particular, if it can be assumed that possibly unknown epistasis between SNPs is present. To develop the best possible prediction models, extensive joint hyperparameter optimizations should be conducted.
Background
The development of complex diseases depends on many factors such as genetic mutations, the lifestyle, or environmental factors. Investigating the effects of genetic variants across the human genome in genomewide association studies (GWAS) has already revealed relevant risk basepair alterations [1]. Single nucleotide polymorphisms (SNPs) may have only a very small effect on the investigated disease. However, when considered jointly, SNPs might be highly relevant [2, 3]. This behavior can be due to many independent SNPs exhibiting minor individual effects, or it can be caused by interactions of genetic variants, i.e., epistasis.
In consequence, summarizing relevant genetic effects in an individual while sufficiently predicting the risk for a certain disease, potentially jointly with nongenetic covariables, would be highly desirable. This would, on the one hand, allow to uncover underlying mechanisms related to this specific disease. On the other hand, accurately predicting the risk of disease for an individual could have a high impact on personalized medicine due to potentially being able to reduce the personal risk by taking specialized preventive measures if an individual has a high genetic risk for a certain disease [4, 5].
One promising approach for the assessment of an individual’s risk is the development of genetic risk scores (GRS). For the construction of GRS, one typically selects a subset of relevant SNPs from a biological pathway or a gene and calculates a weighted sum of the selected genetic variants.
Genomewide approaches with a selection of genetic variants from across the whole genome resulting from prior knowledge are also possible for building GRS [6, 7]. However, such selections typically depend on largescale association studies in which single SNPs were tested individually with regard to the phenotype. Thus, interacting variants which do not exhibit substantial marginal effects might be left out although SNP level interactions might contribute to disease risk [8, 9]. In this context, an alternative to conventional GWAS for identifying diseaserelated SNPs might be genomewide association interaction studies (GWAIS) [9].
The standard procedure for the computation of the GRS is the usage of external weights [10, 11], ideally determined from independent association studies such as GWAS or GWAIS. However, there might be no appropriate association study for the regarded outcome or population available such that suitable weights have to be gathered in a different way.
Internal GRS weights can be estimated by regarding the problem of constructing GRS as a supervised statistical learning problem, where the response would be the disease status or a quantitative biological variable such as the glucose level. In this case, the predictors are genetic variants of the specific pathway or gene, where SNPs are usually coded by the number of minor alleles for this individual. The estimation of proper weights or fitted models which generalize well, i.e., which represent the whole population reasonably well and not only the available sample, requires the partitioning of the whole data set into training and test data sets. Dudbridge [3] and Hüls et al. [11] found in their studies that a random close to onehalf split generalizes well. Sufficient samples are necessary in the test data set for evaluating the association of the GRS with the response which especially holds true for geneenvironment interaction (GxE) studies in which more parameters are to be estimated. A GxE interaction is present if, for different genotypes, different disease susceptibilities to an environmental factor are underlying, e.g., if an individual has a high genetic risk for a certain disease which is enabled by an environmental factor [12].
So far mainly linear methods such as generalized linear models (GLM) or regularization methods based on GLMs, such as the lasso [13] or one of its generalizations, the elastic net [14], have been used in the construction of GRS [11, 15, 16]. The elastic net offers the advantage of properly handling highly correlated predictors, e.g., SNPs in linkage disequilibrium (LD), by employing an \(L_2\) regularization while performing a variable selection due to the \(L_1\) regularization. Nonetheless, these regularized linear regression methods cannot directly take interactions between predictors into account (unless specific interaction terms were specified prior to applying them) and the assumption of an additive relationship between the response and the input variables has to be fulfilled. Therefore, the usage of algorithms which are able to develop more general models and which in fact can find and take interesting interactions into account might be preferable.
The treebased statistical learning method random forests [17] is wellknown and widely used among a variety of use cases [e.g., [18,19,20]]. It builds several individual classification or regression trees (CART) [21], which are fitted by a nonlinear recursive partitioning algorithm, and combines them to one strong ensemble. For a low to moderate amount of SNPs (\(< 100\)), it has been shown that the classic random forests algorithm is able to properly uncover SNP interactions even when the corresponding marginal effects are negligible [22].
Another treebased nonlinear statistical learning procedure is logic regression [23] which mainly considers binary predictors. It searches for Boolean expressions of the input variables and combines multiple expressions in a GLM and already has been used in applications to SNP data [24,25,26]. Both treebased methods are theoretically able to cover each possible prediction scenario for categorical input data. However, their model fitting techniques are highly different.
To the best of our knowledge, it has barely been investigated yet whether the aforementioned statistical learning algorithms can be used as alternative procedures to conventional GRS construction approaches. For random forests, some publications suggest that the ensemble method is able to outperform conventional linear methods such as logistic regression, odds ratio scores or the lasso [27, 28]. However, more recent studies which considered genomewide risk scores, i.e., GRS constructed using SNPs from all over the genome and not just single genes or pathways, were not able to verify that random forests should be used over linear approaches [29, 30]. In the context of disease risk prediction, e.g., Yoo et al. [31] regarded random forests, logic regression, and logistic regression without penalization in one simple genegene interaction simulation study and additionally in a real data application. In their analyses, the treebased algorithms could induce higher predictive performances than logistic regression. Nonetheless, multifaceted analyses taking different realistic data scenarios into account are necessary in order to draw meaningful conclusions about the appropriateness of the treebased methods for the construction of GRS.
The classic random forests and logic regression algorithms have some shortcomings. In particular, random forests can severely overfit the data [32] and logic regression can lead to highly variant models [24]. Thus, we additionally considered modifications of the classic algorithms to overcome these drawbacks.
In this article, we, therefore, evaluate random forests, logic regression, and extensions of these methods in an extensive simulation study and an application to a real data set from a German cohort study for the construction of GRS and compare the results to the elastic net.
Methods
Construction of genetic risk scores
Let \({\mathcal {D}}_{\mathrm {train}} = \lbrace ({\varvec{x}}_i, y_i) \rbrace _{i=1}^N\) be a training data set with N observations and binary outcomes \(y_i \in \lbrace 0, 1 \rbrace\). Further assume that each input vector \({\varvec{x}}_i\) is a collection of p biallelic SNPs, i.e., taking values in the pdimensional space \(\lbrace 0,1,2 \rbrace ^p\), where 0 codes the homozygous reference, 1 the heterozygous variant, and 2 the homozygous variant. Then the problem of constructing a GRS model consists of fitting a proper function
The target space is equal to the probability scale [0, 1], since \({\mathrm {GRS}}({\varvec{x}})\) should be an estimate of \({\mathbb {P}}(Y = 1 \mid {\varvec{X}} = {\varvec{x}})\), i.e., the probability of being a case given some SNPs \({\varvec{x}}\). This fitting procedure is conducted on a designated training data set. Independent observations from a test data set \({\mathcal {D}}_{\mathrm {test}}\) are used to evaluate the GRS, i.e., \({\mathrm {GRS}}({\varvec{x}})\) for \(({\varvec{x}}, \cdot ) \in {\mathcal {D}}_{\mathrm {test}}\).
An overview of the workflow for fitting and evaluating GRS models using the statistical learning approach is given in Fig. 1.
Random forests
In random forests, multiple classification or regression trees (CART) [21] with injected randomness are built to form one strong ensemble. From a graphtheoretical point of view, decision trees are usually binary trees in which each inner knot represents a split based on a predictor and each leaf (terminal node) describes a prediction scenario. Figure 2a illustrates an exemplary classification tree with four disjoint prediction scenarios. New predictions start at the root node and follow the respective edge until a leaf is reached.
Decision trees are induced by a recursive greedy splitting algorithm which searches at each inner node for the best possible split with respect to an impurity measure. The impurity measure is a quantifier for the homogeneity of respective nodes. For binary classification trees, the Gini impurity
for empirical probabilities \(P(Y = c \mid {\varvec{X}} \in t)\) that the response Y is equal to class c given that the input vector \({\varvec{X}}\) falls into the regarded node t is usually chosen.
The tree induction procedure can be locally terminated by stopping criteria. When a node should not be split, it is declared as a leaf and has to receive a prediction value. For classification trees, this is usually the class with the highest empirical probability in the regarded branch.
However, single decision trees suffer from the instability problem which states that a small noiselike modification of the training data set may lead to a disproportional modification of the fitted tree. This issue is mainly caused by the greedy fashion of choosing splits [33].
Random forests tries to address this issue. The algorithm employs bagging [34] which draws a bootstrap sample of the available data for each individual tree as its training data set. The tree fitting procedure is further randomized by adjusting the splitting algorithm to choose \(mtry \le p\) predictors from the total set of input variables at every inner node which qualify for the best split. \(mtry\) is a hyperparameter usually chosen as \(\sqrt{p}\) or p/3 which should be properly tuned in certain applications. Based on these randomizations, the resulting model averages the individual trees, i.e., for classification trees, the class which is classified most often will be chosen as the prediction.
Random forests for constructing genetic risk scores
If one is aiming at constructing GRS for binary traits, one has to keep in mind that probability estimates for showing the regarded feature are needed instead of class estimates. Random forests based on classification trees can be used for probability estimation by averaging the number of trees which voted for class 1 [35]. However, if we, e.g., assume that the true risk for being a case would be equal to \(80\%\) and that all classification trees properly recognize this fact and, therefore, predict class 1 for this particular setting, the forest risk estimate would be given by \(100\%\). Thus, for this reason, we consider probability estimation trees [36] which hold risk estimates in their leaves in contrast to classifications. These estimates are usually chosen as the empirical branch probabilities from which classification trees also draw their estimates. Random forests based on probability estimation trees average the probability estimates of the individual trees similar to regression trees.
If SNP variables coded as 0, 1, or 2 are interpreted as quantitative variables, decision trees and random forests are able to split with respect to \((\lbrace 0 \rbrace , \lbrace 1, 2 \rbrace )\) or \((\lbrace 0, 1 \rbrace , \lbrace 2 \rbrace )\), thus, considering both dominant and recessive modes of inheritance. Therefore, SNPs are directly used as input variables when employing random forests.
Random forests VIM
One issue that arose when fitting the first GRS models with random forests in our initial experiments was a substantial overfitting which could be observed by comparing the test and training data errors. Therefore, performing an appropriate variable selection prior to fitting the final random forests models might reduce the overfitting and lead to better results for noiseintensive data. Kursa and Rudnicki [37] proposed an iterative variable selection approach which relies on variable importance measures (VIM) and which they called Boruta. The permutation VIM can be calculated using the outofbag observations for each tree, thus, avoiding an overfitting of the VIM itself. In each iteration, the Boruta approach adds for each predictor variable a shadow variable with the same values but randomly permutes them to destroy a potential predictorresponse relationship for this variable. Next, a random forest on this extended set of input variables is fitted and the evaluated VIMs for these shadow variables are used to approximate the distribution of VIMs for noninfluential input variables. The computed VIMs of the original variables are then compared to the VIMs of the shadow variables in statistical tests for importance. In particular, the maximum observed importance of all shadow variables is used to decide whether an original variable is temporarily classified as important. More specifically, if a variable yields an importance higher than the maximum observed importance among all shadow variables, it will be temporarily marked as important. Several iterations of creating shadow variables, fitting random forests, and computing VIMs are used to perform binomial tests, which regard how often the variable was temporarily marked as important, testing the alternative of greater or smaller VIM realizations, i.e., important or unimportant variables. More precisely, these binomial tests are based on the null hypothesis that the probability of the regarded input variable yielding a higher VIM than the maximum VIM of all shadow variables is equal to 0.5. The significance threshold of the binomial tests is set to 1%, which is also the recommended threshold by the authors of the Boruta approach. Compared to other randomforestbased variable selection methods such as the Vita algorithm proposed by Janitza et al. [38] which relies on negative VIM values, the Boruta approach does not require a vast amount of (noninfluential) input variables.
As an alternative procedure, we also tried the variable selection method by Altmann et al. [39], which relies on random permutations of the response variable. However, in our experiments, the Boruta approach yielded more stable results in general. In particular, even when considering different significance thresholds for the approach by Altmann et al. [39], the Boruta procedure still could induce more stable variable selections, i.e., leading to variable selections that did not severely differ between independent replicates. This observation is in line with the analyses by Degenhardt et al. [40] who provide an indepth comparison of various random forests variable selection methods.
Hence, we fitted ordinary random forests with probability estimation trees and random forests based on the Boruta variable selection which we call random forests VIM in the following. For random forests, we used the R package ranger [41]. For random forests VIM, the R package Boruta [37], that also relies on the ranger package, was used.
Logic regression
Logic regression [23] is a treebased statistical learning algorithm which is specifically tailored to binary input variables. It searches for ideal Boolean expressions of those and works with binary tree representations of Boolean expressions, logic trees. Logic trees hold the Boolean operators \(\wedge\) (AND) or \(\vee\) (OR) in their inner nodes and contain predictor variables or their negations (indicated through \(^c\)) in their terminal nodes. Figure 2b depicts an exemplary logic tree which is equivalent to the exemplary classification tree from Fig. 2a, i.e., both trees perform the same predictions for each realization of the three input variables. The interpretation as a Boolean expression is obtained recursively by combining expressions in a bottomup fashion, yielding \((X_1^c \wedge X_2) \vee (X_1 \wedge X_3^c)\) for the logic tree from Fig. 2b.
Logic trees themselves can only be used for binary classification tasks, since they represent logic expressions so that their output is also either 0 or 1. To generalize their usage for, e.g., risk prediction, Ruczinski et al. [23] proposed using logic trees \(L_1, \ldots , L_M\) as predictors in a GLM
considering an appropriate link function g such as the \({\mathrm {logit}}\) function \({\mathrm{logit}}(p) = \log (p/(1p))\) for a binary response.
The total model fitting procedure consists of finding the most appropriate logic tree(s). In practice, for each model, a set of neighbor states is defined by simple adjustments of the current model. The moves used in logic regression consist of exchanging variables and operators, adding or removing branches, splitting or removing variables, and adding or removing trees. This set of moves ensures that from every state, every other possible state can be reached in a final number of steps. For more details, see [23].
Based upon this methodology, two model search algorithms are used in practice:

a greedy search which evaluates each neighbor of a given state and moves to the best one

simulated annealing [42], a stochastic search algorithm which only considers one random neighbor per iteration and can also move to worse states to prevent being stuck in a local minimum.
Model ranking is performed using a score function which is chosen to be the deviance for the logistic model. The model which yields the best score among all models visited in the search is chosen as the resulting model. Irrespective of using the greedy approach or simulated annealing, one should configure the model size hyperparameters, i.e., the total number of trees and the total number of leaves, to obtain the best fit on the entire population. For fitting conventional logic regression models, we used the R package LogicReg [43] and used simulated annealing as the search procedure.
Logic regression for constructing genetic risk scores
SNP variables coded as 0, 1, or 2 can be biologically meaningful divided into two binary variables, in \({\mathrm {SNP}}_D = {\mathbbm {1}}({\mathrm {SNP}} \ne 0)\), coding for a dominant effect, and in \({\mathrm {SNP}}_R = {\mathbbm {1}}({\mathrm {SNP}} = 2)\), coding for a recessive effect. With these two binary variables, interactions can be properly expressed. For example, consider a scenario where two SNPs influence the disease risk in such a way that the risk is significantly increased if and only if for both SNPs their respective minor allele occurs at least once. With Boolean logic, this can be expressed as \({\mathrm {SNP}}_{1,D} \wedge {\mathrm {SNP}}_{2,D}\). It might also be possible that two riskincreasing SNPs with a dominant mode of inheritance can only elevate the disease risk once, i.e., if both statuses occur, the risk is not increased beyond the first elevation. This scenario can also be expressed with Boolean logic as \({\mathrm {SNP}}_{1,D} \vee {\mathrm {SNP}}_{2,D}\). Furthermore, SNPs in high linkage disequilibrium (LD) that are, therefore, highly correlated can also be properly addressed with the logical OR. One LD block might then be expressed as a chain of ORconcatenated SNPs, a disjunction. Thus, for the construction of GRS with logic regression, each SNP is divided into two binary variables prior to applying the procedure.
Logic bagging
As an alternative to an exhaustive search with simulated annealing, we also considered applying bagging [34] to logic regression models fitted with a greedy search. We call this approach logic bagging. In contrast to conventional logic regression, logic bagging fits ensembles of individual logic regression models and, similar to random forests, predictions are made using the average of the predictions of the individual logic regression models. This approach is still computationally expensive when using an adequate amount of bagging iterations (e.g., 500) but reduces the variance and does not require the tuning of a cooling schedule. Logic bagging is implemented in the R package logicFS [44]. For fitting logic bagging models, the greedy search is employed mainly due to computational reasons. In particular, in Additional file 1: Fig. S1, the model fitting times are depicted. For example, for fitting and evaluating a single logic bagging model consisting of 500 logic regression models fitted via simulated annealing, it would take about \(500 \cdot 28.82s \approx 4h\) using the mean model fitting and evaluation time of 28.82s for logic regression.
Elastic net
The elastic net [14] is a regularized linear regression model which combines

the lasso (least absolute shrinkage and selection operator) [13], i.e., \(L_1\) regularized regression that reduces the estimate of the regression coefficients of noninfluential predictors to zero, therefore, excluding noninformative input variables,

and ridge regression [45], i.e., \(L_2\) regularized regression for properly handling highly correlated predictors by assigning similar weights to such predictors.
Elastic net, hence, uses a penalty term given by
for the regression coefficients \(\varvec{\beta } = \begin{pmatrix} \beta _1&\ldots&\beta _p \end{pmatrix}^T\) in the fitting procedure solving the optimization problem
for the loglikelihood function \(\ell\). In this article, binary outcomes are considered. Thus, the logistic regression approach for elastic net was employed.
Here, \(\lambda \ge 0\) determines the strength of the regularization, i.e., for larger values of \(\lambda\), the penalty \(\lambda R_\alpha (\varvec{\beta })\) increases, thus, favoring coefficient vectors with smaller norms, i.e., more loosely fitting models. The parameter \(\alpha \in [0,1]\) configures the balance between the lasso and ridge regression, i.e., for \(\alpha = 0\), one would perform ordinary ridge regression and for \(\alpha = 1\), one would apply the lasso. Therefore, these two hyperparameters have to be tuned properly.
In practice, the model coefficients are estimated by employing coordinate descent as optimization algorithm to solve the minimization problem (1) and taking advantage of the fact that similar values of \(\lambda\) lead to similar model coefficients for a fast fitting of different \(\lambda\) settings [46]. We used the R package glmnet [46] with crossvalidation for fitting elastic net models.
The common procedure when constructing GRS with regularized regression procedures such as elastic net is to use the \(\lbrace 0,1,2 \rbrace\) coding for each SNP in the model [11, 16]. We, therefore, follow in our comparison this standard procedure and use the \(\lbrace 0,1,2 \rbrace\) coding in the elastic net.
If interaction effects between SNPs should be included in the elastic net model, they have to be explicitly specified prior to fitting the model. However, in practice, it is usually unknown which loci might interact. Including all possible interactions between SNPs becomes rapidly infeasible, as for a moderate amount of SNPs, the number of possible interaction terms might already be too high. For example, for 50 SNPs, there exist more than \(10^{15}\) interaction terms. The standard procedure for constructing GRS with linear methods such as the elastic net is to only consider marginal genetic effects [16]. Thus, we follow in our evaluations this common procedure and do not include interactions in the elastic net models.
Simulation studies
The treebased statistical learning methods random forests, random forests VIM, logic regression, and logic bagging were evaluated and compared to the elastic net in a simulation study considering three scenarios with several different settings. All SNPs were drawn independently resembling LDbased pruned or clumped SNPs. All simulations and analyses were performed with R version 4.0.3 [47]. Data sets for all simulation scenarios were generated using the R function simulateSNPglm from the scrime package [48].
General workflow
The general workflow for generating the data sets for the simulation study is given as follows for each of the simulation settings, which are described in detail afterwards.

1.
Choose the fixed data parameters, i.e., the odds ratios, number of SNPs, sample size and simulation design.

2.
For each SNP, draw a random minor allele frequency (MAF).

3.
Randomly generate the genotypes based on the MAFs.

4.
If continuous covariables are to be included, randomly generate the data for these variables.

5.
Randomly generate the outcome according to the linear predictor.

6.
Evaluate the fraction of cases in the generated outcome and tune the prevalence such that this fraction becomes approximately balanced, i.e., yielding \(\sim 50\%\) cases. This involves going back to step 5.

7.
Create 100 independent data sets for a certain setting using the steps 2–5 for each repetition.
Simulation setups
Marginal genetic effects
In a first step, we focused on main effects, which represents the ideal case for the elastic net, since no interactions are considered here and the individual effects behave additively with each other. Similar to Hüls et al. [49], we considered six SNPs influencing the value of the outcome, where we simulated a dominant effect for each of these SNPs. Thus, we first considered data sets generated from a logistic regression model
In order to draw conclusions for different realistic scenarios, we varied three parameters:

the effect size, i.e., the odds ratio, of each influential SNP which can be configured by specifying \(\exp (\beta _i)\) [50],

the intensity of statistical noise which we adjusted by adding noninfluential SNPs to each data set,

and the sample size of each data set.
To achieve nearly casecontrol studylike designs, we configured the prevalence, i.e., \(\left( 1+\exp (\beta _0)\right) ^{1}\) [50], to result in nearly balanced data sets for each regarded odds ratio. The MAF was drawn randomly for each SNP and for each data set from the interval [0.15, 0.45] similar to Pan et al. [51]. For each scenario, we generated 100 independent data sets, i.e., we performed 100 replications. Table 1 lists the regarded settings for the aforementioned simulation parameters.
Dominant interactions of SNPs
In a second simulation scenario, we additionally considered a genegene interaction, i.e., an interaction between SNPs. More specifically, we here always considered three SNPs with low main effects, i.e., odds ratios of 1.2 and a dominant mode of inheritance, since we focused on marginal effects in the first scenario. Additionally, we included an interaction term between two SNPs whose odds ratio was varied. Similar to the first scenario, we also varied the amount of statistical noise, i.e., the number of SNPs for which no effect on the outcome is intended. Furthermore, we considered three sub designs that determine which SNPs interact. The data was generated following models such as
The indices \((j, k) \in \lbrace (1,2), (1,4), (4,5) \rbrace\) determine whether both interacting SNPs also do have marginal effects, only one of them exhibits a main effect, or if they only are influential when considered jointly. The prevalence was again configured by \(\beta _0\) to approximately achieve casecontrolbalanced study designs. The MAF was randomly chosen in the interval [0.15, 0.45] and the sample size was fixed to 2000 observations per data set, since we only considered weak marginal effects. 100 independent data sets for each setting were analyzed using a cyclic scheme such as in the first simulation scenario. The study parameters for the second simulation scenario are summarized in Table 2.
Geneenvironment interactions
In the final simulation scenario, we added two correlated continuous variables to the true underlying model from which one forms a GxE interaction with a SNP. One of these two variables exhibits a marginal effect on the outcome, while the second variable only influences the outcome if a certain risk allele occurs at least once. The data for this scenario was generated considering the model
Similar to the genegene interaction simulation scenario, the effects for the first three SNPs were fixed to odds ratio of 1.2, 1.5, and 1.8, respectively. The interaction between \({\mathrm {SNP}}_1\) and \({\mathrm {SNP}}_4\) received a fixed odds ratio of 1.8, since in this analysis, the focus lies on the GxE interaction. The index \(j \in \lbrace 2,5 \rbrace\) determines whether the SNP in the GxE interaction also exhibits a moderate marginal effect or if this SNP only influences the outcome in interaction with the continuous variable \(E_2\). The odds ratios of the terms involving the continuous variables \(E_1\) or \(E_2\) were specified per IQR (interquartile range) of the respective environmental variable as it is regularly done when performing analyses of GxE interactions [11]. For the continuous variable \(E_1\), the (marginal) odds ratio was fixed to 1.2 per IQR. The odds ratio of the GxE interaction between \({\mathrm {SNP}}_j\) and \(E_2\) was varied between 1.2 and 2.4. The continuous variables were generated following a multivariate normal distribution, i.e.,
In particular, the mean \(\mu\) was set to 20, the variance \(\sigma ^2\) was chosen as 10 and the correlation \(\rho\) between these two variables was chosen as either 0.5 or 0.9, resembling moderately and highly correlated variables, respectively. The prevalence was again configured by \(\beta _0\) to approximately achieve casecontrolbalanced study designs. The MAF was randomly chosen in the interval [0.15, 0.45] and the sample size was fixed to 2000 observations per data set as in the previous simulation scenario. The number of additional noise SNPs was fixed to 45. 100 independent data sets for each setting were analyzed. The study parameters for the third simulation scenario are summarized in Table 3. In GxE interaction studies, GRS are usually constructed only using the available genetic data [11]. Thus, we constructed the GRS without utilizing the environmental variables.
Analysis of association and predictive strength
To evaluate and compare the different statistical learning methods in their ability to construct GRS, a cyclic trainingvalidationtest data set scheme was considered. In the ith repetition of this cyclic scheme, the ith data set \({\mathcal {D}}_i\), \(i \in \lbrace 1, \ldots , 100 \rbrace\), was used to train the GRS with the different statistical learning methods. For the evaluation of the performance of these methods, the succeeding data set \({\mathcal {D}}_{i+1}\) if \(i \ne 100\) and \({\mathcal {D}}_1\) otherwise was chosen to be the independent test data set. For tuning the hyperparameters (see “Section Hyperparameter optimization”), we chose the preceding data set, i.e., \({\mathcal {D}}_{i1}\) if \(i \ne 1\) and \({\mathcal {D}}_{100}\) otherwise as validation data.
Since all data sets were generated independently, the cyclic scheme is equivalent to a conventional trainingvalidationtest data set approach in which each of the 100 data sets is once used as training set, once as test set, and once as validation set in a cyclic manner. Due to the high computational costs when considering many different parameter configurations, hyperparameter tuning was performed by averaging the performances over the first 10 validation iterations for each simulation setting and each parameter setting. The setting which yielded the highest validation AUC across the average over these 10 repetitions was chosen as the fixed setting for the particular simulation setting.
The standard approach for testing the association considers the GRS as a predictor in a conventional regression model [2]. For binary outcomes, the logistic regression model is fitted on the test data. The logistic regression model maps the linear predictor with the logistic function from \((\infty , +\infty )\) to (0, 1). Thus, the GRS (probability estimates) are transformed to the scale of the linear predictor by applying the inverse of the logistic function, the \({\mathrm {logit}}\) function. In summary, the univariate association model
is constructed using
for raw risk predictions of the fitted GRS model \({\mathrm {GRS}}_\mathrm {raw}\).
For statistically assessing this association, we conducted Wald tests testing the alternative that the GRS is associated with the response. Based on these test results, we estimated the statistical power and the type I error rate for analyzing and comparing the ability of properly recognizing signals in the genetic data by the GRS construction procedures. The statistical power, which is given by the probability that the GRS is correctly recognized as influential on the response, can be estimated by the fraction of logistic models with statistically significant predictors under all cases which rely on theoretically influential genetic data. The type I error rate, i.e., the false positive rate, can be estimated by the fraction of significantly recognized GRS under all cases in which the response and the predictors are actually independent.
To compare the predictive strength of GRS, which is probably most relevant, we calculated the area under the curve (AUC) with respect to the receiver operating characteristic (ROC). This metric offers two main advantages over classification measures such as the accuracy, sensitivity, or specificity. First, it does not depend on the classification threshold which perhaps should be tuned. Second, the AUC can handle imbalanced data sets due to simultaneously regarding sensitivity and specificity. Moreover, the AUC has an intuitive interpretation as the probability that a random observation from the entire population of cases yields a higher risk estimate than a randomly chosen control from the population [52].
Additionally, we evaluated the classical classification metrics accuracy, sensitivity, and specificity. In particular, we performed hard classifications on the resulting logistic regression model containing the GRS using a classification threshold of 0.5, i.e., classifying an observation as a case if it is predicted that the probability of being a case is higher than the probability of being a control and classifying an observation as a control otherwise. Using these classifications, the overall accuracy, sensitivity, and specificity as defined, e.g., in Alberg et al. [53] were evaluated. The accuracy was not explicitly adjusted for the prevalence, since we generated approximately casecontrolbalanced data sets in the simulation study, thus, yielding a prevalence of 50%. However, the main purpose of GRS does not lie in hard classifying observations as cases or controls. Instead, GRS are used for estimating individual risks, e.g., in precision medicine or for uncovering biological mechanisms involved in the development of diseases. Therefore, a metric such as the AUC which simultaneously considers different sensitivities and specificities seems to be preferable in the evaluation of the performance of GRS.
Hyperparameter optimization
Certain statistical learning procedures require the optimization of hyperparameters using independent validation data sets. This also holds true for the algorithms considered in this article. Table 4 lists the regarded hyperparameter configurations, where each possible combination of these parameters has to be considered in the parameter tuning. A description of each of these parameters is given in Additional file 1: Section 2. For random forests, we fixed the number of total trees grown to 2000, which is a sufficiently large number of trees in our applications, since in preliminary experiments, we could observe that the validation AUC converged using smaller amounts of trees. Analogously, we fixed the number of bagging iterations for logic bagging to 500. The cooling schedule in logic regression was configured manually by observing the cooling behavior for different settings and choosing a start temperature and end temperature such that around 90% of the proposed models were accepted at the beginning of the algorithm and close to no models were accepted when approaching the end temperature. The amount of simulated annealing iterations was fixed to 500000. The regularization parameter \(\lambda\) for the elastic net was automatically chosen by employing crossvalidation in the respective fitting processes and selecting the value which minimizes the loss.
For each considered statistical learning method, a more detailed workflow for tuning and training the respective models is depicted in Additional file 1: Section 3.
Results of the simulation studies
Marginal genetic effects
Figure 3 summarizes the AUC for each of the 27 regarded settings in the main effects simulation scenario. In Additional file 1: Fig. S2, corresponding asymptotic 95% confidence intervals are depicted. Most notably, logic bagging leads in almost every scenario to the highest AUC. For strong effects and large data sets, ordinary logic regression induces similar or even better results which are comparable to the true underlying model. Especially for weak effects, ordinary random forests yields comparably high values for the AUC. Unsurprisingly, random forests with a prior variable selection is more effective in relation to the other procedures when considering a higher amount of statistical noise. For less noisy data, random forests VIM cannot compete with the other treebased methods and shows high variations. The elastic net yields inferior results for large data sets and large effect sizes and also has difficulties detecting a signal for the more challenging scenarios, i.e., for small odds ratios and low observation counts.
The analyses of power resemble the results of the AUC comparison and are depicted in Additional file 1: Fig. S3. The type I error rates for the treebased methods seem to randomly scatter around the prespecified significance level of 5%. However, the elastic net induces type I error rates of around two percent and is, therefore, quite conservative. The corresponding type I error rates are shown in Additional file 1: Fig. S4.
In Additional file 1: Figs. S5–S7, the results for the accuracy, sensitivity, and specificity are depicted. The accuracies resemble the results of the AUC evaluation, while the sensitivities and specificities do not show a clear pattern between the evaluated methods. These figures also show that, for increasing odds ratios, the specificities increase while the sensitivities decrease.
We also evaluated the GRS on the training data itself to compare the degrees of overfitting. Here, ordinary random forests leads to the severest overfitting. For data with high statistical noise and small effect sizes, its AUC almost reaches 100% compared to the true AUC of around 56%. The other treebased algorithms also induce higher training AUCs than the true model, but not larger than random forests. In particular, a prior variable selection can indeed reduce the intensity of overfitting. The elastic net yields in most cases the lowest values for the AUC closely following the AUCs of the true model. Taking the test data analyses into account, this indicates a mixture of underfitting and slight overfitting of the elastic net. The training data results can be found in Additional file 1: Fig. S8.
Dominant interaction effects of SNPs
For the analysis of the scenarios with influential interaction terms, the performances of the statistical learning procedures measured by the AUC are shown in Fig. 4. Additionally, asymptotic 95% confidence intervals can be found in Additional file 1: Fig. S9. Similar to the main effects scenarios, logic bagging induces in each scenario the highest values of the AUC. Also as in the other settings, random forests VIM does not gravely suffer from noisy data compared to standard random forests, but cannot severely outperform its ordinary counterpart. Random forests itself seems to be the secondbest performing method with an almost steady but close distance to logic bagging. Interactions of variables without marginal effects seem to be less of an issue to conventional logic regression, since for Design 2.3 and larger interaction effect sizes, logic regression achieves comparable AUCs to random forests. For weak interaction effects, the elastic net can yield comparative results to random forests and the logic regression. Nonetheless, increasing the interaction effect also increases the discrepancy between the treebased approaches and the elastic net.
The results of the corresponding power and type I error analyses can be found in Additional file 1: Figs. S10 and S11. As in the previous simulation scenario, the comparison of the estimates of the statistical power resembles the corresponding analyses of the AUC. Again, the type I error rates for the treebased methods seem to randomly scatter around 5%, whereas the elastic net leads to substantially lower error rates.
The results for the accuracy, sensitivity, and specificity can be found in Additional file 1: Figs. S12–S14. Similar to the marginal effects simulation scenario, the comparisons of the mean accuracy resemble the results of the AUC evaluation. The other two metrics sensitivity and specificity do not yield clear patterns between the considered procedures.
Evaluations of the GRS on the training data reveal again that conventional random forests seems to induce the severest overfitting. The results of these training data set applications are summarized in Additional file 1: Fig. S15.
Geneenvironment interactions
Figure 5 depicts the predictive performances of the statistical learning procedures for the 20 settings in the GxE interaction simulation scenario. Corresponding asymptotic 95% confidence intervals are shown in Additional file 1: Fig. S16. In contrast to the previous scenario, a true unique GRS model does not exist, since the GRS is based only on the genetic data while the true model of this scenario also consists of environmental covariables. Similar to the genegene interaction scenario, logic bagging leads in each setting to the highest AUCs. Throughout all settings in this simulation scenario, logic regression seems to be the second best performing method yielding AUCs closely below the AUCs of logic bagging. Random forests and random forests VIM induce very similar results such that there is no clear pattern between these two methods. For weak GxE interaction effects, the elastic net induces comparably poor results. However, for increasing GxE interaction effects, the discrepancy between random forests and elastic net decreases such that, for an odds ratio of 2.4, the elastic net yields slightly higher AUCs than random forests which are, however, still below the AUCs of logic bagging.
The correlation \(\rho\) of the two continuous variables does not seem to affect the GRS performance in this simulation scenario. Nonetheless, the overall performance in Design 3.1 is higher than the performance in Design 3.2. This phenomenon can be explained by the absence of a marginal effect of the GxE interacting SNP in Design 3.2 complicating the identification of this SNP.
For this simulation scenario, the statistical power for all considered methods and simulation settings was equal to 100%. Similar to the previous scenarios, the elastic net seems to be more conservative as it induces lower type I error rates than the treebased methods. The estimated type I error rates can be found in Additional file 1: Table S1.
In Additional file 1: Fig. S17–S19, the results for the accuracy, sensitivity, and specificity are depicted. Similar to the power analyses, the mean accuracies of the considered methods are almost identical in each simulation setting. However, for weak GxE interaction effects, the elastic net seems to induce the lowest mean accuracies. The results for the other two metrics, the sensitivity and the specificity, are also very similar.
Training data evaluations reveal again that conventional random forests tends to induce the severest overfitting. The training data results are depicted in Additional file 1: Fig. S20.
Comparison considering binary SNP codings
Additionally to considering the standard way of specifying the input variables for the different methods, we also evaluated the GRS construction approaches using the binary \(\lbrace 0,1 \rbrace\) SNP coding for each method and not exclusively for logic regression and logic bagging. The detailed results for the \(\lbrace 0,1 \rbrace\) SNP coding and the respective simulation scenarios are depicted in Additional file 1: Figs. S21–S23.
In comparison to using the \(\lbrace 0,1,2 \rbrace\) coding, the performance of random forests and random forests VIM decreases. This is not very surprising, since, as pointed out in the methodological description, decision trees and random forests consider the dominant and recessive modes of inheritance when using the \(\lbrace 0,1,2 \rbrace\) coding. Thus, using the \(\lbrace 0,1 \rbrace\) coding doubles the number of input variables without supplying more information to random forests. The increase in the number of input variables complicates identifying the ideal splits when using typical settings for the hyperparameter \(mtry\).
For the elastic net, the performance increases when employing the \(\lbrace 0,1 \rbrace\) coding instead of the conventional \(\lbrace 0,1,2 \rbrace\) coding such that, in the marginal effects simulation scenario and in the GxE interaction scenario, the elastic net yields similar results as logic bagging when considering settings with stronger genetic effects. Nonetheless, in the genegene interaction simulation scenario for the Designs 2.2 and 2.3 in which at least one interacting SNP does not exhibit a marginal effect, the elastic net with the \(\lbrace 0,1 \rbrace\) SNP coding still induces inferior AUCs compared to logic bagging.
Real data application
We also compared the GRS construction approaches using a real data set from a German cohort study, the SALIA study (Study on the Influence of Air Pollution on Lung, Inflammation and Aging) [54], which included in total 4874 women that were at their first examination between 54 and 55 years old. The participants were recruited in 19851994 from highly industrialized areas and less industrialized areas in NorthRhine Westphalia, Germany. In 2006, a followup questionnaire was completed by 4027 women which contained questions about the diagnosis of certain diseases. In a further followup clinical examination conducted in 20072010, genetic data was also gathered. Here, we considered a data set consisting of 517 women from the SALIA study for which the presence of rheumatic diseases and genetic data are available. Furthermore, information about the exposure to specific air pollutants, i.e., nitrogen dioxide (\({\mathrm {NO}}_2\)), nitrogen oxide [nitrogen monoxide \({\mathrm {NO}}\) and nitrogen dioxide \({\mathrm {NO}}_2\)] (\({\mathrm {NO}}_x\)), particulate matter with an aerodynamic diameter of \(\le 2.5 \mu m\) or \(\le 10 \mu m\) (\({\mathrm {PM}}_{2.5}\) or \({\mathrm {PM}}_{10}\)), particulate matter with diameters of \(2.510 \mu m\) (\({\mathrm {PM}}_\mathrm {coarse}\)), and the reflectance of \({\mathrm {PM}}_{2.5}\) filters (\({\mathrm {PM}}_{2.5\ {\mathrm {absorbance}}}\)), is available at the time of performing the examinations in 2008. The assessment of the exposure to air pollution was conducted as part of the ESCAPE (European Study of Cohorts for Air Pollution Effects) project using landuse regression models [55, 56]. We used these air pollution exposures to assess GxE interactions. Information on covariables such as the BMI (body mass index), age, education status, smoking status, or workplace exposure for adjusting the final models is also available. In the questionnaire, it was asked whether any rheumatic disease was diagnosed. Thus, we considered prevalent rheumatic diseases as outcome in our analyses. Details on the SALIA study and the assessment of air pollution in this study are given by Krämer et al. [57] and Hüls et al. [58].
Selection of relevant genetic factors
In order to construct proper GRS for genes potentially having an impact on the development of rheumatic diseases, we selected several genes which showed to be influential in a literature research. For the selection of relevant genes, we mainly focused on rheumatoid arthritis, since it is the most common rheumatic disease besides osteoarthritis [59,60,61].
In around 70% to 90% of rheumatoid arthritis patients, anticitrullinated peptide antibodies (ACPA) can be detected [62]. For ACPApositive rheumatoid arthritis, many identified genetic associations belong to the human leukocyte antigen (HLA) class II complex [63]. Thus, we selected genes from the HLA class II complex for which associations with rheumatoid arthritis have been detected. In particular, we chose the HLADRB1 gene which presumably explains a large portion of the heritability of rheumatoid arthritis in the HLA class II complex [63,64,65,66]. Furthermore, we included the HLADPB1 and HLADOA genes which also might influence the risk of developing rheumatoid arthritis [66,67,68].
Since we started by including all available SNPs within the respective genes, 385 SNPs from the three genes formed our basis which we reduced by exploiting high states of LD. Using PLINK version 1.9 [69, 70], we performed LDbased clumping [71] (considering \(r^2 = 0.5\)). This procedure resulted in 72 tag SNPs which were used to construct the GRS.
We also constructed genomewide GRS based on a recent metaanalysis of GWAS regarding rheumatoid arthritis [72]. In this metaanalysis, only nonHLA loci were considered in contrast to the genebased selection. 70 of the proposed SNPs were available in our data and were used to fit the GRS models.
Geneenvironment interaction analysis
Additionally, we also analyzed GxE interaction effects. For the risk of developing ACPApositive rheumatoid arthritis, GxE interactions between HLA class II alleles and smoking have been discovered [73, 74]. It might be of interest if trafficrelated air pollution also interacts with genetic risk factors in the development of rheumatoid arthritis. Thus, our logistic regression models for the evaluation of GRS have the shape
for the environmental variable E and covariables \(C_1, \ldots , C_l\).
The selection of potential relevant covariables was performed in two steps. First, we applied a stepwise logistic regression with the AIC (Akaike information criterion) as the selection measure. This lead to the inclusion of the age, the BMI, the current smoking status, and the former smoking status. Next, we regarded this selection of variables in the final models jointly with the GRS and air pollutants. We excluded covariables which worsened the models, i.e., which lead to lower AUCs. After this procedure, only the age was left.
Analysis of association and predictive strength
The analysis was conducted in a repeated traintest split scheme. For 100 repetitions, we randomly divided the whole data set into 50% training data and 50% test data similar to Hüls et al. [11]. The respective training data sets were further randomly divided into 75% training data for hyperparameter tuning and 25% validation data (for the considered values of the hyperparameters, see “Section Hyperparameter optimization”). The best performing hyperparameter setting across the average of these 100 validation iterations was chosen.
Results of the real data application
A descriptive summary of the most important variables gathered in the data set from the SALIA study is given by Table 5. Most noticeably, we considered an unbalanced data set with 394 controls and 123 cases considering prevalent rheumatic diseases.
Univariate regression models
In the analysis of the data of the SALIA study, Table 6 summarizes the median pvalues of GRS analyzed in univariate regression models as in Eq. (5). When testing the influence of the GRS on the risk of developing rheumatoid arthritis, conventional random forests and logic bagging are the only models achieving significance at a significance level of 5% for at least 50% of the evaluations.
Figure 6 summarizes the test AUC values for the treebased statistical learning procedures and elastic net induced by univariate regression models only based on the GRS. For the genebased approach, most noticeably, random forests and logic bagging yield the highest AUCs where random forests achieves a slightly better performance than logic bagging. Ordinary logic regression and random forests with a prior variable selection induce similar results which cannot compete with conventional random forests and logic bagging. However, the elastic net yields the lowest AUCs. Here, the lower quartile of the AUCs yielded by the elastic net reaches 50%, i.e., the predictive performance of a (noninformative) constant classifier.
In addition to genebased GRS, we also constructed genomewide GRS based on a recent GWAS metaanalysis regarding rheumatoid arthritis [72]. A specific comparison of the predictive power between the genebased and genomewide approaches is summarized in Fig. 6. However, for the genomewide selection of SNPs, barely a signal can be observed in our sample as the AUCs on the test data sets were close to 50%. Thus, the genomewide GRS construction approach was not included in subsequent analyses. The inferior predictive performance compared to the genebased selection is possibly caused by the exclusion of HLA genes in the underlying metaanalysis. Nonetheless, the elastic net induces the lowest values for the AUC compared to the treebased methods which is in line with our previous experiments. In contrast to the genebased approach, random forests VIM yields a predictive power that can compete with ordinary random forests and logic bagging.
Geneenvironment interaction analysis
In the final adjusted models of the form as in Eq. (6), we regarded each air pollutant indicator separately and included the respective GxE interaction term. Neither the GRS themselves nor the GxE interaction terms are significant at a significance level of 5%. The concrete median pvalues of the 100 repetitions for the final adjusted models can be found in Additional file 1: Table S2.
Figure 7 depicts the predictive performance of the considered statistical learning algorithms for the induction of genebased GRS in multivariate regression models. Analogously to the univariate analysis, random forests and logic bagging yield the highest predictive power where the overall best values are reached for \({\mathrm {PM}}_{2.5}\). For this air pollutant, random forests achieves the best performance. The elastic net, random forests VIM, and logic regression yield similar performances which, again, cannot compete with random forests and logic bagging.
We also evaluated the GRS on the training data sets themselves. The best performing procedures random forests and logic bagging tend to heavily overfit the data as can be seen by the high discrepancy between the test and the training data analyses. These two algorithms achieve training AUCs of nearly 100% whereas the other methods lead to more homogeneous results. The corresponding AUCs can be found in Additional file 1: Fig. S24.
Smoking is a major risk factor for rheumatoid arthritis [75]. As can be seen in Table 5, the fractions of current smokers and former smokers in the excerpt from the SALIA study are higher among controls than among cases which is in contradiction to the literature. Since only 19.7% of the study participants in the data excerpt are current or former smokers, we conducted a sensitivity analysis excluding all current and former smokers from the data. Again, we are not able to identify any significant GxE interactions. The resulting AUCs are very similar to the former analysis. Random forests and logic bagging yield the highest test AUC values, whereas elastic net induces substantially lower values. The concrete results can be found in Additional file 1: Fig. S25.
Discussion
In this analysis, we evaluated treebased statistical learning approaches for the construction of GRS. We used the elastic net as a reference model and analyzed the treebased statistical learning methods in a simulation study considering several scenarios, focusing on marginal and epistatic genetic effects, respectively. To confirm our findings, we constructed and assessed GRS on a real data set from the German SALIA cohort study.
As our analyses showed, a modification of logic regression, namely logic bagging, was able to outperform the reference GRS construction procedure, the elastic net, in almost every scenario of the simulation study.
Similarly, logic bagging lead to a comparably strong predictive performance in the real data application. Logic regression could only compete when considering large effect sizes in the simulation studies and yielded inferior results in the analysis of the SALIA data. This indicates that logic regression fits highly variant models which can indeed benefit from a variance reduction via an ensemble approach like bagging. For larger genetic effects, bagging does not seem to be necessary due to a more consequent identification of the underlying signal.
Random forests lead to the best predictive performance on the real data set. Considering the simulation study, in a likewise comparable scenario, i.e., small data sets, low marginal genetic effects, and higher amounts of statistical noise, random forests could induce comparably high values for the AUC as well. In the analysis of marginal genetic effects, random forests’ performance decreased for increasing amounts of noise. This phenomenon can be partly explained by the random selection scheme of predictors for partitioning. The input variables are drawn with equal probabilities without replacement. Therefore, considering the setting with 44 noise SNPs in the first simulation scenario, in a decision tree branch where already three of the six influential SNPs and no noise are included, the probability of regarding one of the three remaining influential SNPs for the next split with the standard setting \(mtry = \lfloor \sqrt{50} \rfloor = 7\) is about only 39%. Thus, choosing a set of SNPs containing only statistical noise is more likely in this case. We also allowed higher settings for \(mtry\) in the hyperparameter optimization as could be seen in Table 4. For higher amounts of statistical noise, the higher setting for \(mtry\) could in fact increase the performance of random forests.
A related issue was the high amount of overfitting by random forests which could be observed in all three simulation scenarios as well as in the real data application. We addressed this by considering minimum terminal node sizes of up to 10% of the number of observations in each leaf and by performing a prior variable selection based on variable importance measures. The former solution, i.e., the tuning of the minimum node size, was important to optimize the performance on the general population, since the standard setting is set to one observation for classification trees. However, for appropriate probability estimates, Malley et al. [35] recommend choosing 10% of the total sample size.
The latter approach, i.e., the usage of random forests VIM, needed higher amounts of statistical noise and stronger marginal genetic effects to achieve test data performances comparable to random forests. Nonetheless, this alternative approach could substantially reduce the amount of overfitting in any case. Presumably caused by weak individual genetic effects, random forests VIM yielded an inferior predictive performance compared to ordinary random forests on the application to the SALIA data. However, in the analyses conducted by Speiser et al. [76], the random forests VIM approach utilizing the Boruta variable selection was able to yield lower error rates than conventional random forests. Thus, studies specifically comparing random forests variable selection procedures with conventional random forests in low signaltonoise ratio scenarios, such as applications considering SNP data, might be beneficial.
The reference procedure, the elastic net, could not compete with logic bagging and random forests when considering stronger genegene interaction effects. Even for solely marginal genetic effects, the regularization procedure had difficulties achieving AUCs as high as the ones of logic bagging. However, for strong GxE interaction effects, the elastic net could induce similar predictive performances as random forests. Before deciding to choose the penalty parameter \(\lambda\) based on the minimum crossvalidation error, we evaluated the elastic net based on the maximum \(\lambda\) which yielded a crossvalidation error in the range of one standard error of the minimum error. This approach is also recommended by Waldmann et al. [77] for GWASlevel amounts of SNPs and used by Hüls et al. [49] for the construction of GRS. However, in our applications including both the simulation study and the real data application, the elastic net had difficulties recognizing a signal at all with this approach which was presumably caused by high errors in general. Thus, we chose the minimizing \(\lambda\) which enhanced our fitted elastic net models.
In practice, the conventional \(\lbrace 0,1,2 \rbrace\) SNP coding is utilized when constructing GRS with regularized regression approaches such as the elastic net [11, 16]. Thus, we focused on this standard procedure in our analyses, which lead to comparatively weak performances. However, when splitting each considered SNP into two binary variables, i.e., when using the binary \(\lbrace 0,1 \rbrace\) SNP coding also for the elastic net, its performance in the simulation study increased due to now being able to differentiate between the dominant and recessive modes of inheritance. Therefore, the results for the \(\lbrace 0,1 \rbrace\) SNP coding suggest that it might be preferable to employ the \(\lbrace 0,1 \rbrace\) coding when fitting GRS using the elastic net. Nonetheless, logic bagging still yielded higher predictive performances than the elastic net in the genegene interaction simulation scenario when considering the \(\lbrace 0,1 \rbrace\) coding for all procedures.
The most important advantage of the treebased methods regarded in this article is to not being restricted to model assumptions such as linearity, i.e., being able to autonomously detect genegene interactions. The assumption of oversimplified genetic architectures in linear models might be the main cause for random forests and logic bagging outperforming the elastic net in most analyses. However, it is well known that genegene interactions also play a role in the heritability of diseases [8, 9].
Another practically interesting question would be, how well the introduced treebased methods can construct GRS for significantly larger amounts of SNPs, e.g., when using a broader SNP selection from GWAS. Winham et al. [22] found in their studies that for increasing amounts of SNPs, the identification of interactions becomes more difficult for random forests. For logic regression, with increasing amounts of explanatory variables, the amount of possible states increases linearly, therefore, requiring more simulated annealing iterations and generally deeper greedy searches and, hence, increasing the model fitting time. This model building time must be further increased when considering higher values for the parameters of maximum trees and maximum leaves which is reasonable due to potentially more influential predictors for more total input variables.
Unsurprisingly, elastic net models could be fitted and evaluated in the least amount of time due to their simplicity compared to the considered treebased models. Random forests with 2000 trees could be fitted and evaluated in less than 10 s in most cases. Random forests VIM needed slightly more time which was to be expected. Logic bagging models needed more time, however, conventional logic regression models utilizing simulated annealing as search procedure consumed the most amount of time and needed up to 1 minute for fitting and evaluating the GRS. In Additional file 1: Fig. S1, the concrete times for the third simulation scenario are depicted.
For increasing odds ratios, the measured sensitivity decreases in the marginal effects and genegene interaction effect simulation scenarios, which does not seem to be plausible at first glance. However, this phenomenon can be explained by the data structure considered in this analysis and the requirement to dichotomize the risk predictions into two classes for estimating the sensitivity and specificity. For constructing GRS, discrete input variables, more exactly SNPs exhibiting three different outcomes, are used. Thus, the constructed and possibly true underlying GRS also follow a discrete pattern depending on the SNP setting. For the marginal effects simulation scenario, there are 7 distinct GRS values in the true underlying model due to Eq. (2). In Additional file 1: Fig. S26, a corresponding GRS distribution is depicted. Due to the additivity in this model, the GRS just below 0.5 occurs in approximately 30% of all observations. Therefore, dichotomizing the GRS at 0.5 leads to classifying only 35% of all observations as cases which explains the low sensitivity in this setting. Lowering the classification threshold to a value such as 0.45 shifts the issue to the specificity, since, in this case, only 35% of all observations will be classified as controls. Thus, the sensitivities and specificities determined in this analysis need to be interpreted with caution because of the discrete nature of the considered input variables.
In our real data application, we analyzed a relatively small data set containing 517 observations with only 123 cases. The missing balance as well as the comparably low sample size complicated meaningful analyses, especially when considering the need for splitting the data set into training and test data sets. Generally, important covariates such as the smoking status and the BMI were not included in the final models due to lowering the predictive performance. This decrease in performance was presumably caused by the low sample size and amount of cases yielding unintuitive statistics such as the higher fraction of smokers among controls.
Conclusion
As our analyses on simulated as well as on real data showed, the treebased statistical learning methods random forests and logic bagging can be valuable tools for constructing GRS. Especially when little prior knowledge about the generesponse relationships is available or if no appropriate external weights for the regarded disease or population are available, these two algorithms should also be taken into consideration when building GRS. Regardless of the presence of genegene interactions in the heritability of a certain disease, the discussed methods have the potential to outperform regularized linear methods.
Availability of data and materials
All code for generating and accessing data for the simulation study is included in this published article as a supplementary information file (Additional file 2).
Abbreviations
 ACPA:

Anticitrullinated peptide antibody
 AIC:

Akaike information criterion
 AUC:

Area under the curve
 BMI:

Body mass index
 CART:

Classification and regression tree
 GLM:

Generalized linear model
 GRS:

Genetic risk score(s)
 GWAIS:

Genomewide association interaction study
 GWAS:

Genomewide association study
 GxE:

Geneenvironment
 HLA:

Human leukocyte antigen
 IQR:

Interquartile range
 LD:

Linkage disequilibrium
 MAF:

Minor allele frequency
 ROC:

Receiver operating characteristic
 SNP:

Single nucleotide polymorphism
 VIM:

Variable importance measure
References
Billings LK, Florez JC. The genetics of type 2 diabetes: what have we learned from GWAS? Ann N Y Acad Sci. 2010;1212(1):59–77.
Choi SW, Mak TSH, O’Reilly PF. Tutorial: a guide to performing polygenic risk score analyses. Nat Protoc. 2020;15(9):2759–72.
Dudbridge F. Power and predictive accuracy of polygenic risk scores. PLoS Genet. 2013;9(3):1–17.
Torkamani A, Wineinger NE, Topol EJ. The personal and clinical utility of polygenic risk scores. Nat Rev Genet. 2018;19(9):581–90.
Wray NR, Lin T, Austin J, McGrath JJ, Hickie IB, Murray GK, et al. From basic science to clinical application of polygenic risk scores: a primer. JAMA Psychiat. 2021;78(1):101–9.
Thomas M, Sakoda LC, Hoffmeister M, Rosenthal EA, Lee JK, van Duijnhoven FJB, et al. Genomewide modeling of polygenic risk score in colorectal cancer risk. Am J Hum Genet. 2020;107(3):432–44.
Kooperberg C, LeBlanc M, Obenchain V. Risk prediction using genomewide association studies. Genet Epidemiol. 2010;34(7):643–52.
GilbertDiamond D, Moore JH. Analysis of gene–gene interactions. Curr Protocols Human Genet. 2011;70(1):1.14.1–1.14.12.
Ritchie MD, Van Steen K. The search for genegene interactions in genomewide association studies: challenges in abundance of methods, practical considerations, and biological interpretation. Ann Transl Med. 2018;6(8):157.
Che R, MotsingerReif A. Evaluation of genetic risk score models in the presence of interaction and linkage disequilibrium. Front Genet. 2013;4:138.
Hüls A, Ickstadt K, Schikowski T, Krämer U. Detection of geneenvironment interactions in the presence of linkage disequilibrium and noise by using genetic risk scores with internal weights from elastic net regression. BMC Genet. 2017;18(1):55.
Ottman R. Geneenvironment interaction: definitions and study design. Prev Med. 1996;25(6):764–70.
Tibshirani R. Regression shrinkage and selection via the Lasso. J R Stat Soc Ser B (Methodol). 1996;58(1):267–88.
Zou H, Hastie T. Regularization and variable selection via the elastic net. J R Stat Soc Ser B (Stat Methodol). 2005;67(2):301–20.
Mavaddat N, Michailidou K, Dennis J, Lush M, Fachal L, Lee A, et al. Polygenic risk scores for prediction of breast cancer and breast cancer subtypes. Am J Human Genet. 2019;104(1):21–34.
Privé F, Aschard H, Blum MGB. Efficient implementation of penalized regression for genetic risk prediction. Genetics. 2019;212(1):65–74.
Breiman L. Random forests. Mach Learn. 2001;45(1):5–32.
Fu H, Zhang Q, Qiu G. Random forest for image annotation. In: Computer Vision—ECCV 2012. Berlin: Springer; 2012. p. 86–99.
Elagamy MN, Stanier C, Sharp B. Stock market random foresttext mining system mining critical indicators of stock market movements. In: 2018 2nd international conference on natural language and speech processing (ICNLSP); 2018. p. 1–8.
Hao M, Jiang D, Ding F, Fu J, Chen S. Simulating spatiotemporal patterns of terrorism incidents on the Indochina Peninsula with GIS and the random forest method. ISPRS Int J GeoInf. 2019;8(3):133.
Breiman L, Friedman J, Stone CJ, Olshen RA. Classification and regression trees. Boca Raton: CRC Press; 1984.
Winham SJ, Colby CL, Freimuth RR, Wang X, de Andrade M, Huebner M, et al. SNP interaction detection with Random Forests in highdimensional genetic data. BMC Bioinform. 2012;13(1):164.
Ruczinski I, Kooperberg C, LeBlanc M. Logic Regression. J Comput Graph Stat. 2003;12(3):475–511.
Schwender H, Ickstadt K. Identification of SNP interactions using logic regression. Biostatistics. 2007;9(1):187–98.
Kooperberg C, Ruczinski I. Identifying interacting SNPs using Monte Carlo logic regression. Genet Epidemiol. 2005;28(2):157–70.
Dinu I, Mahasirimongkol S, Liu Q, Yanai H, Sharaf Eldin N, Kreiter E, et al. SNPSNP interactions discovered by logic regression explain Crohn’s disease genetics. PLoS ONE. 2012;7(10):1–6.
Kruppa J, Ziegler A, König IR. Risk estimation and risk prediction using machinelearning methods. Hum Genet. 2012;131(10):1639–54.
Botta V, Louppe G, Geurts P, Wehenkel L. Exploiting SNP correlations within random forest for genomewide association studies. PLoS ONE. 2014;9(4):1–11.
Gola D, Erdmann J, MüllerMyhsok B, Schunkert H, König IR. Polygenic risk scores outperform machine learning methods in predicting coronary artery disease status. Genet Epidemiol. 2020;44(2):125–38.
Badré A, Zhang L, Muchero W, Reynolds JC, Pan C. Deep neural network improves the estimation of polygenic risk scores for breast cancer. J Hum Genet. 2021;66(4):359–69.
Yoo W, Ference BA, Cote ML, Schwartz A. A comparison of logistic regression, logic regression, classification tree, and random forests to identify effective genegene and geneenvironmental interactions. Int J Appl Sci Technol. 2012;2(7):268.
Hastie T, Tibshirani R, Friedman J. The elements of statistical learning: data mining, inference, and prediction. New York: Springer; 2009.
Li RH, Belford GG. Instability of decision tree classification algorithms. In: Proceedings of the eighth ACM SIGKDD international conference on knowledge discovery and data mining. New York: Association for Computing Machinery; 2002. p. 570–575.
Breiman L. Bagging predictors. Mach Learn. 1996;24(2):123–40.
Malley JD, Kruppa J, Dasgupta A, Malley KG, Ziegler A. Probability machines: consistent probability estimation using nonparametric learning machines. Methods Inf Med. 2012;51(1):74–81.
Provost F, Domingos P. Tree induction for probabilitybased ranking. Mach Learn. 2003;52(3):199–215.
Kursa MB, Rudnicki WR. Feature selection with the Boruta package. J Stat Softw. 2010;36(11):1–13.
Janitza S, Celik E, Boulesteix AL. A computationally fast variable importance test for random forests for highdimensional data. Adv Data Anal Classif. 2018;12(4):885–915.
Altmann A, Toloşi L, Sander O, Lengauer T. Permutation importance: a corrected feature importance measure. Bioinformatics. 2010;26(10):1340–7.
Degenhardt F, Seifert S, Szymczak S. Evaluation of variable selection methods for random forests and omics data sets. Brief Bioinform. 2017;20(2):492–503.
Wright MN, Ziegler A. ranger: a fast implementation of random forests for high dimensional data in C++ and R. J Stat Softw. 2017;77(1):1–17.
Kirkpatrick S, Gelatt CD, Vecchi MP. Optimization by simulated annealing. Science. 1983;220(4598):671–80.
Kooperberg C, Ruczinski I. LogicReg: Logic Regression; 2021. R package version 1.6.3.
Schwender H, Tietz T. logicFS: Identification of SNP Interactions; 2020. R package version 2.10.0.
Hoerl AE, Kennard RW. Ridge regression: biased estimation for nonorthogonal problems. Technometrics. 1970;12(1):55–67.
Friedman J, Hastie T, Tibshirani R. Regularization paths for generalized linear models via coordinate descent. J Stat Softw. 2010;33(1):1–22.
R Core Team. R: A Language and Environment for Statistical Computing. Vienna, Austria; 2020. Available from: https://www.Rproject.org/.
Schwender H, Fritsch A. scrime: Analysis of HighDimensional Categorical Data Such as SNP Data; 2018. R package version 1.3.5.
Hüls A, Krämer U, Carlsten C, Schikowski T, Ickstadt K, Schwender H. Comparison of weighting approaches for genetic risk scores in geneenvironment interaction studies. BMC Genet. 2017;18(1):115.
Li Q, Fallin MD, Louis TA, Lasseter VK, McGrath JA, Avramopoulos D, et al. Detection of SNPSNP interactions in trios of parents with schizophrenic children. Genet Epidemiol. 2010;34(5):396–406.
Pan D, Li Q, Jiang N, Liu A, Yu K. Robust joint analysis allowing for model uncertainty in twostage genetic association studies. BMC Bioinform. 2011;12(1):9.
Hanley JA, McNeil BJ. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology. 1982;143(1):29–36.
Alberg AJ, Park JW, Hager BW, Brock MV, DienerWest M. The use of “overall accuracy” to evaluate the validity of screening or diagnostic tests. J Gen Internal Med. 2004;19(5p1):460–465.
Schikowski T, Sugiri D, Ranft U, Gehring U, Heinrich J, Wichmann HE, et al. Longterm air pollution exposure and living close to busy roads are associated with COPD in women. Respir Res. 2005;6(1):152.
Beelen R, RaaschouNielsen O, Stafoggia M, Andersen ZJ, Weinmayr G, Hoffmann B, et al. Effects of longterm exposure to air pollution on naturalcause mortality: an analysis of 22 European cohorts within the multicentre ESCAPE project. Lancet. 2014;383(9919):785–95.
Eeftens M, Beelen R, de Hoogh K, Bellander T, Cesaroni G, Cirach M, et al. Development of land use regression models for PM2.5, PM2.5 absorbance, PM10 and PMcoarse in 20 European Study areas; results of the ESCAPE project. Environ Sci Technol. 2012;46(20):11195–205.
Krämer U, Herder C, Sugiri D, Strassburger K, Schikowski T, Ranft U, et al. Trafficrelated air pollution and incident type 2 diabetes: results from the SALIA cohort study. Environ Health Perspect. 2010;118(9):1273–9.
Hüls A, Krämer U, Herder C, Fehsel K, Luckhaus C, Stolz S, et al. Genetic susceptibility for air pollutioninduced airway inflammation in the SALIA study. Environ Res. 2017;152:43–50.
Vanhoof J, Declerck K, Geusens P. Prevalence of rheumatic diseases in a rheumatological outpatient practice. Ann Rheum Dis. 2002;61(5):453–5.
Jokar M, Jokar M. Prevalence of inflammatory rheumatic diseases in a rheumatologic outpatient clinic: analysis of 12626 cases. Rheumatol Res. 2018;3(1):21–7.
Sangha O. Epidemiology of rheumatic diseases. Rheumatology. 2000;39(suppl\_2):3–12.
Song YW, Kang EH. Autoantibodies in rheumatoid arthritis: rheumatoid factors and anticitrullinated protein antibodies. QJM Int J Med. 2009;103(3):139–46.
Kampstra AS, Toes RE. HLA class II and rheumatoid arthritis: the bumpy road of revelation. Immunogenetics. 2017;69(8):597–603.
Clarke A, Vyse TJ. Genetics of rheumatic disease. Arthritis Res Therapy. 2009;11(5):1–9.
Eyre S, Bowes J, Diogo D, Lee A, Barton A, Martin P, et al. Highdensity genetic mapping identifies new susceptibility loci for rheumatoid arthritis. Nat Genet. 2012;44(12):1336–40.
Raychaudhuri S, Sandor C, Stahl EA, Freudenberg J, Lee HS, Jia X, et al. Five amino acids in three HLA proteins explain most of the association between MHC and seropositive rheumatoid arthritis. Nat Genet. 2012;44(3):291–6.
Jiang L, Jiang D, Han Y, Shi X, Ren C. Association of HLADPB1 polymorphisms with rheumatoid arthritis: a systemic review and metaanalysis. Int J Surg. 2018;52:98–104.
Okada Y, Suzuki A, Ikari K, Terao C, Kochi Y, Ohmura K, et al. Contribution of a nonclassical HLA gene, HLADOA, to the risk of rheumatoid arthritis. Am J Human Genet. 2016;99(2):366–74.
Purcell S, Chang C. PLINK 1.9; 2021. Available from: www.coggenomics.org/plink/1.9/.
Chang CC, Chow CC, Tellier LC, Vattikuti S, Purcell SM, Lee JJ. Secondgeneration PLINK: rising to the challenge of larger and richer datasets. GigaScience. 2015;4:7.
Purcell S, Neale B, ToddBrown K, Thomas L, Ferreira MA, Bender D, et al. PLINK: a tool set for wholegenome association and populationbased linkage analyses. Am J Hum Genet. 2007;81(3):559–75.
Ha E, Bae SC, Kim K. Largescale metaanalysis across East Asian and European populations updated genetic architecture and variantdriven biology of rheumatoid arthritis, identifying 11 novel susceptibility loci. Ann Rheum Dis. 2021;80(5):558–65.
Källberg H, Padyukov L, Plenge RM, Rönnelid J, Gregersen PK, van der Helmvan Mil AHM, et al. Genegene and geneenvironment interactions involving HLADRB1, PTPN22, and smoking in two subsets of rheumatoid arthritis. Am J Human Genet. 2007;80(5):867–75.
Karlson EW, Deane K. Environmental and geneenvironment interactions and risk of rheumatoid arthritis. Rheum Dis Clin. 2012;38(2):405–26.
Ishikawa Y, Terao C. The impact of cigarette smoking on risk of rheumatoid arthritis: a narrative review. Cells. 2020;9(2):475.
Speiser JL, Miller ME, Tooze J, Ip E. A comparison of random forest variable selection methods for classification prediction modeling. Expert Syst Appl. 2019;134:93–101.
Waldmann P, Mészáros G, Gredler B, Fürst C, Sölkner J. Evaluation of the lasso and the elastic net in genomewide association studies. Front Genet. 2013;4:270.
Acknowledgements
Not applicable.
Funding
Open Access funding enabled and organized by Projekt DEAL. This work has been supported (in part) by the Research Training Group “Biostatistical Methods for HighDimensional Data in Toxicology” (RTG 2624, Project R3) funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation—Project Number 427806116). The SALIA followup study 2007–2010 was funded by the German Statutory Accident Insurance (DGUV) Grant No: 617.0FP266 and the European Community’s Seventh Framework Programme (FP 7/20072011) under the Grant Agreement Number 211250.
Author information
Authors and Affiliations
Contributions
ML, HS, and CW designed the simulation study. ML, SK, CW, and TS conceived the analyses of the real data application. The simulation study and real data evaluations were conducted by ML. ML was the major contributor in writing the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
The study was conducted in accordance to the declaration of Helsinki. The SALIA cohort study has been approved by the Ethics Committees of the RuhrUniversity Bochum and the Heinrich Heine University Düsseldorf. We received written informed consent from all participants.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 1
. Further evaluation results, hyperparameter descriptions, and method workflows. Additional methodological descriptions and results for the simulation study and the real data application.
Additional file 1
. Simulation study data generating code. R code for generating and accessing all data sets used in the simulation study.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Lau, M., Wigmann, C., Kress, S. et al. Evaluation of treebased statistical learning methods for constructing genetic risk scores. BMC Bioinformatics 23, 97 (2022). https://doi.org/10.1186/s1285902204634w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1285902204634w
Keywords
 Polygenic risk scores
 Epistasis
 Statistical learning
 Random forests
 Variable selection
 Logic regression
 Bagging
 Elastic net
 Simulation study