- Research article
- Open access
- Published:

# Extension of the bayesian alphabet for genomic selection

*BMC Bioinformatics*
**volume 12**, Article number: 186 (2011)

## Abstract

### Background

Two Bayesian methods, BayesC*π* and BayesD*π*, were developed for genomic prediction to address the drawback of BayesA and BayesB regarding the impact of prior hyperparameters and treat the prior probability *π* that a SNP has zero effect as unknown. The methods were compared in terms of inference of the number of QTL and accuracy of genomic estimated breeding values (GEBVs), using simulated scenarios and real data from North American Holstein bulls.

### Results

Estimates of *π* from BayesC*π*, in contrast to BayesD*π*, were sensitive to the number of simulated QTL and training data size, and provide information about genetic architecture. Milk yield and fat yield have QTL with larger effects than protein yield and somatic cell score. The drawback of BayesA and BayesB did not impair the accuracy of GEBVs. Accuracies of alternative Bayesian methods were similar. BayesA was a good choice for GEBV with the real data. Computing time was shorter for BayesC*π* than for BayesD*π*, and longest for our implementation of BayesA.

### Conclusions

Collectively, accounting for computing effort, uncertainty as to the number of QTL (which affects the GEBV accuracy of alternative methods), and fundamental interest in the number of QTL underlying quantitative traits, we believe that BayesC*π* has merit for routine applications.

## Background

High-density single nucleotide polymorphisms (SNPs) covering the whole genome are available in animal and plant breeding to estimate breeding values. First, individuals having SNP genotypes and trait phenotypes are used to estimate SNP effects (training), and then genomic estimated breeding values (GEBVs) are obtained for every genotyped individual using those effects. Currently, the number of SNP genotypes per individual amounts to tens of thousands, but, owing to the rapid advances in genomics, it will soon exceed millions at comparable costs. The statistical challenge is to estimate SNP effects in a situation where the number of training individuals is much smaller than the vast number of SNPs. For this purpose, Meuwissen et al. [1] presented two hierarchical Bayesian models, termed BayesA and BayesB, that are discussed extensively in animal and plant breeding research (e.g., [2–6]). The reason for their popularity is that their implementation as single site locus sampler is straightforward, computing time is reasonable, and both simulations [1, 7, 8] and real data analyses [9, 10] have shown that linkage disequilibrium (LD) between SNPs and quantitative trait loci (QTL) is exploited better than with least-squares or ridge-regression; hence, accuracies of GEBVs were higher for these Bayesian methods. Gianola et al. [11] pointed to statistical drawbacks of BayesA and BayesB that center around the prior for SNP effects. A priori, a SNP effect is zero with probability *π*, and normally distributed having mean zero and a locus-specific variance with probability (1-*π*). This locus-specific variance has a scaled inverse chi-square prior with few degrees of freedom and a scale parameter, , that is often derived from an assumed additive-genetic variance under certain genetic assumptions [11, 12]. It can be shown that the full-conditional posterior of a locus-specific variance has only one additional degree of freedom compared to its prior regardless of the number of genotypes or phenotypes. This conflicts with the concept of Bayesian learning, and as a consequence, shrinkage of SNP effects depends strongly on as detailed by [11]. This problem becomes even more important with increasing SNP density as shown later. There are at least two possibilities to overcome this drawback: First, a single effect variance that is common to all SNPs is used instead of locus-specific variances. Then, as shown later, the influence of is smaller. Second, the scale parameter of the inverse chi-square prior for locus-specific variances is treated as an unknown with its own prior. The first strategy is referred to as BayesC in the following and the second as BayesD.

Another drawback of BayesA and BayesB is that the probability *π* that a SNP has zero effect is treated as known. In BayesA, *π* = 0 so that all SNPs have non-zero effect, whereas in BayesB, *π* > 0 to accommodate the assumption that many SNPs have a zero effect. The shrinkage of SNP effects is affected by *π*, and thus should be treated as an unknown being inferred from the data. In the following, *π* is treated as an unknown in BayesC and BayesD, which will be referred to as BayesC*π* and BayesD*π*, respectively. Finally, the question arises how the estimated *π* is related to the number of QTL.

The objective of this study was to present two Bayesian model averaging methods that address the drawback of BayesA and BayesB regarding the impact of on shrinkage of SNP effects, and treat *π* as an unknown by using BayesC*π* and BayesD*π*. Simulations were conducted to analyze estimates of *π* for the ability to infer the number of QTL depending on the genetic architecture of a quantitative trait and training data size. Field data from North American Holstein bulls were used to estimate *π* for milk production traits, and to compare accuracies of GEBVs obtained by BayesA, BayesB, BayesC*π*, BayesD*π*, and ridge-regression. Cross-validations were applied in a setting where the additive-genetic relationships between training and validation bulls were low so that the accuracies of GEBVs were dominated by LD information. This criterion reveals the potential of genomic selection better than accuracy obtained by using training data sets that contain close relatives of validation bulls such as parents, full and half sibs. The reason is that future selection candidates in cattle breeding programs may not have close relatives in training when genomic selection is applied early in life [9].

## Methods

### Statistical Model

The general statistical model can be written as

where ** y** is an

*N*× 1 vector of trait phenotypes,

**X**is an incidence matrix of the fixed effects in

**,**

*β***is a vector with polygenic effects of all individuals in the pedigree,**

*u**K*is the number of SNPs,

*z*_{ k }is an

*N*× 1 vector of genotypes at SNP

*k*,

*a*

_{ k }is the additive effect of that SNP, and

**is a vector of residual effects. In this study, the only fixed effect in**

*e***was the overall mean**

*β**μ*, and SNP genotypes were coded as the number of copies of one of the SNP alleles, i.e., 0, 1 or 2.

### Prior specifications

The prior for *μ* was a constant; the prior for ** u**|

**A**, was normal with mean zero and variance , where

**A**is the numerator-relationship matrix and is the additive-genetic variance not explained by SNPs. The prior for

*a*

_{ k }depends on the variance, , and the prior probability

*π*that SNP

*k*has zero effect:

The models of this study differed in their specifications for *π* and . In BayesA, BayesB and BayesD*π*, denotes that each SNP has its own variance. Each of these variances has a scaled inverse chi-square prior with degrees of freedom *ν*_{
a
} and scale , and thus with probability (1-*π*) the marginal prior of *a*_{
k
} |*ν*_{
a
} , is a univariate student's t-distribution, . This is the model hierarchy proposed by [1], where was derived here from the expected value of a scaled inverse chi-square distributed random variable, ; hence,

where *ν*_{
a
} was 4.2 as in [1], and is the variance of the additive effect for a randomly sampled locus, which can be related to the additive-genetic variance explained by SNPs, , as

where *p*_{
k
} is the allele frequency of SNP *k*[11–13]. BayesC*π* and BayesD*π* are constructed as follows to address the lack of Bayesian learning in BayesA and BayesB.

In BayesC*π*, , i.e., the priors of all SNP effects have a common variance, which has a scaled inverse chi-square prior with parameters *ν*_{
a
} = 4.2 and , where is derived as for BayesA and BayesB. As a result, the effect of a SNP fitted with probability (1-*π*) comes from a mixture of multivariate student's t-distributions, . For example, assume that only 3 SNPs are used in the analysis, resulting in 4 possible models in which the effect of SNP 1, say, is not zero (Table 1). Each of these models has a different multivariate t-prior, where the univariate t-distribution is regarded here as a special case of the multivariate distribution. Thus, across the 4 models, the effect of SNP 1 comes from a mixture of multivariate t-distributions.

In BayesD*π*, the degrees of freedom for the scaled inverse chi-square prior of the locus-specific variances, *ν*_{
a
} , are treated as known with a value of 4.2 as in all other models, but the scale parameter, , is treated as an unknown with Gamma(1,1) prior. Thus, for a SNP fitted with probability (1-*π*), its effect comes from a mixture of univariate student's t-distributions. In this case, the mixture is due to treating as unknown with a gamma prior.

The other parameter that must be specified for the prior of *a*_{
k
} in (1) is *π*, which is treated as known with *π* = 0 for BayesA and, in this paper, with *π* = 0.99 for BayesB. In BayesC*π* and BayesD*π*, in contrast, *π* is treated as an unknown with uniform(0,1) prior.

The prior for the residual effects is normal with mean zero and variance , and the priors for and are scaled inverse chi-square with arbitrarily small value of 4.2 for the degrees of freedom, and scale parameters and , respectively. These scale parameters were derived by the formula , where is the *a priori* value of or .

## Inference of model parameters

Two Markov Chain Monte Carlo (MCMC) algorithms were implemented to infer model parameters: one for BayesA, BayesB, and BayesD*π* and the other one for BayesC*π*. The differences between these two algorithms result from how the variances of SNP effects are modeled and lead to different strategies for including a SNP in the model.

### Algorithm for BayesA, BayesB and BayesD*π*

BayesA is a special case of BayesB with *π* = 0. The variables *μ*, *a*_{
k
} , ** u**, , , as well as and

*π*of BayesD

*π*are sampled by Gibbs-steps using their full-conditional posteriors, whereas the decision to fit SNP

*k*into the model and the value of its locus-specific variance, , are sampled by a Metropolis-Hastings (MH) step. In contrast to Meuwissen et al. [1], who implemented BayesA using Gibbs sampling, BayesA is implemented here as BayesB with

*π*= 0 and a reduced number of MH steps.

The MH step used in this study differs from that described for BayesB in [1]. In their implementation, the candidate for is sampled from the scaled inverse chi-square prior with probability (1 - *π*), whereas a model without SNP *k* is proposed with probability *π*. In the latter case both *a*_{
k
} and are equal to zero. The acceptance probability for the candidate sample in iteration *t* from the currently accepted variance, , to the candidate value, , is

where and denote densities of the data model given and , respectively, and all other model parameters denoted by ELSE as in Sorensen and Gianola [14], except for *a*_{
k
} which is integrated out here. Values of *π* close to 1 lead to candidate samples that are mostly 0, and thus in poor mixing. To increase the probability of non-zero candidates, the MH step is repeated 100 times in each iteration of the MCMC algorithm.

The proposal distribution for used here is different from the prior. Regardless of *π*, the candidate for is sampled with probability 0.5 from a scaled inverse chi-square, and with probability 0.5 from a point mass on zero, which reduces the number of MH steps required for mixing. The number of MH steps used here was 10. Further, the scale parameter of the candidate is chosen depending on whether SNP *k* was in the model in the previous iteration *t* - 1 or not, i.e., whether or equals to zero:

The acceptance probability is

where the prior for is

and its proposal is

This proposal is expected to have better mixing than that of [1] for extreme values of *π*. The acceptance probability is equivalent to equation 2.4 in Godsill (2001) [15].

After has been updated, *a*_{
k
} is sampled from

where and . After and *a*_{
k
} have been updated for all *K* SNPs, the polygenic effects in ** u** are sampled by the technique of [16] as described in [14] using an iterative algorithm to solve the mixed model equations; is sampled from a scaled inverse chi-square with degrees of freedom and scale , where

*n*

_{ u }is the number of individuals in the pedigree; is sampled from a scaled inverse chi-square with and , where

*n*is the number of training individuals. In BayesD

*π*, is sampled from a gamma with shape and scale , where

*m*

^{(t)}is the number of SNPs fitted in the model for iteration

*t*. The parameters of this gamma posterior show that information from all loci contributes to the posterior of the unknown scale parameter and therefore through it to the posteriors of the locus-specific variances. Finally,

*π*is drawn from Beta(

*K*-

*m*

^{(t)}+ 1,

*m*

^{(t)}+ 1). The starting value for

*π*was 0.5.

### Algorithm for BayesC*π*

The MCMC algorithm for BayesC*π* consists of Gibbs steps only, where those for *μ*, ** u**, , , and

*π*are identical to those in BayesD

*π*. In contrast, the decision to include SNP

*k*in the model depends on the full-conditional posterior for the indicator variable

*δ*

_{ k }, which is introduced for this very purpose. This indicator variable equals 1 if SNP

*k*is fitted to the model and is zero otherwise. Following general Bayesian rules, the full-conditional posterior probability that

*δ*

_{ k }= 1 is

where *p*(** y**|ELSE) =

*p*(

**|**

*y**δ*

_{ k }= 0, ELSE)

*p*(

*δ*

_{ k }= 1|

*π*); denotes the density of the data model given that SNP

*k*is fitted with common effect variance and the currently accepted values of all other parameters,

*p*(

**|**

*y**δ*

_{ k }= 0, ELSE) is the density of the data model without SNP

*k*,

*p*(

*δ*

_{ k }= 0|

*π*) =

*π*is the prior probability that SNP

*k*has zero effect, and correspondingly

*p*(

*δ*

_{ k }= 1|

*π*) = 1 -

*π*. The posterior for

*a*

_{ k }is identical to (4) except that replaces the locus-specific variance in

*c*

_{ k }so that . The common effect variance is sampled from a full-conditional posterior, which is a scaled inverse chi-square with degrees of freedom and scale , where

*m*

^{(t)}is the number of SNPs fitted with non-zero effect in iteration

*t*.

The starting value for *π*, *π*_{0}, determines as can be seen from equations (2) and (3). However, can affect to what extent *π* is used to shrink SNP effects, hence the estimate of *π*. As increases with *π*_{0}, less shrinkage is expected through , but shrinkage can be increased with larger *π* values, which can be regarded as a compensation for the lower shrinkage through . To examine the effect of *π*_{0} in BayesC*π*, results are given for *π*_{0} equal to 0.5, 0.8 and 0.95. The degrees of freedom of the scaled inverse chi-square prior, *ν*_{
a
} , also determine through formula (2), and thus can affect *π* estimates. However, in this study *ν*_{
a
} was not varied, but held constant at 4.2.

### Impact of on shrinkage in BayesC*π* compared to BayesA and BayesB

The parameters of this full-conditional distribution can be used to contrast the impact of on shrinkage in BayesC*π* compared to that in BayesA and BayesB. In the latter, the posterior of the locus-specific variance of SNP *k* is a scaled inverse chi-square distribution with degrees of freedom and scale [11]. That is, that posterior has only one more degree of freedom than the prior. In contrast, the full-conditional of the posterior of the common effect variance in BayesC*π* will have more degrees of freedom when *m*^{(t)}> 1 and the scale is less influenced by and more a function of the information contained in the data through .

The impact of on the shrinkage of SNP effects, especially for BayesA, increases with SNP density. The scale parameter, , decreases with increasing number of SNPs in the analyses due to , which depends on *π*. Hence, small SNP effects are regressed more towards zero than with a smaller number of SNPs in the model. Consider a chromosomal segment where a QTL is surrounded by many SNPs that are all in LD with the QTL. In the worst case, all these SNPs are collinear, which might occur for low effective population sizes. The QTL effect, even if large, will be distributed to all SNPs such that each SNP effect is small. As these effects are strongly regressed towards zero, the QTL effect can be completely lost.

### Software implementation

These Bayesian model averaging methods were implemented in *GenSel* software [17] and are available for web-based analysis of genomic data. It is accessible through BIGS.ansci.iastate.edu, and a user manual is attached to this manuscript in additional file 1.

## Simulations

Simulations were conducted to analyze estimates of *π* from BayesC*π* and BayesD*π* depending on the genetic architecture of an additive quantitative trait, and training data size. Two types of scenarios were simulated in this study. The first was an ideal scenario in which all loci were in mutual linkage equilibrium and genotypes of both SNPs and QTL were available for training and validation. The true value of *π* is the number of QTL divided by the total number of loci in this analysis. The second was a realistic scenario in which the loci were in LD and only SNPs were modeled. As a consequence the true value of *π* was unknown. In both scenarios, loci were biallelic with initial allele frequency of 0.5, and QTL effects were sampled either from a standard normal or from a gamma with shape 0.4 and scale 1.66 as in [1]. Figure 1 depicts the cumulative distribution functions of these two distributions to illustrate the different effect sizes simulated. The sampled QTL effects were standardized before training to exhibit the additive-genetic variance calculated from a specified heritability and a residual variance of 1. Trait phenotypes were simulated by adding residual effects sampled from a standard normal to the sum of the genotypic values. Simulations were varied with different numbers of QTL and training data sizes, which was either 1,000 or 4,000 individuals. The MCMC algorithms were run for 50,000 iterations with a burn-in of 20,000 iterations. A higher number of iterations did not change the results.

In the ideal simulations, a total of 2,000 loci were simulated as if they were all located on different chromosomes to ensure linkage equilibrium. The number of QTL among those loci was 10, 200, or 1,000 and trait heritability was 0.5. The realistic simulations started with a base population of 1,500 individuals that were randomly mated over 1,000 generations to generate LD from mutations and drift. Individuals of generation 1,000 were used as founders of a real pedigree from the North American Holstein population, which included 7,094 bulls used in the real data analysis. This simulated LD similar to that in real livestock populations [9]. Individuals from the last generation of pedigree individuals were parents of the training individuals, with each parent represented once. The simulated genome consisted of a single chromosome of length 1 M that had 4,000 evenly-spaced SNPs and either 10, 20, or 40 QTL that were randomly distributed on the chromosome. The mutation rate was 2.5•10^{-5} for both SNPs and QTL, which is larger than estimates of actual mutation rates to ensure that a sufficient number of loci was segregating after 1,000 generations of random mating; it can be shown that mutation rate has only a small effect on LD in this simulation using the formula derived by [18]. Recombinations were modeled according to a binomial map function, where the maximum number of uniformly and independently distributed crossovers on a chromosome of 1 M was 4 [19], i.e., assuming interference. The proportion of segregating loci after 1,000 generations of random mating was 0.98, hence the number of segregating QTL in the scenarios with 10, 20 and 40 QTL was 9.8, 19.6 and 39.2 on average, respectively. To select 2,000 SNPs for training and validation, the chromosome was first divided into 2,000 evenly-spaced bins and then one SNP with minor allele frequency greater than 0.05 was randomly selected in each bin. The heritability was varied with the values 0.03, 0.2 and 0.9 to modify the size of QTL effects. All simulations were repeated 24 times.

## Real data analyses

Data from North American Holstein bulls were used to gain information about the number of QTL affecting quantitative traits in real populations and to compare the different Bayesian methods with respect to GEBV accuracy that results mainly from LD information.

### Genotyped bulls

The data set consisted of 7,094 progeny tested North American Holstein bulls that were genotyped for the Illumina Bovine50K array, excluding bulls that had more than 5% missing genotypes. De-regressed breeding values obtained from the official genetic evaluation of the USDA in August 2009 were used as trait phenotypes and were available for the quantitative traits milk yield, fat yield, protein yield and somatic cell score. The de-regressed proofs of the bulls used had a reliability greater than 0.7 and the square root of the reliability was used to weight residual effects [20]. The average reliability of milk, fat and protein yield was 0.89 and that of somatic cell score 0.81. Furthermore, a pedigree, containing the bulls in cross-validation and their ancestors born after 1950, was available to model polygenic effects and to quantify additive-genetic relationships between training and validation bulls.

### SNP data

SNPs were selected for the analyses based on minor allele frequency (> 3%), proportion of missing genotypes (< 5%), proportion of mismatches between homozygous genotypes of sire and offspring (< 5%) and Hardy-Weinberg equilibrium (*p* < 10^{-10}). The total number of SNPs in the analyses was 40,764.

### Training and validation data sets

Bulls born between 1995 and 2004 were used for training, whereas 113 bulls born before 1995 and with additive-genetic relationships to the training bulls smaller than 0.1 were used for validation. The reason for generating this cross-validation scenario was that LD rather than additive-genetic relationships was to determine the accuracy of GEBVs. The contribution of LD information to the estimates of SNP effects is sensitive to the size of the training data set, and thus 1,000, 4,000 and 6,500 training bulls were randomly selected from the bulls born between 1995 and 2004.

The MCMC algorithms were run for 200,000 iterations with a burn-in of 150,000 iterations for 1,000 training bulls, 100,000 iterations with a burn-in of 50,000 iterations for 4,000 training bulls, and 50,000 iterations with a burn-in of 20,000 iterations for 6,500 training bulls. These numbers of iterations were sufficient in that a higher number did not change the results. Posterior distributions were visually inspected for convergence. In addition to the GEBVs obtained by the Bayesian model averaging methods, breeding values for the validation bulls were estimated using an animal model with the numerator-relationship matrix [21, 22], which provided standard pedigree-based BLUP-EBVs (P-BLUP) to quantify the genetic-relationship information from the pedigree. An animal model with a genomic relationship matrix [13] was used to obtain GEBVs (G-BLUP), which is equivalent to ridge-regression.

### Evaluation criteria

Estimates of *π* were studied as , where *K* is the number of loci used in the statistical analysis. This represents the posterior mean of the number of loci fitted in each iteration of the MCMC algorithm (*N*_{SNP}), which is more practical than *π* for comparisons of scenarios that differ in the number of simulated QTL. The reason is that the true value of *π* is usually unknown unless QTL are among the loci in the model. The accuracy of GEBVs was estimated by correlation between GEBVs and de-regressed proofs divided by the average accuracy of de-regressed proofs of the validation bulls. The GEBV of validation bull *i* was calculated as

where *z*_{
ik
} is the genotype score (0, 1, or 2) for bull *i* at SNP *k* and is the posterior mean of the effect at that locus. The EBVs from P-BLUP and G-BLUP were obtained from solutions of the animal model.

## Results

### Ideal scenario

Table 2 depicts the posterior number of SNPs fitted in the model (*N*_{SNP}) estimated by BayesC*π* and BayesD*π* starting with *π* = 0.5 according to the number of training individuals, number of QTL (*N*_{QTL}) and distribution of QTL effects, which all had a considerable effect on the results. A sufficiently large set of training data would include in the model only QTL and no spurious SNPs. With normal distributed QTL effects, BayesC*π* was more accurate in this regard than BayesD*π*, especially as training data size increased. BayesD*π* fitted substantially more loci than *N*_{QTL} in the scenarios with 10 and 200 QTL; in addition, *N*_{SNP} did not approach *N*_{QTL} as training data size increased in the scenario with 1,000 QTL. With gamma distributed QTL effects, *N*_{SNP} was always lower than *N*_{QTL} with BayesC*π*; BayesD*π*, in contrast, overestimated *N*_{QTL} when 10 and 200 QTL were simulated, but underestimated it for 1,000 QTL. Starting with *π* = 0.8 or 0.95 hardly changed *N*_{SNP} from BayesC*π* (results not shown); the only notable change was obtained for normal distributed QTL effects and *N*_{QTL} = 1,000, where *N*_{SNP} increased with training data size from 640 to 957.

### Realistic scenario

Table 3 shows *N*_{SNP} estimated by BayesD*π* for *h*^{2} = 0.9. Although *N*_{SNP} declined with decreasing *N*_{QTL}, *N*_{SNP} overestimated *N*_{QTL} considerably, and the training data size did not have an effect on *N*_{SNP}. The overestimation was even higher for heritabilities of 0.03 and 0.2 (not shown), but *N*_{SNP} decreased somewhat with increasing training data size. BayesC*π* with a starting value of *π* = 0.5 (Table 4) overestimated *N*_{QTL} less than BayesD*π* for *h*^{2} = 0.9, and significant trends were obtained for *N*_{SNP} with increasing training data size, which depended on the distribution of QTL effects, *N*_{QTL}, and *h*^{2}. For *h*^{2} = 0.9, *N*_{SNP} increased with training data size and the overestimation of *N*_{QTL} decreased with *N*_{QTL}. For *h*^{2} = 0.2, in contrast, the overestimation was higher with 1,000 training bulls, and *N*_{SNP} decreased significantly with training data size. For *h*^{2} = 0.03, *N*_{SNP} was generally high, and decreased with training data size except for 20 and 40 QTL with normally distributed effects. However, the trend with training data size for *h*^{2} = 0.03 was smaller than for the other two *h*^{2} values relative to the high *N*_{SNP} with 1,000 training individuals. Starting with *π* values of 0.8 and 0.95 (results not shown) did not change results for *h*^{2} = 0.9, but decreased the decay of *N*_{SNP} with training data size for *h*^{2} = 0.2, because estimates for *N*_{SNP} were smaller with 1,000 training individuals. The latter was also observed for *h*^{2} = 0.03 along with a decreasing trend for *N*_{SNP}.

### Real data analyses

Additive-genetic relationships between training and validation bulls were small: No validation bull had an additive-genetic relationship to a training bull exceeding 0.092. The distribution of the maximum additive-genetic relationships between training and validation bulls had a lower quartile, median, and upper quartile of 0.016, 0.05 and 0.07, respectively. The main cause of the low additive-genetic relationships was a separation of about three generations between the bulls of both data sets, because 90% of the validation bulls were born before 1975. Table 5 shows accuracies of P-BLUP, G-BLUP, and the Bayesian model averaging methods according to the quantitative trait and training data size. The accuracies of P-BLUP for fat and protein yield as well as somatic cell score were close to zero as expected, but the accuracy for milk yield was unexpectedly high with 0.15 and 0.24 for 1,000 and 4,000 training individuals, respectively.

Accuracies of GEBVs were similar for the different methods with the following exceptions: BayesB with *π* = 0.99 had the lowest accuracies for all traits but fat yield, and G-BLUP had the lowest accuracies for fat yield. Furthermore, the accuracies for milk yield obtained by BayesC*π* tended to be lower than for G-BLUP, BayesA and BayesD*π*. In general, BayesA tended to give the highest accuracies for all traits except for fat yield. The accuracies of BayesC*π* did not differ depending on the starting values for *π* (results only shown for starting *π* = 0.5).

The accuracy of GEBVs improved markedly with training data size for milk yield, fat yield and somatic cell score from 1,000 to 4,000 bulls, but improved only slightly or reduced from 4,000 to 6,500 bulls. The increase in accuracy with training data size for protein yield was less than for the other traits from 1,000 to 4,000 bulls, but tended to be more from 4,000 to 6,500 bulls. Somatic cell score had the highest relative increase in accuracy of all traits because accuracies were lowest for 1,000 training bulls. Interestingly, G-BLUP had the lowest accuracy for somatic cell score with 1,000 training bulls, but the increase was largest such that the accuracy for 6,500 bulls was as high as for BayesA.

The posterior distributions for *N*_{SNP} (not shown) were unimodal, symmetric, and standard deviations decreased notably with increasing training data size as in Table 6. Exceptions were the posterior distributions for protein yield and somatic cell score of BayesC*π* with 1,000 training bulls, which were bimodal and rather flat. Although the accuracies of BayesC*π* and BayesD*π* were very similar, they fitted very different numbers of SNPs (Table 6). As in the realistic simulations, *N*_{SNP} from BayesD*π* was insensitive to training data sizes for all traits, whereas BayesC*π* showed clear trends with training data size that differed across traits; *N*_{SNP} was comparatively low for milk and fat yield and increased with training data size, and estimates were very similar for the different starting values of *π*, *π*_{0}. *N*_{SNP} always decreased with training data size for protein yield, but estimates increased for all training data sizes as *π*_{0} decreased. For somatic cell score, however, the trends changed depending on *π*_{0}; *N*_{SNP} increased with training data size for *π*_{0} = 0.95, but decreased with lower *π*_{0} values.

## Discussion

Two Bayesian model averaging methods that address the statistical drawbacks of BayesA and BayesB were developed for genomic prediction. These two models were termed BayesC*π* and BayesD*π* to emphasize that the prior probability *π* that a SNP has zero effect was treated as an unknown. The objectives of this study were to evaluate the ability of these methods to make inferences about the number of QTL (*N*_{QTL}) of a quantitative trait by simulated and real data, and to compare accuracies of GEBVs from these new methods compared to existing methods.

### Simulations

In ideal simulations, all loci were in linkage equilibrium and both SNPs and QTL were modeled. BayesC*π* was able to distinguish the QTL that had non-zero effects from the SNPs that had zero effects as training data size increased and when QTL effects were normally distributed. In contrast, when QTL effects were gamma distributed many QTL remained undetected. This may have been because the gamma distribution generated fewer large effects and more small effects than the normal (Figure 1). Further, the prior of SNP effects in BayesC*π* given the common effects variance was normal and not gamma; a gamma prior may produce better results and should be investigated in a subsequent study. In conclusion, even in this ideal case the estimate of obtained from BayesC*π* is a poor indicator for *N*_{QTL}, unless the QTL distribution is normal. BayesD*π* was insensitive to *N*_{QTL} and inappropriate to estimate *N*_{QTL}.

In realistic simulations, SNPs and QTL were in LD and only the SNP genotypes were known. As expected, BayesC*π* fitted more SNPs than there were QTL, because every QTL was in LD with several SNPs. However, the number of SNPs fitted per QTL depended on both training data size and effect size of a QTL, which was varied here by the distribution of QTL effects, *h*^{2} and *N*_{QTL}; the size of simulated QTL effects increased with *h*^{2} and decreased with *N*_{QTL}. If QTL effects were generally large and easy to detect (Table 4, *h*^{2} = 0.9), *N*_{SNP} was small with 1,000 training individuals and increased with training data size. In addition, the larger a QTL effect, the more SNPs were fitted per QTL (Table 4, *h*^{2} = 0.9, 10 vs. 20 *N*_{QTL}). The cause for these findings may be that SNPs in low LD with the QTL were more likely to be fitted as either QTL effect size or training data size increased. The increase in *N*_{SNP} with training data size could also have been the result of detecting QTL with smaller effects. If QTL effects were smaller and less easy to detect (Table 4, *h*^{2} = 0.2), *N*_{SNP} was larger with 1,000 training individuals, which may be explained by false positive SNPs in the model, because the power of detection was likely to be low. In contrast to *h*^{2} = 0.9, *N*_{SNP} decreased substantially with training data size. However, the fact that *N*_{SNP} increased with training data size for *h*^{2} = 0.03, normally distributed QTL effects, and a starting value of *π* = 0.5 ( small) points to another explanation why many SNPs were fitted with small QTL effect size or small training data size: a higher number of SNPs explains differences between training individuals better than a smaller number, and thus more SNPs were required to explain those differences as training data size increased. In conclusion, BayesC*π* overestimates *N*_{QTL}, the extent depending on the size of QTL effects, which makes inference difficult. However, information about *N*_{QTL} can be gained by analyzing the trend of *N*_{SNP} with training data size, and starting with different *π* values. Furthermore, as SNP density increases in the future, overestimation of *N*_{QTL} is expected to be smaller, because LD between SNPs and QTL will be higher such that fewer SNPs are modeled per QTL. Sufficiently high SNP density guarantees near perfect LD between at least one SNPs and each QTL in which case the scenario of the ideal simulations will be approached.

### Real data analysis

#### Number of QTL and size of QTL effects

In agreement with the realistic simulations, estimates of *π* from BayesD*π* were insensitive to both trait and training data size (Table 6). BayesC*π*, in contrast, showed clear differences for both: *N*_{SNP} increased with training data size for milk and fat yield, and decreased for protein yield and somatic cell score. Thus, milk and fat yield may have more QTL with large effects than protein yield and somatic cell score, which can be derived from the trends of *N*_{SNP} in the realistic simulations. This is also supported by the accuracies of GEBVs where milk and fat yield had a higher accuracy than protein yield and somatic cell score. Furthermore, fat yield may have more QTL with large effect than milk yield, because both the increase of *N*_{SNP} from 1,000 to 4,000 training bulls and accuracy of GEBVs was higher for fat yield.

The number of SNPs in the model estimated by BayesC*π* may primarily result from the QTL with the largest effects, assuming that QTL with small effects were not detectable. The rather low accuracies of GEBVs and especially the low increase in accuracy from 4,000 to 6,500 bulls may also point to this conclusion, because many more training individuals seem to be necessary to estimate small QTL effects. Another reason may be that LD between SNPs and QTL was still too low, but this will change as SNP density increases; QTL with large effects will be estimated with fewer SNPs and additional QTL with smaller effect will be detected.

As mentioned earlier, a possible overestimation of *N*_{QTL} results from the fact that several SNPs are in LD with a QTL, where each of these SNPs explains a part of the QTL effect. These SNPs are likely to surround the QTL on the chromosome, and thus *N*_{QTL} can be estimated more precisely by calculating the variance of GEBVs explained by the effects of all SNPs in a specified chromosomal region. This can be done by defining a window containing a certain number of consecutive SNPs that are used to calculate this variance. By sliding the window over the chromosome and observing peaks that are higher than for single SNPs, *N*_{QTL} may be inferred better. This can be done with all methods that estimate SNP effects.

#### Comparison of the accuracy of GEBVs

North American Holstein bulls were partitioned into training and validation data sets such that bulls of both data sets were as unrelated as possible. As a result, the contribution of additive-genetic relationships to the accuracy of GEBVs was negligible for fat yield, protein yield and somatic cell score as demonstrated by the low accuracy of P-BLUP. However, that accuracy was unexpectedly high for milk yield, which might be an artifact of previous selection for milk yield because genotypes in the validation data set were only available from selected parents. Accuracies of GEBVs were similar for the different methods, and no one outperformed all the others across all traits or training data sizes. Nevertheless, BayesA performed remarkably well for this SNP density despite the statistical drawback of BayesA as described by [11]. However, as demonstrated in [11], it is important that the degrees of freedom used for the scaled inverse chi-square prior of the locus-specific variances express little prior belief. BayesA always fits all SNPs, hence the shrinkage of SNP effects results completely from the locus-specific variances, and, in contrast to the other methods, SNP effects are not fully shrunk to zero. Thus, even SNPs that truly have zero effects are expected to have small estimated effects adding noise to the GEBVs. This applies also to G-BLUP, which is equivalent to ridge-regression fitting all SNPs with equal variance. This did not seem to affect the accuracy of GEBVs here, but in the simulations of [1] BayesB performed better than BayesA and ridge-regression. The explanation may be that the traits analyzed here are determined by many more QTL than in those simulations. Thus, BayesA may be inferior to BayesC*π* and BayesD*π* for traits that are determined by only a few QTL and when many more SNPs effects are modeled as SNP density increases. Applying BayesA to the data sets of the realistic simulations with only 10 QTL confirmed its inferiority to BayesC*π* and BayesD*π*.

Treating *π* as known with a high value as in BayesB may be a poor choice. This agrees with Daetwyler et al. [23] who reported that G-BLUP outperformed BayesC with a fixed *π* when the number of simulated QTL was large. This can be explained partly by the fact that [23] considered the GEBV accuracy of the offspring of training individuals, meaning that genetic-relationships were important; these were captured better by the SNPs, when more SNPs were fitted as in G-BLUP [7]. Note further that BayesC with *π* = 0 is similar to G-BLUP. Consider ridge-regression as the equivalent model of G-BLUP to see this similarity. Both methods are equivalent either 1) if the single effect variance of BayesC is treated as known, 2) if *ν*_{
a
} is very large and equals to the single effect variance of ridge regression, or 3) if the single effect variance of ridge regression is treated as unknown with own scaled inverse chi-square prior. Thus the lower accuracy for BayesC in that study results most likely from treating *π* as known. Another reason may be that the scale parameter of the inverse chi-square prior for the common effect variance in [23] did not depend on the additive-genetic variance nor on the fixed *π* value as proposed by [1].

The finding that BayesC*π* and BayesD*π* give similar accuracies but different *π* values reveals that the two methods have different mechanisms for shrinking SNP effects. BayesD*π* primarily used the locus-specific variances, whereas BayesC*π* was only able to vary the shrinkage at different SNPs by using *δ*_{
k
} ; if a SNP is not fitted to the model the effect is shrunk completely to zero, otherwise they are all shrunk using the same ratio of residual to common effect variance. In principle, BayesD*π* is expected to be more flexible in shrinking SNP effects because it could use both locus-specific variances and *δ*_{
k
} for this purpose. The poor mixing of *π* in BayesD*π* indicates that locus-specific variances dominated over *δ*_{
k
} , which may explain why *π* is not an indicator for *N*_{QTL}.

#### Effect of training data size on Bayesian model averaging

Another insight into the mechanisms of Bayesian model averaging comes from the large increase in accuracy of GEBVs with training data size obtained by BayesB for milk yield. The parameter *π* was treated as known with value 0.99 resulting in about 400 SNPs fitted in each iteration of the MCMC algorithm for both 1,000 and 4,000 training bulls (Table 6). This indicates that *π* is a strong prior for *δ*_{
k
} = 0. Therefore, setting *π* = 0.99 is analogous to searching for models that fit about 400 SNPs in each iteration of the algorithm and to average them. These models change from one iteration to another as some SNPs are removed from the model, while others are included. This interchange of SNPs, however, is expected to be more frequent with a small training data size, because the power to detect significant SNPs is low. On the other hand, if the training data size is large, fewer SNPs are interchanged less often so that models differ less from one iteration to the other. This becomes most apparent in the increasing number of SNPs having moderate to high model frequency as training data size increased from 1,000 to 4,000 bulls as shown in Figure 2. The implication is that the effects of those SNPs were less shrunk with larger training data size, whereas effects of all other SNPs were shrunk more.

#### Comparison of GEBV accuracy with other studies

Accuracies of GEBVs reported by [24] and [2] for the North American and Australian Holstein populations, respectively, are not comparable to the accuracies found here. Accuracies for the milk production traits were higher in those studies, because validation bulls were closely related to those comprising the training data as demonstrated by [9]. In that study, accuracy of GEBVs due to LD was estimated from 1,048 and 2,096 German Holstein bulls using BayesB with *π* = 0.99. Most of those bulls were born between 1998 and 2004, and 60% were offspring of North American Holstein bulls revealing the high genetic relationships between the German and the North American Holstein population. The strategy used to estimate the accuracy due to LD was a regression approach based on pairs of training and validation data sets with different additive-genetic relationships between the bulls of both data sets. That strategy is very time-consuming when several methods must be compared, and therefore a different approach was chosen here. GEBV accuracies obtained by BayesB compared to those in [9] were similar for milk yield, comparable for fat yield when the training data size was greater than 1,000, but lower for protein yield and somatic cell score. The increase of accuracy with training data size tended to be higher in [9]. Moreover, in contrast to the present study, G-BLUP was inferior in [9]. The difficulties in comparing the accuracies found here to those in [9], apart from the standard errors, is that there might be genotype-environment interactions, because the environment in which the daughters of the bulls born before 1975 have been tested might be different from the environment of the last decade relevant to the daughters of the training bulls. In addition, selection and genetic drift may have changed the LD structure in the population so that the accuracies of this study may not represent the GEBV accuracies due to LD in the current population.

#### Computing time

Computing time, which may become more important as SNP density increases, is an advantage of BayesC*π*, because its Gibbs algorithm is faster than the Metropolis-Hastings algorithm of the other methods. The reason is that the MH step for sampling the locus-specific variances in this implementation of BayesA, BayesB and BayesD*π* is repeated in each iteration to improve mixing; the Gibbs step for fitting a SNP in BayesC*π* is only performed once. Furthermore, computing time depends largely on the number of SNPs fitted in each iteration, because the following two computation steps are the most demanding ones in the algorithm: The phenotypes have to be unadjusted for the genotypic effects of a SNP if that SNP was fitted in the previous iteration; similarly, if a new SNP effect was sampled in the current iteration, the phenotypes have to be adjusted for the genotypic effects of that SNP. BayesC*π* was more sensitive to both the genetic architecture of a trait and training data size than BayesD*π*, and thus computing time was shorter for BayesC*π*. In this implementation, BayesA always had the longest computing time because all SNPs were fitted. For example, using 1,000 training bulls for milk yield and a 2.4 GHz *AMD 280 Opteron* processor, computing time for 100,000 iterations was 10.3, 14.1, 18 and 21.3 hr for BayesC*π*, BayesB, BayesD*π* and BayesA, respectively.

## Conclusions

BayesC*π* and BayesD*π* that address the drawback of BayesA and BayesB regarding the impact of the prior hyperparameters on shrinkage of SNP effects and that treat as an unknown the prior probability *π* that a SNP has zero effect were developed for genomic prediction. Estimates of *π* from BayesC*π*, in contrast to those from BayesD*π*, are sensitive to training data size and SNP density, and provide information about the genetic architecture of a quantitative trait; the traits milk yield and fat yield measured in North American Holsteins have QTL with larger effects than protein yield and somatic cell score. The statistical drawback of BayesA and BayesB did not impair the GEBV accuracy that is mainly due to LD information. Accuracies of the alternative Bayesian methods were similar and none of them outperformed all others across all traits and training data sizes. Therefore the best method must be determined for each quantitative trait separately. In contrast to simulation studies, BayesA was a good model choice for genomic prediction in the North American Holstein population at this current SNP density. Treating *π* as known with a high value is not recommended as alternative methods such as BayesC*π* or BayesD*π* gave better accuracies. In general, computing time is shorter for BayesC*π* than for BayesD*π*, and longest for BayesA. Collectively, accounting for computing effort, uncertainty as to the number of QTL (which affects the GEBV accuracy of alternative methods), and fundamental interest in the number of QTL underlying quantitative traits, we believe that BayesC*π* has merit for routine applications.

## References

Meuwissen THE, Hayes BJ, Goddard ME: Prediction of total genetic value using genome-wide dense marker maps. Genetics 2001, 157(4):1819–1829. [http://www.genetics.org/cgi/content/abstract/157/4/1819]

Hayes BJ, Bowman PJ, Chamberlain AJ, Goddard ME: Invited review: Genomic selection in dairy cattle: Progress and challenges. J Dairy Sci 2009, 92(2):433–443. [http://jds.fass.org/cgi/content/abstract/92/2/433] 10.3168/jds.2008-1646

Jannink JL, Lorenz AJ, Iwata H: Genomic selection in plant breeding: from theory to practice. Briefings in Functional Genomics 2010, 9(2):166–177. [http://bfg.oxfordjournals.org/cgi/content/abstract/9/2/166] 10.1093/bfgp/elq001

Piepho HP: Ridge Regression and Extensions for Genomewide Selection in Maize. Crop Sci 2009, 49(4):1165–1176. [https://www.crops.org/publications/cs/abstracts/49/4/1165] 10.2135/cropsci2008.10.0595

Heffner EL, Sorrells ME, Jannink JL: Genomic Selection for Crop Improvement. Crop Sci 2009, 49: 1–12. [https://www.crops.org/publications/cs/abstracts/49/1/1] 10.2135/cropsci2008.08.0512

Zhong S, Dekkers JCM, Fernando RL, Jannink JL: Factors Affecting Accuracy From Genomic Selection in Populations Derived From Multiple Inbred Lines: A Barley Case Study. Genetics 2009, 182: 355–364. [http://www.genetics.org/cgi/content/abstract/182/1/355] 10.1534/genetics.108.098277

Habier D, Fernando RL, Dekkers JCM: The Impact of Genetic Relationship Information on Genome-Assisted Breeding Values. Genetics 2007, 177(4):2389–2397. [http://www.genetics.org/cgi/content/abstract/177/4/2389]

Habier D, Fernando RL, Dekkers JCM: Genomic Selection Using Low-Density Marker Panels. Genetics 2009, 182: 343–353. [http://www.genetics.org/cgi/content/abstract/182/1/343] 10.1534/genetics.108.100289

Habier D, Tetens J, Seefried FR, Lichtner P, Thaller G: The impact of genetic relationship information on genomic breeding values in German Holstein cattle. Genet Sel Evol 2010, 42: 5. 10.1186/1297-9686-42-5

Hayes B, Bowman P, Chamberlain A, Verbyla K, Goddard M: Accuracy of genomic breeding values in multi-breed dairy cattle populations. Genet Sel Evol 2009, 41: 51. [http://www.gsejournal.org/content/41/1/51] 10.1186/1297-9686-41-51

Gianola D, de los Campos G, Hill WG, Manfredi E, Fernando R: Additive Genetic Variability and the Bayesian Alphabet. Genetics 2009, 183: 347–363. [http://www.genetics.org/cgi/content/abstract/183/1/347] 10.1534/genetics.109.103952

Fernando RL, Habier D, Stricker C, Dekkers JCM, Totir LR: Genomic selection. Acta Agric Scand A Anim Sci 2008, 57(4):192–195.

VanRaden PM: Efficient Methods to Compute Genomic Predictions. J Dairy Sci 2008, 91(11):4414–4423. [http://jds.fass.org/cgi/content/abstract/91/11/4414] 10.3168/jds.2007-0980

Sorensen D, Gianola D: Likelihood, Bayesian, and MCMC Methods in Quantitative Genetics. Springer-Verlag; 2002.

Godsill SJ: On the Relationship Between Markov chain Monte Carlo Methods for Model Uncertainty. Journal of Computational and Graphical Statistics 2001, 10(2):230–248. [http://pubs.amstat.org/doi/abs/10.1198/10618600152627924] 10.1198/10618600152627924

García-Cortés LA, Sorensen D: On a multivariate implementation of the Gibbs-sampler. Genet Sel Evol 1996, 28: 121–126. 10.1186/1297-9686-28-1-121

Fernando RL, Garrick DJ: GenSel - user manual. 2009.

Ohta T, Kimura M: Linkage disequilibrium at steady state determined by random genetic drift and recurrent mutation. Genetics 1969, 63: 229–238.

Karlin S: Theoretical aspects of genetic map functions in recombination processes. In Human population genetics: The pittsburgh symposium. Edited by: Chakravarti A. New York: Van Nostrand Reinhold; 1984:209–228.

Garrick D, Taylor J, Fernando R: Deregressing estimated breeding values and weighting information for genomic regression analyses. Genet Sel Evol 2009, 41: 55. [http://www.gsejournal.org/content/41/1/55] 10.1186/1297-9686-41-55

Henderson CR: Sire evaluation and genetic trends. Animal Breeding and Genetics Symposium in Honor of Dr. Jay L. Lush. Champaign, IL., American Society of Animal Science and American Dairy Science Association 1973, 10–41.

Henderson CR: Best linear unbiased estimation and prediction under a selection model. Biometrics 1975, 31(2):423–447. 10.2307/2529430

Daetwyler HD, Pong-Wong R, Villanueva B, Woolliams JA: The Impact of Genetic Architecture on Genome-Wide Evaluation Methods. Genetics 2010, 185(3):1021–1031. [http://www.genetics.org/cgi/content/abstract/185/3/1021] 10.1534/genetics.110.116855

VanRaden PM, Van Tassell CP, Wiggans GR, Sonstegard TS, Schnabel RD, Taylor JF, Schenkel FS: Invited Review: Reliability of genomic predictions for North American Holstein bulls. J Dairy Sci 2009, 92: 16–24. [http://jds.fass.org/cgi/content/abstract/92/1/16] 10.3168/jds.2008-1514

## Acknowledgements

This research was further supported by the United States Department of Agriculture, National Research Initiative grant USDA-NRI-2009-35205-05100, and State of Iowa Hatch and Multi-state Research Funds.

## Author information

### Authors and Affiliations

### Corresponding author

## Additional information

### Competing interests

The authors declare that they have no competing interests.

### Authors' contributions

All authors contributed to the development of the statistical methods and to the program code of *GenSel* software. DH conducted the analyses and drafted the manuscript. All other authors contributed to the final manuscript, read and approved it.

## Electronic supplementary material

## Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

## Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## About this article

### Cite this article

Habier, D., Fernando, R.L., Kizilkaya, K. *et al.* Extension of the bayesian alphabet for genomic selection.
*BMC Bioinformatics* **12**, 186 (2011). https://doi.org/10.1186/1471-2105-12-186

Received:

Accepted:

Published:

DOI: https://doi.org/10.1186/1471-2105-12-186