Open Access

ParticleCall: A particle filter for base calling in next-generation sequencing systems

BMC Bioinformatics201213:160

DOI: 10.1186/1471-2105-13-160

Received: 14 March 2012

Accepted: 9 July 2012

Published: 9 July 2012

Abstract

Background

Next-generation sequencing systems are capable of rapid and cost-effective DNA sequencing, thus enabling routine sequencing tasks and taking us one step closer to personalized medicine. Accuracy and lengths of their reads, however, are yet to surpass those provided by the conventional Sanger sequencing method. This motivates the search for computationally efficient algorithms capable of reliable and accurate detection of the order of nucleotides in short DNA fragments from the acquired data.

Results

In this paper, we consider Illumina’s sequencing-by-synthesis platform which relies on reversible terminator chemistry and describe the acquired signal by reformulating its mathematical model as a Hidden Markov Model. Relying on this model and sequential Monte Carlo methods, we develop a parameter estimation and base calling scheme called ParticleCall. ParticleCall is tested on a data set obtained by sequencing phiX174 bacteriophage using Illumina’s Genome Analyzer II. The results show that the developed base calling scheme is significantly more computationally efficient than the best performing unsupervised method currently available, while achieving the same accuracy.

Conclusions

The proposed ParticleCall provides more accurate calls than the Illumina’s base calling algorithm, Bustard. At the same time, ParticleCall is significantly more computationally efficient than other recent schemes with similar performance, rendering it more feasible for high-throughput sequencing data analysis. Improvement of base calling accuracy will have immediate beneficial effects on the performance of downstream applications such as SNP and genotype calling.

ParticleCall is freely available at https://​sourceforge.​net/​projects/​particlecall.

Background

The advancements of next-generation sequencing technologies have enabled inexpensive and rapid generation of vast amounts of sequencing data [13]. At the same time, high-throughput sequencing technologies present us with the challenge of processing and analyzing large data sets that they provide. A fundamental computational challenge encountered in next-generation sequencing systems is the one of determining the order of nucleotides from the acquired measurements, the task typically referred to as base calling. The accuracy of base calling is of essential importance for various downstream applications including sequence assembly, SNP calling, and genotype calling [4]. Moreover, improving base calling accuracy may enable achieving desired performance of downstream applications with smaller sequencing coverage, which translates to a reduction in the sequencing cost.

A widely used sequencing-by-synthesis platform, commercialized by Illumina, relies on reversible terminator chemistry. Illumina’s sequencing platforms are supported by a commercial base-calling algorithm called Bustard. While Bustard is computationally very efficient, its base-calling error rates can be significantly improved by various computationally more demanding schemes [5]. Such schemes include work presented in [69]. Among the proposed methods, the BayesCall algorithm [8] has been shown to significantly outperform Bustard in terms of the achievable base calling error rates. By relying on a full parametric model of the acquired signal, BayesCall builds a Bayesian inference framework capable of providing valuable probabilistic information that can be used in downstream applications. However, its performance gains come at high computational costs. A modified version of the BaseCall algorithm named naiveBayesCall [9] performs base calling in a much more efficient way, but its accuracy deteriorates (albeit remains better than Bustard’s). Both BayesCall and naiveBayesCall rely on expectation-maximization (EM) framework that employs a Markov chain Monte Carlo (MCMC) sampling strategy to estimate the parameters of the statistical model describing the signal acquisition process. This parameter estimation step turns out to be very time-consuming, limiting practical feasibility of the proposed schemes. Highly accurate and practically feasible parameter estimation and base-calling remain a challenge that needs to be addressed.

In this paper, we propose a Hidden Markov Model (HMM) representation of the signal acquired by Illumina’s sequencing-by-synthesis platforms and develop a particle filtering (i.e., sequential Monte Carlo) base-calling scheme that we refer to as ParticleCall. When relying on the BayesCall’s Markov Chain Monte Carlo implementation of the EM algorithm (MCEM) to estimate system parameters, ParticleCall achieves the same error rate performance as BayesCall while reducing the time needed for base calling by a factor of 3. To improve the speed of parameter estimation, we develop a particle filter implementation of the EM algorithm (PFEM). PFEM significantly reduces parameter estimation time while leading to a very minor deterioration of the accuracy of base calling. Finally, we demonstrate that ParticleCall has the best discrimination ability among all of the considered base calling schemes.

Methods

In this section, we first review the data acquisition process and the basic mathematical model of the Illumina’s sequencing-by-synthesis platform. Then we introduce a Hidden Markov Model (HMM) representation of the acquired signals. Relying on the HMM model and particle filtering (i.e., sequential Monte Carlo) techniques, we develop a novel base calling and parameter estimation scheme and discuss some important practical aspects of the proposed method.

Illumina sequencing platform

A sequencing task on the Illumina’s platform is preceded by the preparation of a library of single-stranded short templates created by performing random fragmentation of the target DNA sample. Each single-stranded fragment in the library is placed on a glass surface (i.e., the flow cell [10]) and subjected to bridge amplification in order to create a cluster of identical copies of DNA templates [11]. The flow cell contains eight lanes where each lane is divided into a hundred of nonoverlapping tiles. The order of nucleotides in a DNA template is identified by synthesizing its complementary strand while relying on reversible terminator chemistry [3]. Ideally, in every sequencing cycle, a single fluorescently labeled nucleotide is incorporated into the complementary strand on each copy of the template in a cluster. The incorporated nucleotide is a Watson-Crick complement of the first unpaired base of the template. In reversible terminator chemistry, four distinct fluorescent tags are used to label the four bases, and are detected by CCD imaging technology. The acquired images are processed in order to obtain intensity signals indicating the type of nucleotide incorporated in each cycle. These raw signal intensities are then analyzed by a base-calling algorithm to infer the order of nucleotides in each of the templates.

Quality of the acquired raw signals is adversely affected by the imperfections in the underlying sequencing-by-synthesis and signal acquisition processes. The imperfections are manifested as various sources of uncertainties. For instance, a small fraction of the strands being synthesized may fail to incorporate a base, or they may incorporate multiple bases in a single test cycle. These effects are referred to as phasing and pre-phasing, respectively, and they result in an incoherent addition of the signals generated by the synthesis of the complementary strands on the copies of the template. Other sources of uncertainty are due to cross-talk and delay effects in the optical detection process, the residual effects that are readily observed between subsequent test cycles, signal decay, and measurement noise.

Overview of the mathematical model

To describe the signal acquired by the Illumina’s sequencing-by-synthesis platform, a parametric model was proposed in [8]. Basic components of the model are overviewed below.

A length-L DNA template sequence is represented by a 4×L matrix S, where the i th column of S, s i , is considered to be a randomly generated unit vector with a single non-zero entry indicating the type of the i th base in the sequence. We follow the convention where the first component of the vector s i corresponds to the base A, the second to C, the third to G, and the fourth to T and denote them as e A , e C , e G , e T . The goal of base-calling is to infer unknown S from the signals obtained by optically detecting nucleotides incorporated during the sequencing-by-synthesis process.

Let p denote the average fraction of strands that fail to extend in a test cycle. Phasing is modeled as a Bernoulli random variable with probability p. Let q denote the average fraction of strands which extend by more than one base in a single test cycle. Pre-phasing is modeled as a Bernoulli random variable with probability q. Length of the synthesized strand changes from i to j with probability
P ij = p , if j = i , 1 p q , if j = i + 1 , q , if j = i + 2 , 0 , otherwise. https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ1_HTML.gif
(1)
Let P denote an (L + 1)×(L + 1) transition matrix with entries P ij defined above, 1≤i,jL + 1. The signal generated over L cycles of the synthesis process is affected by phasing and pre-phasing and can be expressed as X=SH, where H=(H i,j) is an L×L matrix with entries H i , j = [ P j ] 1 ( i + 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq1_HTML.gif, the probability that a synthesized strand is of length i after j cycles. Here P j denotes the j th power of matrix P. The decay in signal intensities over cycles (caused by DNA loss due to primer-template melting, digestion by enzymatic impurities, DNA dissociation, misincorporation, etc.) is modeled by the per-cluster density random parameter λ t ,
λ t = ( 1 d t ) λ t 1 + ( 1 d t ) λ t 1 t , https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ2_HTML.gif
(2)
where t ( 0 , σ t 2 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq2_HTML.gif is a one-dimensional Gaussian random variable and d t is the per-cluster density decay parameter within [0,1]. We represent the t th column of H as h t and the t th column of X as x t . Incorporating the decay into the model, the signal generated in cycle t is expressed as
x t = λ t S h t , https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ3_HTML.gif
(3)
where x t = x t A x t C x t G x t T https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq3_HTML.gif is the vector of signals generated in each of the optical channels. Assuming Gaussian observation noise, the measured intensities at cycle t are given by
y t = K t x t + b { A , C , G , T } x t b η t b , https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ4_HTML.gif
(4)

where K t denotes the 4×4 crosstalk matrix describing overlap of the emission spectra of the four fluorescent tags, and η t A , η t C , η t G , η t T https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq4_HTML.gif are independent, identically distributed (i.i.d.) 4×1 Gaussian random vectors with zero mean and a common 4×4 covariance matrix Σ t .

Note that, due to typically small values of p and q, the components of the vector h t around its t th entry are significantly greater than the remaining ones. This observation can be used to simplify the expressions (2) and (3). In particular, let h t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq5_HTML.gif denote the vector obtained by windowing h t around its t th entry, i.e., by setting small components of h t to 0. In general, we consider l + r + 1 dominant components of h t centered at position t, H tl,t,H tl + 1,t,…,H t,t,…H t + r−1,t,H t + r,t, and then expression (2) becomes
x t x t w = λ t S h t w = λ t i = l r s t + i H t + i , t . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ5_HTML.gif
(5)
Finally, note that the signal measured in cycle t is empirically observed to contain residual effect from the previous cycle. The residual effect is modeled by adding α t (1− d t )y t−1to y t , where the unknown parameter α t (0,1). Therefore, the model can be summarized as
λ t | λ t 1 ( 1 d t ) λ t 1 , ( 1 d t ) 2 λ t 1 2 σ t 2 , y t | y t 1 , S , λ t K t x t w + α t ( 1 d t ) y t 1 , x t w 2 2 Σ t , s t Unif e A , e C , e G , e T , x t w = λ t S h t w t = 1 , 2 , , L https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ6_HTML.gif
(6)

where ·2 denotes the l2-norm of its argument, and where y0=0, λ0=1.

Hidden Markov Model of DNA base-calling

In this section, we reformulate the statistical description of the signal acquired by the Illumina’s sequencing-by-synthesis platform as a Hidden Markov Model (HMM) [12]. HMMs comprise a family of probabilistic graphical models which describe a series of observations by a “hidden” stochastic process and are generally suitable for representing time series data. Sequencing data obtained from the Illumina’s platform is a set of time-series intensities y1:L, motivating the HMM representation. HMMs provide a convenient framework for state and parameter estimation, which we exploit to develop a particle filter base-calling scheme in the next section.

For the sake of convenience, we remove the dependency between subsequent observations y t−1 and y t by defining y t = y t α t ( 1 d t ) y t 1 , t = 1 , 2 , , L https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq6_HTML.gif. Therefore, we can write
y t | S , λ t K t x t w , x t w 2 2 Σ t . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ7_HTML.gif
(7)
Components of y 1 : L https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq7_HTML.gif are the observations of our HMM, and depend on the underlying signals x1:L. Moreover, let S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq8_HTML.gif denote the 4×(l + r + 1) windowed submatrix of S, i.e.,
S t w = [ s t l s t l + 1 s t s t + r ] . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ8_HTML.gif
(8)

Since x t w = λ t S h t w = λ t i = t l t + r s i H i , t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq9_HTML.gif, it is clear that y t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq10_HTML.gif depends on λ t and S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq11_HTML.gif. Therefore, we define the state of the HMM to be the combination of λ t and S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq12_HTML.gif – the per-cluster density at cycle t and the collection of (l + r + 1) bases around (and including) the base in position t, respectively.

The proposed HMM representation is illustrated in Figure 1. The observation dynamics that characterize the relationship between y t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq13_HTML.gif and the hidden states ( S t w , λ t ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq14_HTML.gif are given by the distribution g ( y t | S t w , λ t ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq15_HTML.gif. It is straightforward to show from (5) that
g y t | S t w , λ t K t x t w , x t w 2 2 Σ t . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ9_HTML.gif
(9)
https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Fig1_HTML.jpg
Figure 1

A hidden Markov model of the generated signal in Illumina sequencing-by-synthesis platforms. An illustration of the graphical HMM of the Illumina’s sequencing platform. The observations y represent signal intensities after the removal of residual effects. The states are the combinations of S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq16_HTML.gif and λ t , which represent a subsequence of the template centered at position t and per-cluster density, respectively.

On the other hand, the state transition dynamics is described by the transition probability between subsequent states, ( S t 1 w , λ t 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq17_HTML.gif and ( S t w , λ t ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq18_HTML.gif. Since S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq19_HTML.gif and λ t are independent, the transition probability is
f ( S t w , λ t | S t 1 w , λ t 1 ) = f 1 ( S t w | S t 1 w ) f 2 ( λ t | λ t 1 ) . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ10_HTML.gif
(10)
The second term on the right-hand side of (8), f2( λ t |λ t−1), is known from the density decay model (1),
f 2 ( λ t | λ t 1 ) ( ( 1 d t ) λ t 1 , ( 1 d t ) 2 λ t 1 2 σ t 2 ) . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ11_HTML.gif
(11)
For notational convenience, we use s t , 1 w , , s t , l + r + 1 w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq20_HTML.gif to denote the set of l + r + 1 column vectors of S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq21_HTML.gif. Note that for k=2,3,…,l + r + 1, the column vectors s t 1 , k w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq22_HTML.gif in S t 1 w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq23_HTML.gif and the column vectors s t , k 1 w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq24_HTML.gif in S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq25_HTML.gif actually represent the same base. Therefore, the transition model between them can be represent by a δ function as
p ( s t , k 1 w | s t 1 , k w ) = δ { s t , k 1 w = s t 1 , k w } = 1 , if s t , k 1 w = s t 1 , k w , 0 , if s t , k 1 w s t 1 , k w . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ12_HTML.gif
(12)
Let U({ e A , e C , e G , e T }) denote a uniform distribution on the support set of unit vectors ({ e A , e C , e G , e T }). We assume no correlation between consecutive bases of the template sequence, i.e., s t , l + r + 1 w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq26_HTML.gif is generated from U({ e A , e C , e G , e T }). Therefore, f 1 ( S t w | S t 1 w ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq27_HTML.gif can be written as
f 1 ( S t w | S t 1 w ) = k = 2 l + r + 1 δ { s t 1 , k w = s t , k 1 w } u s t , l + r + 1 w , https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ13_HTML.gif
(13)

where u(·)U({ e A , e C , e G , e T }). Hereby, all the components of the HMM are specified.

ParticleCall base-calling algorithm

The goal of base calling is to determine the order of nucleotides in a template from the acquired signal y1:t. This can be rephrased as the problem of inferring the most likely sequence of states ( S t w , λ t ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq28_HTML.gif of the HMM in (7)-(8) from the observed sequence y 1 : t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq29_HTML.gif(clearly, s1:L follows directly from S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq30_HTML.gif). We assume that the parameters Λ={p,q,d1:L,α1:L,σ1:L,K1:L,Σ1:L} are common for all clusters within a tile, and that they are provided by a parameter estimation step discussed in the following section. In this section, we introduce a novel base calling algorithm ParticleCall which relies on particle filtering techniques to sequentially infer ( S t w , λ t ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq31_HTML.gif and, therefore, recover the matrix S.

In general, particle filtering (i.e., sequential Monte Carlo) methods generate a set of particles with associated weights to estimate the posteriori distribution of unknown variables given the acquired measurements [13]. In the proposed HMM framework, we sequentially calculate the posteriori distribution of the columns of S p(s t | y 1 : t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq32_HTML.gif),t=1,2,…,L, and find the maximum a posteriori (MAP) estimates of s t by solving
s ̂ t = arg max s t { e A , e C , e G , e T } { p ( s t | y 1 : t ) } . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ14_HTML.gif
(14)

Our algorithm relies on a sequential importance sampling/resampling (SISR) particle filter scheme [14] to calculate p ( S t w , λ t | y 1 : t ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq33_HTML.gif. Different choices and approximation methods of proposal densities are considered in [1517]. We directly use the transition (8) as the proposal density. This sequential importance sampling suffers from degeneracy and the variance of the importance weights will increase over time. To address the degeneracy problem, a resampling step is introduced in order to eliminate samples which have small normalized importance weights. Common resampling methods include multinomial resampling [14], residual resampling [18] and systematic resampling [19, 20]. We measure degeneracy of the algorithm using the effective sample size Keff and, for the sake of simplicity, employ multinomial resampling strategy. If we denote the number of particles by N p and associated weights by w, then K eff = ( k = 1 N p ( w t ( i ) ) 2 ) 1 https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq34_HTML.gif and resampling step is used when Keff is below a fixed threshold N threshold . N threshold of size O( N p ) is typically sufficient [14]. In our implementation, we set N threshold = N p /2.

We omit further details for brevity and formalize the ParticleCall algorithm below.

Algorithm 1

ParticleCall base-calling algorithm
  1. 1.

    Initialization:

     

1.1 Initialize particles:for i=1→ N p do

Sample each column of the submatrix S 1 w , ( i ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq35_HTML.gif from U({ e A , e C , e G , e T }); Sample λ 1 ( i ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq36_HTML.gif from a Gaussian distribution with mean 1, and the variance calculated using Bustard’s estimates of λin the first 10 test cycles.end for

1.2 Compute and normalize weights for each particle according to w 1 ( i ) g ( y 1 | S 1 w , ( i ) , λ 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq37_HTML.gif as in (7).
  1. 2.

    Run iteration t(t≥2):

     

2.1 Sampling:for i=1→ N p do

Sample S t w , ( i ) , λ t ( i ) f ( · , · | S t 1 w , ( i ) , λ t 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq38_HTML.gif according to (8).end for

2.2 Update the importance weight
w t ( i ) w t 1 ( i ) g y t | S t w , ( i ) , λ t ( i ) . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ15_HTML.gif
(15)

2.3 Normalize the weights. Calculate the posteriori probability of s t and obtain the estimate s ̂ t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq39_HTML.gif.

2.4 Resampling:if K eff = k = 1 N p ( w t ( i ) ) 2 1 N threshold https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq40_HTML.gifthen

Draw N p samples { S ̄ t w , ( j ) , λ ̄ t ( j ) , j = 1 , , N p } https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq41_HTML.gif from { S t w , ( i ) , λ t ( i ) , i = 1 , , N p } https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq42_HTML.gif with probabilities proportional to { w t ( i ) , i = 1 , , N p } https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq43_HTML.gif. Assign equal weight to each particle, w ̄ t ( i ) = 1 / N p https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq44_HTML.gif.end if

Since S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq45_HTML.gif in the HMM states are discrete with a finite alphabet, and the transitions of S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq46_HTML.gif and λ t are independent according to (8), it is possible to Rao-Blackwellize the ParticleCall algorithm. Rao-Blackwellization is used to marginalize part of the states in the particle filter, hence reducing the number of needed particles N p [16]. We marginalize the discrete states S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq47_HTML.gif and reduce the hidden process to λ t , while relying on the particle filter to calculate p(λ1:t| y 1 : t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq48_HTML.gif).

The original posterior distribution of the states can be expressed as
p λ 1 : t , S 1 : t w | y 1 : t = p S 1 : t w | y 1 : t , λ 1 : t p λ 1 : t | y 1 : t . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ16_HTML.gif
(16)

Since p ( λ 1 : t | y 1 : t ) p ( y t | y 1 : t 1 , λ 1 : t ) p ( λ t | λ t 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq49_HTML.gif, where λ t 1 ( i ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq50_HTML.gif is a sample from p(λ1:t−1| y 1 : t 1 https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq51_HTML.gif), we can state the Rao-Blackwellized ParticleCall algorithm as below.

Algorithm 2

Rao-Blackwellized ParticleCall algorithma

1. Initialization:

1.1 Initialize particles:for i=1→ N p do

Sample λ 1 ( i ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq52_HTML.gif from a Gaussian distribution with mean 1, and the variance calculated using Bustard’s estimates of λin the first 10 test cycles.end for

1.2 Compute and normalize weights for each particle according to w 1 ( i ) g ( y 1 | λ 1 ( i ) ) S 1 w g ( y 1 | S 1 w , λ 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq53_HTML.gif.

1.3 Calculate the discrete distribution p ( S 1 w | y 1 , λ 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq54_HTML.gif for each i.

2. Run iteration t(t≥2):

2.1 Sampling:for i=1→ N p do

Sample λ t ( i ) f ( · | λ t 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq55_HTML.gif.end for

2.2 Update the importance weight w t ( i ) w t 1 ( i ) g ( y t | y 1 : t 1 , λ 1 : t ( i ) ) . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq56_HTML.gif and normalize the weights.

2.3 Resample if Keff N threshold

2.4 Update p ( S t w | y 1 : t , λ 1 : t ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq57_HTML.giffor i=1→ N p do

Update p ( S t w | y 1 : t , λ 1 : t ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq58_HTML.gif using p ( S t 1 w | y 1 : t 1 , λ 1 : t 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq59_HTML.gif and λ t ( i ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq60_HTML.gif.end for

In step 2.2 of Algorithm 2, the quantity g ( y t | y 1 : t 1 , λ 1 : t ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq61_HTML.gif can be obtained by marginalizing over discrete states S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq62_HTML.gif and S t 1 w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq63_HTML.gif,
g y t | y 1 : t 1 , λ 1 : t ( i ) = S t w p y t | y 1 : t 1 , S t w , λ 1 : t ( i ) p S t w | y 1 : t 1 , λ 1 : t ( i ) = S t w p y t | S t w , λ t ( i ) S t 1 w p S t w | S t 1 w , y 1 : t 1 , λ 1 : t ( i ) × p S t 1 w | y 1 : t 1 , λ 1 : t ( i ) , https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ17_HTML.gif
(17)

where p ( y t | S t w , λ t ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq64_HTML.gif is the observation density, p ( S t 1 w | https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq65_HTML.gif y 1 : t 1 , λ 1 : t ( i ) ) = p ( S t 1 w | y 1 : t 1 , λ 1 : t 1 ( i ) ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq66_HTML.gif due to the independence of the state transitions, and p ( S t w | S t 1 w , y 1 : t 1 , λ 1 : t ( i ) ) = https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq67_HTML.gif p ( S t w | S t 1 w ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq68_HTML.gif due to the Markov property and the independence of the state transitions.

In step 2.4 of Algorithm 2, the update equation is obtained as
p S t w | y 1 : t , λ 1 : t ( i ) p S t w , y t , λ t ( i ) | y 1 : t 1 , λ 1 : t 1 ( i ) = S t 1 w p y t , S t w , λ t ( i ) | y 1 : t 1 , S t 1 w , λ 1 : t 1 ( i ) × p S t 1 w | y 1 : t 1 , λ 1 : t 1 ( i ) = S t 1 w p y t | S t w , λ t ( i ) , y 1 : t 1 , S t 1 w , λ 1 : t 1 ( i ) × p S t w , λ t ( i ) | y 1 : t 1 , S t 1 w , λ 1 : t 1 ( i ) × p S t 1 w | y 1 : t 1 , λ 1 : t 1 ( i ) = p y t | S t w , λ t ( i ) p λ t ( i ) | λ t 1 ( i ) × S t 1 w p S t w | S t 1 w p S t 1 w | y 1 : t 1 , λ 1 : t 1 ( i ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ18_HTML.gif
(18)

Parameter estimation

To determine the set of parameters Λneeded to run the proposed ParticleCall base calling algorithm, one could rely on the MCMC implementation of the EM algorithm (MCEM) proposed in [8]. In section Results and discussion, we demonstrate the performance of the ParticleCall algorithm that relies on the MCEM parameter estimation scheme. Note, however, that the MCMC sampling strategy employed by MCEM requires a lengthy burn-in period and a very large sample size to perform the expectation step. Therefore, the MCEM parameter estimation scheme is computationally rather intensive and requires significant computational resources if it is to be used for processing large sequencing data sets. As an alternative, we develop an EM parameter estimation scheme which relies on the proposed HMM and uses samples generated by a particle filter to evaluate the expectation of the likelihood function. We refer to this algorithm as the particle filter EM (PFEM). The speed and accuracy of the proposed scheme is practically sound for use in next generation sequencing platforms.

Assumptions on parameters

Recall that the set of parameters needed to run ParticleCall is Λ={p qd1:Lα1:Lσ1:LK1:LΣ1:L}. The phasing and prephasing parameters p and q are assumed to be the same for each sequencing lane and are estimated using the same procedure as Bustard (see, e.g., [8]). The remaining parameters are assumed to be cycle-dependent and need to be estimated for each tile. The cycle-dependency assumption on the parameters can lead to a substantial improvement in the base-calling accuracy [5]. In order to avoid over-fitting, we assume that parameters remain constant within a short window of cycles and then change to a different set of values. To track the changes in the parameters, we first divide the total read length L into several non-overlapping windows and then perform our parameter estimation window-by-window. To further reduce the number of parameters and improve the estimation efficiency, we assume that the parameters d1:L and σ1:L are uniformly distributed over an interval and incorporate them into the hidden states of the HMM model. Therefore, only the mean and variance of these parameters, i.e., d mean , d var , σ mean , and σ var need to be estimated. Computational results demonstrate that these two assumptions does not affect the accuracy of base-calling.

Particle filter EM algorithm

In the early sequencing cycles, effects of phasing and prephasing are relatively small. Therefore, we may ignore phasing and prephasing to facilitate straight-forward computation of the initial estimates of the remaining parameters. In particular, the signal generated in the early cycles t is approximated as
x t = λ t s t . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ19_HTML.gif
(19)

Replacing (2) by (10) leads to a simplified model that allows for straightforward base calling and inference of the parameters by means of linear regression. We use these values to obtain the estimates of d mean , d var , σ mean , and σ var , and to initialize the remaining parameters α, K, Σ, in the particle filter EM parameter estimation procedure.

The parameter estimation is performed window-by-window and is conducted using n reads randomly chosen from a tile (in our experiments, we use n=200). Assume the window length is w, and denote the window index by m. The particle filter EM (PFEM) algorithm finds parameters for one window and then uses these values to initialize the search for parameters in the next window. We illustrate the procedure for the first window here (the same procedure is repeated in the following windows). Let Λ 1 i = { α i , K i , Σ i } https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq69_HTML.gif denote the set of parameters for window 1 in the ith iteration of the EM scheme. The estimate of Λ 1 i https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq70_HTML.gif is given by
Λ 1 i = arg max Λ 1 L 1 Λ 1 i 1 , https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ20_HTML.gif
(20)
where L 1 ( Λ 1 i 1 ) = j = 1 n L 1 , j ( Λ 1 i 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq71_HTML.gif is the sum of the log-likelihood functions over the reads in the training set. The log-likelihood function for each read, L 1 , j ( Λ 1 i 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq72_HTML.gif, is obtained as
L 1 , j ( Λ 1 i 1 ) = log P ( y 1 : w | Λ 1 i 1 ) = E log P ( y 1 : w , s 1 : w , λ 1 : w | Λ 1 i 1 ) , https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ21_HTML.gif
(21)

where the expectation is taken with respect to P ( s 1 : w , λ 1 : w | y 1 : w , Λ 1 i 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq73_HTML.gif. We rely on an SISR particle filtering scheme to generate equally weighted sample trajectories from P ( s 1 : w , λ 1 : w | y 1 : w , Λ 1 i 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq74_HTML.gif. Based on (7) and (8), we calculate log P ( y 1 : w , s 1 : w , λ 1 : w | Λ 1 i 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq75_HTML.gif for these samples and compute their average to approximate the expectation in (12). The maximization (11) is performed by solving equations obtained after taking gradients of L 1 ( Λ 1 i 1 ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq76_HTML.gif over the parameters and setting them to 0. In our experiment, the PFEM parameter estimation scheme performs 30 EM iterations and uses 600 samples from the particle filter for each window.

Results and discussion

The proposed method is evaluated on a data set obtained by sequencing phiX174 bacteriophage using Illumina Genome Analyzer II with the cycle length 76. This is a short genome with a known sequence which enables reliable performance comparison of different base-calling techniques. We tested ParticleCall and several other algorithms on a tile containing 77337 reads, and present the results here. All the codes are written in C and the tests are run on a desktop with an Intel Core i7 4-core 3GHz processor.
https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Fig2_HTML.jpg
Figure 2

Per-cycle error rates of ParticleCall, BayesCall, naiveBayesCall, Rolexa and Bustard. The figure compares the per-cycle error rates of different base-calling algorithms. ParticleCall and BayesCall are the most accurate ones.

Performance of ParticleCall

The base calling error rates are computed by aligning the reads to the reference genome and evaluating frequency of mismatches. Reads that could not be aligned to the reference with at least 70% matches are discarded. Note that the error rates and speed of the proposed ParticleCall algorithm and the parameter estimation scheme are affected by the parameters l, r, particle number N p , and parameter estimation window length w. We ran ParticleCall with l=r{1,2,4}. Increasing l and r beyond l=r=1 did not affect the performance while it significantly slowed down the algorithm. This is due to small values of the phasing and prephasing probabilities, which are estimated to be p=3.54×10−8 and q=0.00335. Therefore, in the remainder of the paper, we set l=r=1. The accuracy of base-calling for different N p is shown in Table 1. As seen there, for the original ParticleCall algorithm, N p =800 leads to high performance with reasonable speed. Rao-Blackwellized ParticleCall can achieve the same accuracy with fewer particles (in particular, N p =300); however, its effective running time is 3 times that of the original ParticleCall with the same performance. This is because the Rao-Blackwellization steps in (9) and (9) require evaluating a sum over all possible S t w https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq77_HTML.gif (43=64 for our choice l=r=1), resulting in a fairly large number of basic operations needed to calculate exact distribution of the discrete variables. Therefore, for further performance comparisons, we rely on the original ParticleCall algorithm (formalized as Algorithm 1). Table 2 shows the ParticleCall base calling error rate and parameter estimation times for different window lengths w. In the remainder of the paper, we set w=5 as it leads to desirable performance/speed characteristics of the algorithm.
Table 1

Comparison of ParticleCall with different N p

Method

N p

error rate

base-calling time (min)

ParticleCall (via MCEM)

400

0.0126

46

 

800

0.0124

88

 

1200

0.0124

130

ParticleCall (via PFEM)

400

0.0128

46

 

800

0.0125

91

 

1200

0.0125

133

Rao-Blackwellized ParticleCall (via MCEM)

100

0.0128

103

 

200

0.0125

190

 

300

0.0124

287

 

400

0.0124

386

ParticleCall is run using parameters obtained via the MCEM parameter estimation scheme as well as via the PFEM parameter estimation algorithm proposed in this paper. Rao-Blackwellized ParticleCall is run using parameters via the MCEM parameter estimation scheme.

Table 2

ParticleCall parameter estimation

  

parameter estimation

Window length w

base-calling error rate

time (min)

4

0.0125

50

5

0.0125

39

6

0.0127

29

7

0.0130

25

ParticleCall base-calling error rate and the parameter estimation time of the proposed PFEM parameter estimation algorithm.

Performance comparison of different algorithms

The error rates and speed of the proposed ParticleCall algorithm are compared with those of BayesCall, naiveBayesCall, Rolexa, and Bustard. We run ParticleCall both with parameters provided by the computationally intensive MCEM algorithm as well as with those inferred by the PFEM parameter estimation scheme proposed in this paper. The results are reported in Table 3. Note that Rolexa generally outputs the so-called IUPAC codes, unlike all the other considered algorithms which provide sequences of nucleotides A, C, G, and T. To allow a comparison, we enforce Rolexa to output sequences of nucleotides as well. The comparison of per-cycle error rates is shown in Figure 2. It can be seen from Table 3 and Figure 2 that ParticleCall, BayesCall and naiveBayesCall all have improved base-calling accuracy compared to Bustard. BayesCall is highly accurate but relatively slow – it requires approximately 4 hours to complete base-calling for one tile of the data. naiveBayesCall significantly improves base-calling speed over BayesCall but it does so at the expense of incurring higher error rate. Our ParticleCall base-calling algorithm has the same accuracy as BayesCall, while being roughly 3 times faster. Figure 2 shows that both ParticleCall and BayesCall are more accurate than naiveBayesCall in the early cycles and improve over Bustard in all cycles. Note that Bustard outperforms Rolexa, which is consistent with the results in [5]. Moreover, we see from Table 3 that performing parameter estimation via the MCEM algorithm proposed in [8] requires 19 hours, while the particle filter implementation of the EM estimation scheme proposed in this paper takes only 39 minutes. As evident from Table 3, running ParticleCall with parameters obtained by the PFEM scheme leads to only a minor performance degradation compared to running it with parameters obtained by the MCEM algorithm. Running ParticleCall base calling along with the PFEM parameter estimation scheme takes about 2 hours per tile, which is 9 times faster than the total time required by the less accurate naiveBayesCall.
Table 3

Comparison of error rates and speed

  

base-calling

parameter estimation

Method

error rate

time (min)

time (min)

Bustard

0.0152

2 (total)

 

Rolexa

0.0170

35 (total)

 

naiveBayesCall

0.0132

21

1139

BayesCall

0.0124

231

1139

ParticleCall

   

(via MCEM)

0.0124

88

1139

ParticleCall

   

(via PFEM)

0.0125

91

39

The base-calling error rate and the running times of different algorithms. ParticleCall is run using parameters obtained via the MCEM parameter estimation scheme as well as via the PFEM parameter estimation algorithm proposed in this paper. For Bustard and Rolexa, only the total running times are reported.

Quality scores

Quality scores are used to characterize confidence in the outcome of the base-calling procedures. They are computed as part of the analysis of the acquired raw data and may be used to filter out reads of suspect quality, or to shorten the reads if the quality scores of individual bases fall below certain thresholds. They can also provide confidence information for downstream analysis including sequence assembly and SNP and genotype calling. Frequently used are the so-called phred quality scores, which were originally developed to assess the quality of the conventional Sanger sequencing and automate large-scale sequencing projects. Phred scores are also often provided by the algorithms used for base-calling in next generation sequencing platforms. Formally, the phred score for a called base in the cycle t, s ̂ t https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq78_HTML.gif, is defined as
Q phred ( s ̂ t ) = 1 0 log 1 0 P ( s ̂ t s t ) . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ22_HTML.gif
(22)
Essentially, Q phred ( s ̂ t ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq79_HTML.gif is the scaled logarithm of the error probability. Higher quality scores imply smaller probability of the base-calling error. For the proposed ParticleCall algorithm, probability of correctly calling a base can be obtained from the posteriori probability as
P ( s ̂ t s t ) = 1 p s t | y 1 : t . https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Equ23_HTML.gif
(23)
Quality scores can be used to compare the discrimination ability of different algorithms. The discrimination score D(ε) at error tolerance ε is defined as the ratio of the correctly called bases having P ( s ̂ t s t ) < https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq80_HTML.gif (i.e., the quality score higher than 1 0 log 1 0 ( ) https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_IEq81_HTML.gif) to all called bases. Figure 3 compares the discrimination ability of ParticleCall, BayesCall, naiveBayesCall and Bustard. It shows that for a reasonable error tolerance ε, ParticleCall with parameters obtained through MCEM has better discrimination ability than BayesCall, naiveBayesCall and Bustard, while ParticleCall with parameters obtained through PFEM has discrimination ability close to naiveBayesCall and better than other algorithms. In other words, when a small cutoff error tolerance ε is set and all the bases with quality scores below εare considered invalid, ParticleCall provides the most accurate results among the considered base-calling schemes.
https://static-content.springer.com/image/art%3A10.1186%2F1471-2105-13-160/MediaObjects/12859_2012_Article_5277_Fig3_HTML.jpg
Figure 3

Discrimination ability D(ε ) of quality scores vs error tolerance. The figure shows the percentage of correctly called bases under different error tolerance ε.

Effects of improved base-calling accuracy on de novo sequence assembly

In shotgun sequencing, a long target sequence is oversampled by a library of randomly fragmented copies of the target, and the overlaps between short reads obtained by a high-throughput platform are used to assemble the target. In de novo assembly, the target is reconstructed without consulting any reference [21, 22]. Performance of assembly algorithms highly depends on the accuracy of base calling. To demonstrate the effects of base-calling accuracy on assembly, we apply the Velvet assembly algorithm [22] on reads provided by Bustard, Rolexa, naiveBayesCall, BayesCall, and ParticleCall. In particular, we randomly subsample the set of reads provided by each of the base calling algorithms to emulate 5X, 10X, 15X, and 20X coverage. Then we run Velvet on each of the subsets, and evaluate commonly used metrics that quantify the quality of the assembly procedure. Specifically, we evaluate the maximum contig length and the N50 contig length. The described procedure is repeated 200 times to obtain average values of these two quality metrics. The results are shown in Table 4. As can be seen there, ParticleCall provides the largest N50 and maximum contig length among all of the considered base calling schemes, for all of the considered coverages.
Table 4

de novo assembly results

         

ParticleCall

ParticleCall

Coverage

Bustard

Rolexa

naiveBayesCall

BayesCall

via MCEM

via PFEM

 

N50

Max

N50

Max

N50

Max

N50

Max

N50

Max

N50

Max

5X

271

607

259

565

278

604

292

629

299

637

289

632

10X

1169

1750

971

1557

1180

1731

1269

1831

1316

1900

1341

1865

15X

3624

3823

2885

3170

3726

3908

3466

3741

3742

3935

3697

3918

20X

4694

4744

4529

4614

4756

4816

4827

4875

5102

5116

4795

5039

The maximum contig length and N50 length of de novo assembly using Velvet. The average values over 200 experiments are shown in the table.

Conclusions

In this paper we presented ParticleCall, a particle filtering algorithm for base calling in the Illumina’s sequencing-by-synthesis platform. The algorithm is developed by relying on an HMM representation of the sequencing process. Experimental results demonstrate that the ParticleCall base calling algorithm is more accurate than Bustard, Rolexa, and naiveBayesCall. It is as accurate as BayesCall while being significantly faster. Quality score analysis of the reads indicates that ParticleCall has better discrimination ability than BayesCall, naiveBayesCall and Bustard. Moreover, a novel particle filter EM (PFEM) parameter estimation scheme, much faster than the existing Monte Carlo implementation of the EM algorithm, was proposed. When relying on the PFEM scheme, ParticleCall has near-optimal performance while needing much shorter total parameter estimation and base calling time.

Author’s contributions

Algorithms and experiments were designed by Xiaohu Shen (XS) and Haris Vikalo (HV). Algorithm code was implemented and tested by XS. The manuscript was written by XS and HV. Both authors read and approved the final manuscript.

Declarations

Acknowledgements

This work was funded by the National Institute of Health under grant 1R21HG006171-01.

Authors’ Affiliations

(1)
Department of Electrical and Computer Engineering, University of Texas at Austin, 1 University Station C0803

References

  1. Shendure J, Ji H: Next-generation DNA sequencing. Nat Biotechnology. 2008, 26: 1135-10.1038/nbt1486.View Article
  2. Metzker M: Emerging technologies in DNA sequencing. Genome Research. 2005, 56: 1767-View Article
  3. Bentley D: Whole-genome re-sequencing. Curr Opin Genet Dev. 2006, 16: 545-10.1016/j.gde.2006.10.009.View ArticlePubMed
  4. Nielsen R, Paul JS, Alvrechtsen A, Song YS: Genotype and SNP calling from next-generation sequencing data. Nature Reviews. 2011, 12: 443-10.1038/nrg2986.PubMed CentralView ArticlePubMed
  5. Ledergerber C, Dessimoz C: Base-calling for next-generation sequencing platforms. Briefings in Bioinformatics. 2011, 12: 489-10.1093/bib/bbq077.PubMed CentralView ArticlePubMed
  6. Rougemont J, Amzallag A, Iseli C, Farinelli L, Xenarios I, Naef F: Probabilistic base calling of solexa sequencing data. BMC Bioinformatics. 2008, 9: 431-10.1186/1471-2105-9-431.PubMed CentralView ArticlePubMed
  7. Erlich Y, Mitra P, Delabastide M, McCombie W, Hannon G: Alta-Cyclic: a self-optimizing base caller for next-generation sequencing. Nat Methods. 2008, 5: 679-10.1038/nmeth.1230.PubMed CentralView ArticlePubMed
  8. Kao W, Stevens K, Song Y: BayesCall: A model-based base-calling algorithm for high-throughput short-read sequencing. Genome Research. 2009, 19: 1884-10.1101/gr.095299.109.PubMed CentralView ArticlePubMed
  9. Kao W, Stevens K, Song Y: naiveBayesCall: an efficient model-based base-calling algorithm for high-throughput sequencing. Journal of Computational Biology. 2011, 18: 365-10.1089/cmb.2010.0247.View ArticlePubMed
  10. Fedurco M, Romieu A, Williams S, et al: BTA, a novel reagent for DNA attachment on glass and efficient generation of solid-phase amplified DNA colonies. Nucleic Acids Res. 2006, 34 (3): e22-10.1093/nar/gnj023.PubMed CentralView ArticlePubMed
  11. Turcatti G, Romieu A, Fedurce M, et al: A new class of cleavable fluorescent nucleotides: synthesis and optimization as reversible terminators for DNA sequencing by synthesis. Nucleic Acids Res. 2008, 36 (4): e25-PubMed CentralView ArticlePubMed
  12. Eddy S: Hidden Markov models. Current Opinion in Structural Biology. 1996, 6 (3): 361-10.1016/S0959-440X(96)80056-X.View ArticlePubMed
  13. Doucet A, Wang X: Monte Carlo methods for signal processing: A review in the statistical signal processing context. IEEE Signal Processing Magzine. 2005, 22: 152-View Article
  14. Cappé O, Moulines E, Rydén T: Inference in hidden Markov models. 2005, Springer Verlag, New York
  15. Pitt M, Shephard N: Filtering via simulation: Auxiliary particle filters. Journal of the American Statistical Association. 1999, 94: 590-10.1080/01621459.1999.10474153.View Article
  16. Doucet A, Godsill S, Andrieu C: On sequential Monte Carlo sampling methods for Bayesian filtering. Statistics and computing. 2000, 10 (3): 197-10.1023/A:1008935410038.View Article
  17. Kim S, Shephard N, Chib S: Stochastic volatility: likelihood inference and comparison with ARCH models. The Review of Economic Studies. 1998, 65 (3): 361-10.1111/1467-937X.00050.View Article
  18. Liu J, Chen R: Sequential Monte Carlo methods for dynamic systems. Journal of the American statistical association. 1998, 93: 1032-10.1080/01621459.1998.10473765.View Article
  19. Kitagawa G: Monte Carlo filter and smoother for non-Gaussian nonlinear state space models. Journal of computational and graphical statistics. 1996, 5: 1-
  20. Carpenter J, Clifford P, Fearnhead P: Improved particle filter for nonlinear problems. Radar, Sonar and Navigation, IEE Proceedings-, Volume 146, IET. 1999, , , 2-7.
  21. Butler J, MacCallum I, Kleber M, Shlyakhter I, Belmonte M, Lander E, Nusbaum C, Jaffe D: ALLPATHS: de novo assembly of whole-genome shotgun microreads. Genome Research. 2008, 18 (5): 810-10.1101/gr.7337908.PubMed CentralView ArticlePubMed
  22. Zerbino D, Birney E: Velvet: algorithms for de novo short read assembly using de Bruijn graphs. Genome research. 2008, 18 (5): 821-10.1101/gr.074492.107.PubMed CentralView ArticlePubMed

Copyright

© Shen and Vikalo; licensee BioMed Central Ltd. 2012

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://​creativecommons.​org/​licenses/​by/​2.​0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.