 Methodology article
 Open Access
 Published:
Reconstructing DNA copy number by joint segmentation of multiple sequences
BMC Bioinformatics volume 13, Article number: 205 (2012)
Abstract
Background
Variations in DNA copy number carry information on the modalities of genome evolution and misregulation of DNA replication in cancer cells. Their study can help localize tumor suppressor genes, distinguish different populations of cancerous cells, and identify genomic variations responsible for disease phenotypes. A number of different high throughput technologies can be used to identify copy number variable sites, and the literature documents multiple effective algorithms. We focus here on the specific problem of detecting regions where variation in copy number is relatively common in the sample at hand. This problem encompasses the cases of copy number polymorphisms, related samples, technical replicates, and cancerous subpopulations from the same individual.
Results
We present a segmentation method named generalized fused lasso (GFL) to reconstruct copy number variant regions. GFL is based on penalized estimation and is capable of processing multiple signals jointly. Our approach is computationally very attractive and leads to sensitivity and specificity levels comparable to those of stateoftheart specialized methodologies. We illustrate its applicability with simulated and real data sets.
Conclusions
The flexibility of our framework makes it applicable to data obtained with a wide range of technology. Its versatility and speed make GFL particularly useful in the initial screening stages of large data sets.
Background
Genomic duplications and deletions are common in cancer cells and known to play a role in tumor progression [1]. As our ability to survey the fine scale of the human genome has increased, it has become apparent that normal cells can also harbor a number of variations in copy number (CN) [2, 3]. The last few years have witnessed a steady increase in our knowledge of the size and frequency of these variants [4–7] and their implications in complex diseases [8, 9]. At the same time, statistical methods and algorithms have been developed to better harness the information available. At the cost of oversimplification, two different approaches have become particularly popular. One is based on the hidden Markov model (HMM) machinery and explicitly aims to reconstruct the unobservable discrete DNA copy number; the other, which we will generically call “segmentation”, aims at identifying portions of the genome that have constant copy number, without specifically reconstructing it.
The HMM approach takes advantage of the implicitly discrete nature of the copy number process (both when a finite number of states is assumed and when, as in some implementations, less parametric approaches are adopted); furthermore, by careful modeling of the emission probabilities, one can fully utilize the information derived from the experimental results. In the case of genotyping arrays, for example, quantification of total DNA amount, relative allelic abundance, and prior information such as minor allele frequencies can be considered.
No apriori knowledge of the number of copy number states is required in the segmentation approach—an advantage in the study of cancer where polyploidy and contamination with normal tissues result in a wide range of fractional copy numbers. Possibly for the reasons outlined, HMMs are the methods of choice in the analysis of normal samples [10–14], while segmentation methods are the standard in cancer studies [15, 16]. A limitation of segmentation methods is that they rely on data in which the variation in copy number is reflected in the differences in means of the segments. This fact makes segmentation methods applicable directly to a substantial portion of the data derived from recent technologies, but not to relative allelic abundance. However, see the modification suggested in [17] and the following description for an exception.
While a number of successful approaches have been derived along the lines described above, there is still a paucity of methodology for the joint analysis of multiple sequences. It is clear that if multiple subjects share the same variation in copy number, there exists the potential to increase power by joint analysis. Wang et al. [18] present a methodology that extends [1] to reconstruct the location of tumor suppressor genes from the identification of regions lost in a larger number of samples. The initial steps of the Birdsuite algorithm rely on the identification of suspect signals in the context of multiple sample. PennCNV [13] includes an option of joint analysis of trios. Methodology to process multiple samples within the context of change point analysis has been developed in [16, 19–21]. Efron and Zhang [22] consider FDR analysis of independent samples to identify copy number polymorphisms (CNPs), and Nowak et al. [23] use a latent feature model to capture, in joint analysis of arrayCGH data from multiple tumor samples, shared copy number profiles, on each of which a fusedlasso penalty is enforced for sparsity.
In the present work we consider a setting similar to [16] in that we want joint analysis to inform the segmentation of multiple samples. Our main focus is the analysis of genotyping array data, but the methodology we develop is applicable to a variety of platforms. By adopting a flexible framework we are able, for example, to define a segmentation algorithm that uses all information from Illumina genotyping data. As in [19], we are interested in the situation where some but not all the samples under consideration carry a copy number variant (CNV). We prefer to enforce a certain sparsity in the vector that identifies which samples carry a given variant. We tackle this problem using a penalized estimation approach, originally proposed in this context by [24], for which we have developed an algorithmic implementation [25]. Appreciable results are achieved in terms of speed, accuracy, and flexibility.
In concluding this introduction, we would like to make an important qualification. The focus of our contribution is on segmentation methods, knowing that this is only one of the steps necessary for an effective recovery of CNVs. In particular, normalization and transformation of the signal from experimental sources are crucial and can have a very substantial impact on final results, as documented in [26–31], for example. Indeed, preprocessing to eliminate systematic variation in intensities is particularly important for joint analysis of multiple sequences, when repeated deviances are more likely to be interpreted as true signal. Furthermore, calling procedures that classify results of segmentation while possibly controlling global error measures [22] are also needed. Indeed, in the data analysis included in this paper, we need to resort to both of these additional steps, and we will describe then briefly the fairly standard choices we make.
Before describing in detail the proposed methods for joint segmentation of multiple sequences, we start by illustrating various contexts where joint analysis appears to be useful.
Genotyping arrays and CNV detection
Genotyping arrays have been used on hundreds of thousands of subjects. The data collected through them provides an extraordinary resource for CNV detection and the study of their frequencies in multiple populations. Typically, the raw intensity data (representing hybridization strength) is processed to obtain two signals: quantification of total DNA amount (from now on log R Ratio, LRR, following Illumina terminology) and relative abundance of the two queried alleles (from now on B allele frequency, BAF). Both these signals contain information on CNVs, and one of the strengths of HMMs has been that they can easily process them jointly. Segmentation models like CBS have traditionally relied only on LRR. While this is a reasonable choice, it can lead to substantial loss of information, particularly in tumor cells, where polyploidy and contamination make information in LRR hard to decipher. To exploit BAF in the context of a segmentation method, a signal transformation has been suggested [17]: mirrored BAF (mBAF) relies on exchangeability of the two alleles and the low information content of homozygous SNPs. The resulting mBAF is defined on a coarser grid than the original BAF, but is characterized by changing means in the presence of a CNV. While [17] shows that the analysis of BAF alone can be advantageous and more powerful than segmentation of LRR in some contexts, clearly a joint analysis of LRR and mBAF should be preferable to an arbitrary selection of one or the other signal.
Multiple platforms
LRR and BAF are just one example of the multiple signals available in some samples. Often, as research progresses, the samples are assessed with a variety of technologies. For example, a number of subjects who have been genotyped at high resolution are now being resequenced. Whenever the technology adopted generates a signal that contains some information on copy number, there is an incentive to analyze the available signals jointly.
Tumor samples from the same patient obtained at different sites or different progression stages
In an effort to identify mutations that are driving a specific tumor, as well as study its response to treatment, researchers might want to study CNVs in cells obtained at different tumor sites or at different time points [32]. Copy number is highly dynamic in cancer cells, so that it is to be expected that some differences will be detected over time or across sites. By contrast, the presence of the same CNVs across these samples can be taken as an indication that the tumors share the same origin. Therefore, a comparative analysis of CNV can be used to distinguish resurgence of the same cancer from insurgence of a new one, or to identify specific cancer cell populations. Given that the tissue extracted always consists of a mixture of normal and cancer cells, which are in turn a mixture of different populations, joint analysis of the signals from the varied materials is much more likely to lead to the identification of common CNVs when these exist.
Related subjects
Family data is crucial in genetic investigations, and hence it is common to analyze related subjects. When studying individuals from the same pedigree, it is reasonable to assume that some CNVs might be segregating in multiple people and that joint analysis would reduce Mendelian errors and increase the power of detection.
The rest of the paper is organized as follows: In the Methods section, we first present the penalized estimation framework, and then describe how the model can be used for data analysis by: (a) outlining an efficient estimation algorithm, (b) generalizing it to the case of uncoordinated data, and (c) describing the choice of the penalization parameters. In the results section, we discuss our findings on two simulated data sets (descriptive of normal and tumor samples) and two real data sets. In one case multiple platforms are used to analyze the same sample, and in the other case samples from related individuals benefit from joint analysis.
Methods
A model for joint analysis of multiple signals
Assume we have observed M signals, each measured at N locations, corresponding to ordered physical positions along the genome, with y_{ i j } being the observed value of sequence i at location j. The copy number process can be modeled as
where ε_{ i j } represent noise, and the mean values β_{ i j } are piecewise constant. Thus, there exists a linearly ordered partition $\{{R}_{1}^{\left(i\right)},{R}_{2}^{\left(i\right)},\dots ,{R}_{{K}_{i}}^{\left(i\right)}\}$ of the location index {1,2,…,N} such that ${\beta}_{\mathrm{is}}=\xb7\; \xb7\; \xb7={\beta}_{\mathrm{it}}={\mu}_{k}^{\left(i\right)}$ for $s,\dots ,t\in {R}_{k}^{\left(i\right)}$ and 1 ≤ k ≤ K_{ i }. In other words, most of the increments β_{ i j } − β_{i,j−1}  are assumed to be zero. When two sequences k and l share a CNV with the same boundaries at location j, both β_{ k j } − β_{k,j−1}  and β_{ l j } − β_{l,j−1}  will be different from zero at the change point j. Modulo an appropriate signal normalization, β_{ i j } = 0 can be interpreted as corresponding to the appropriate normal copy number equal to 2. We propose to reconstruct the mean values β by minimizing the following function, called hereafter the generalized fused lasso (GFL):
which includes a goodnessoffit term and three penalties, whose roles we will explain one at a time. The ℓ_{1} penalty $\sum _{i=1}^{M}\sum _{j=1}^{N}\left{\beta}_{\mathit{ij}}\right$ enforces sparsity within β, in favor of values β_{ i j } = 0, corresponding to the normal copy number. The total variation penalty $\sum _{j=2}^{N}{\beta}_{\mathit{ij}}{\beta}_{i,j1}$ minimizes the number of jumps in the piecewise constant means of each sequence and was introduced by [24] in the context of CNV reconstruction from arrayCGH data. Finally, the Euclidean penalty on the column vector of jumps $\sqrt{\sum _{i=1}^{M}{({\beta}_{\mathit{ij}}{\beta}_{i,j1})}^{2}}$ is a form of the group penalty introduced by [33] and favors common jumps across sequences. As clearly explained in [34], “the local penalty around 0 for each member of a group relaxes as soon as the β_{ i j } − β_{i,j−1}  for one member i of the group moves off 0.” Bleakley and Vert (2011) [35] also suggested the use of this groupfusedlasso penalty to reconstruct CNV. We consider here the use of both the total variation and the Euclidean penalty on the jumps to achieve the equivalent effect of the sparse group lasso, which, as pointed out in [36], favors CNV detection in multiple samples, allowing for sparsity in the vector indicating which subjects are carriers of the variant. This property is important in situations of multiple tumor samples and related subjects, where one does not want to assume that all the M sequences carry the same CNV.
The incorporation of the latter two penalties can also be naturally interpreted in view of image denoising. To restore an image disturbed by random noise while preserving sharp edges of items in the image, a 2D total variation penalty $\lambda \sum _{i=1}^{M}\sum _{j=2}^{N}{\beta}_{\mathit{ij}}{\beta}_{i,j1}+\rho \sum _{j=1}^{N}\sum _{i=2}^{M}{\beta}_{\mathit{ij}}{\beta}_{i1,j}$ is proposed in a regularized leastsquare model [37], where β_{ i j } is the true underlying intensity of pixel (i j). In CNV detection problems, signals from multiple sequences can be aligned in the shape of an image, except that pixels in each sequence are linearly ordered while sequences as a group have no certain order a priori. Thus, one of the two total variation penalties is replaced by the group penalty on the column vector of jumps.
Using matrix notation, and allowing the tuning parameter λ_{1}, λ_{2} and λ_{3} to be sequence specific, we can reformulate the objective function as follows. Let Y = (y_{ i j })_{M×N} and β = (β_{ i j })_{M×N}. Let β_{ i } be the i th row of β and β_{(j)} the j th column of β. Also, let λ_{3} = (λ_{3,i})_{M×1}. Then we have
where ·_{ F } is the Frobenius matrix norm, $\left\right\xb7{}_{{\ell}_{1}}$ and $\left\right\xb7{}_{{\ell}_{2}}$ are the ℓ_{1} and ℓ_{2} vector norms, β_{i,s:t} indicates the subvector with elements β_{i,s},…,β_{i,t} of the row vector β_{ i }, and “∗” denotes entrywise multiplication between two vectors. It would be easy to modify the tuning parameters so as to make them location specific by reducing the penalty for a jump in genomic regions known to harbor CNVs.
An MM algorithm
While the solution to the optimization problem (3) might have interesting properties, this approach is useful only if an effective algorithm is available. The last few years have witnessed substantial advances in computational methods for ℓ_{1} regularization problems, including the use of coordinate descent [38, 39] and path following methods [35, 40–42]. The computational complexity of these methods in the best situation is O(MNK), where k indicated the number of knots along the solution path. Here knots are conjunction points between a series of piecewise functions of tuning parameters. It is important to note that these algorithms – some of which are designed for more general applications – may not be the most efficient for large scale CNV analysis for at least two reasons. On the one hand, reasonable choices of λ might be available, making it unnecessary to solve for the entire path; on the other hand, the number of knots k can be expected to be as large as O(N), making the computational costs of path algorithms prohibitive.
With specific regard to the fusedlasso application to CNV detection, we were successful in developing an algorithm with per iteration cost O(N) and empirically fast convergence rate for the analysis of one sequence [25]. We apply the same principles here and start by modifying the norms in the penalty to achieve better computational stability. For the ℓ_{1} norm we substitute $\left\rightx{}_{2,\epsilon}=\sqrt{{x}^{2}+\epsilon}$ for sufficiently small ε, and for the ℓ_{2} norm we substitute $\left\right\mathbf{x}{}_{2,\epsilon}={\left(\sum _{i=1}^{n}{x}_{i}^{2}+\epsilon \right)}^{\frac{1}{2}}$. This produces the differentiable objective function
Adopting an MM framework [43], we want to find a surrogate function ${\u210a}_{\epsilon}(\mathit{\beta}\mid {\mathit{\beta}}^{\left(m\right)})$ for each iteration M such that ${\u210a}_{\epsilon}({\mathit{\beta}}^{\left(m\right)}\mid {\mathit{\beta}}^{\left(m\right)})={f}_{\epsilon}\left({\mathit{\beta}}^{\left(m\right)}\right)$ and ${\u210a}_{\epsilon}(\mathit{\beta}\mid {\mathit{\beta}}^{\left(m\right)})\ge {f}_{\epsilon}\left(\mathit{\beta}\right)$ for all β. At each iteration, the MM principle chooses ${\mathit{\beta}}^{(m+1)}=\text{arg min\hspace{0.17em}}{\u210a}_{\epsilon}(\mathit{\beta}\mid {\mathit{\beta}}^{\left(m\right)})$. A majorizing function with the above properties is readily obtained using the concavity of the squareroot function $\left\rightx{}_{2,\epsilon}\le \frac{1}{2\left\rightz{}_{2,\epsilon}}({x}^{2}{z}^{2})$, and its vector equivalent $\left\right\mathbf{x}{}_{2,\epsilon}\le \frac{1}{2\left\right\mathbf{z}{}_{2,\epsilon}}(\left\right\mathbf{x}{}_{{\ell}_{2}}^{2}\left\mathbf{z}\right{}_{{\ell}_{2}}^{2})$. The resulting surrogate function
separates as a sum of similar functions in the the row vectors β_{ i } ; namely,
where
Here each ${\mathbf{A}}_{i}^{\left(m\right)}$ is a tridiagonal symmetric matrix, and ${\stackrel{~}{c}}_{i}^{\left(m\right)}$ is a constant, irrelevant for optimization purposes. In view of the strict convexity of the surrogate function, each ${\mathbf{A}}_{i}^{\left(m\right)}$ is also positive definite. The nonzero entries of ${\mathbf{A}}_{i}^{\left(m\right)}$ and ${\mathbf{b}}_{i}^{\left(m\right)}$ (i = 1,…,M) are listed in Additional file 1: Supplementary Text . Each of the surrogate functions in (5) can be minimized by solving the linear system ${\mathit{\beta}}_{i}={\left[{\mathit{\beta}}_{i}^{\left(m\right)}\right]}^{T}{\left[{\mathbf{A}}_{i}^{\left(m\right)}\right]}^{1}$ by the Tridiagonal Matrix (TDM) algorithm [44]. This results in a periteration computational cost of O(MN). This algorithm is empirically observed to achieve an exponential convergence rate [25], although we do not yet have an analytic proof. In practice, this method scales well with joint analysis of tens to hundreds of samples with measurements at millions of locations, with limitations dictated by memory requirements. For analysis of real data, we suggest one or a group of samples to be analyzed chromosome by chromosome, since a CNV region can never extend beyond one chromosome to another. Actual computation times are shown along with different examples in the results section. Readers might be interested in comparing the outlined approach with other segmentation methods that are not based on the use of ℓ_{1} penalties, as [45].
Stacking observations at different genomic locations
While copy number is continuously defined across the genome, experimental procedures record data at discrete positions, for which we have used the indexes j=1,…,N. In reality, repeated evaluations of the same sample (or related samples) will typically result in measurements at only partially overlapping genomic locations, either because different platforms use different sets of probes or missing data occur at different positions across the sequences. For example, for mBAF and LRR measurements from the same experiment on the same subject, the mBAF signal is defined on a subset of the locations where the LRR signal is defined.
Let S indicate the union of all genomic positions where some measurement is available among the M signals under study. And let S_{ i } be the subset of locations with measurements in sequence i. We reconstruct β_{ i j } for all j ∈ S. When j ∉ S_{ i }, β_{ i j } will be determined simply on the basis of the neighboring data points, relying on the regularizations introduced in (3). The goodnessoffit portion of the objective function is therefore redefined as
The MM strategy previously described applies with slight modifications of the matrix ${\mathbf{A}}_{i}^{\left(m\right)}$ (see Additional file 1: Supplementary Text).
The attentive reader will have noted that the y_{ i j } values with j ∉ S_{ i } can be considered missing data, and evaluation of the missingness pattern is appropriate. In general, the y_{ i j } cannot be considered missing at random. The most important example is the case of mBAF, where homozygous markers result in missing values. Homozygosity is clearly more common when copy number is equal to 1 than when copy number is equal to 2. Therefore, there is potentially more information on β_{ i j } to be extracted from the signals than what we capture with the proposed method. Although most of the information on deletions is obtained through LRR, BAF does convey additional information on duplications, where the changes in LRR are limited by saturation effects. On the other hand, it does appear that our method does not increase the rate of false positives. Hence, it can be considered as an operational improvement over segmentation based on LRR only, even if in theory, it does not completely use the information on BAF.
Choice of tuning constants and segmentation
One of the limitations of penalization procedures is that values for the tuning parameters need to be set, and clear guidelines are not always available. Path methods that obtain a solution of the optimization problem (3) for every value of a tuning parameter can be attractive, but recent algorithmic advances [35, 41, 42] remain impractical for problems of the size of ours. A number of recent publications obtain optimal values of penalty parameters under a series of conditions [46–49]. We rely on these findings to propose a strategy for obtaining a solution of (3) for reasonably liberal values of the tuning parameters, followed by a sequencebysequence hard thresholding of the detected jumps with a dataadaptive threshold.
We have found the following guidelines to be useful in choosing penalty parameter values:
for i = 1,…,M, where ${\widehat{\sigma}}_{i}$ is a robust estimate of standard deviation of y_{ i }, p is roughly the proportion of the M sequences we anticipate to carry CNVs, and c_{1}, c_{2} and c_{3} are positive multipliers adjusted to account for different signaltonoise ratios and CNV sizes. We discuss the function ρ below.
While a more rigorous justification is provided in the Additional file 1: Supplementary Text, we start by underscoring some of the characteristics of this proposal.
● The sequencespecific penalizing parameters are proportional to an estimate of the standard deviation of the sequence signal. In other words, after initial normalization, the same penalties would be used across all signals.
● The tuning parameter for the total variation (fused lasso) and the Euclidean (group fused lasso) penalties on the jumps depend on $\sqrt{\text{log}\phantom{\rule{0.3em}{0ex}}N}$, where N is the possible number of jumps. This has a “multiple comparison controlling” effect and resembles rates that have been proven optimal under various sparse scenarios [46–49]. This term does not appear in the expression of λ_{1}, as the lasso penalty can be understood as providing a soft thresholding of the solution of (3) when λ_{1} = 0. Given the penalization due to λ_{2} and λ_{3}, the solution of (3) when λ_{1} = 0 will have much smaller dimension than N.
● The group penalty depends on $\sqrt{M}$, where M is the number of grouped sequences, as in the original proposal [33].
● The relative weight of the fusedlasso and groupfusedlasso penalties is regulated by ρ, which depends on p, the proportion of the M sequences expected to carry the same CNV. For example, if M = 2 and the two sequences record LRR and mBAF from the same individual, one expects p = 1 with ρ = 0, enforcing jumps at identical places in the two signals. Or, when studying the sequences relative to one parent and her/his offspring, it is reasonable to set p = 1/2, reflecting the fact that the two share half of their genome. For completely unrelated sequences, p = 0 and ρ = 1. In this setting, p is defined by genomewide characteristic of the samples. However, when analyzing specific genomic segments, corresponding to copy number polymorphisms, it is possible to use p to reflect the population frequencies of each of the copy number variants. We do not consider the problem of estimating a variable p in the present work.
The standard deviation ${\widehat{\sigma}}_{i}$ can be estimated robustly as follows. Let Δ_{ i j } = y_{i,j + 1} − y_{i,j}, for j = 1,…,N − 1, be the firstorder differences of adjacent y_{ i j } for sequence i. Then most $\text{Var}\left({\Delta}_{\mathit{ij}}\right)=2{\sigma}_{i}^{2}$ except those bridging real change points, so we can take
where $\hat{\mathit{SD}}\left({\mathit{\Delta}}_{i}\right)=\text{Standard Deviation}\left({\mathit{\Delta}}_{i}\right)$ or $\hat{\mathit{SD}}\left({\mathit{\Delta}}_{i}\right)=\text{Median Absolute Deiviation}\left({\mathit{\Delta}}_{i}\right)$ for Δ_{ i } = {Δ_{i,1},…,Δ_{i,N−1} }.
As mentioned before, the exact values of the penalty parameters should be adjusted depending on the expectations of signal strengths. Following the approach in [50], one can approximate the bias induced by each of the penalties and hence work backwards in terms of acceptable levels. As detailed in Additional file 1: Supplementary Text,
Following again the approach in [50], one can show that under some relatively strong assumptions, the choices in (7) lead to a consistent behavior as $N\to \infty $ and M stays bounded (see Additional file 1: Supplementary Text). Despite the fact that N is indeed large in our studies, it is not clear that we can assume it to be in the asymptotic regime. As finer scale measurements become available, scientists desire to investigate CNVs of decreasing length. The CNVs we are interested in discovering are often covered by a small number of probes. Furthermore we have often little information on the sizes and frequencies of CNVs. In this context, we find it advisable to rely on a twostage strategy:

1.
Sequences are jointly segmented by minimizing (3) for a relatively lax choice of the penalty parameters.

2.
Jumps are further thresholded on the basis of a datadriven cutoff.
Step 2 allows us to be adaptive to the signal strength and can be carried on with multiple methods. For example, one can adopt the modified Bayesian Information Criteria (mBIC) [51]. For sequence i, the jumps are sorted as $\{{\widehat{d}}_{i\left(1\right)},\dots ,{\widehat{d}}_{i(N1)}\}$ in the descending order of their absolute values. Then we choose the first $\widehat{k}$ change points where $\widehat{k}$ is given by
In data analysis, we often apply an even simpler procedure where the threshold for jumps is defined as a fraction of the maximal jump size observed for every sequence. Specifically, for sequence i, let ${\widehat{D}}_{i}={\text{max}}_{2\le j\le N}\left\{\right{\widehat{d}}_{\mathit{ij}}\left\right\}$, where ${\widehat{d}}_{\mathit{ij}}={\widehat{\beta}}_{\mathit{ij}}{\widehat{\beta}}_{i,j1}$ is the largest observed jump for sequence i. Then we define
as a “ruler” reflecting the scale of a possible real jump size, taking c γ_{ i } as the cutoff in removal of most small jumps. In all analyses for this paper, we fix a = 1, b = 5 and c = 0.2. In our experience, this heuristic procedure works well for both tumor and normal tissue CNV data.
Calling procedure
Even if this is not the focus of our proposal, in order to compare the performance of our segmentation algorithm with HMM approaches, it becomes necessary to distinguish gains from losses of copy number. While the same segmentation algorithm can be applied to a wide range of data sets, calling procedures depend more closely on the specific technology used to carry out the experiments. Since our data analysis relies on Illumina genotyping arrays, we limit ourselves to this platform and briefly describe the calling procedure adopted in the results section.
Analyzing one subject at the time, each segment with constant mean is assigned to one of five possible copy number states (c = 0,1,2,3,4). Let R collect the indexes of all SNPs comprising one segment and let (x_{ R } y_{ R }) = {(x_{ j }y_{ j }), j ∈ R} be the vectors of values for BAF and LRR in the segment. On the basis of typical patterns for BAF and LRR in the different copy number states (see [10, 13, 18]), we can write loglikelihood ratio
explicitly defined in Additional file 1: Supplementary Text. Segment R is assigned a CNV state ĉ that maximize LR(c), only if $\mathrm{LR}\left(\mathit{\u0109}\right)>{r}_{1}$, where r_{1} is a prespecified cutoff.
As noted in [16], the LRR data for a segment with c = 2, ideally normalized to have mean 0, often has a small nonzero mean due to experimental artifacts. If the number of SNPs in R is sufficiently large, the above loglikelihoodratio criterion would result in the erroneous identification of a copy number different from 2. To avoid this, we also require that the size of the absolute difference of the mean of LRR from zero be larger than a threshold $\left{\stackrel{\u0304}{y}}_{R}\right>{r}_{2}\sigma $.
Results and discussion
We report the results of the analysis of two simulated and two real data sets, which overall exemplify the variety of situations where joint segmentation of multiple sequences is attractive, as described in the motivation section. In all cases, we compare the performance of the proposed procedure with a set of relevant, often specialized, algorithms. The penalized estimation method we suggest in this paper shows competitive performance in all cases and often a substantial computational advantage. Its versatility and speed make it a very convenient tool for initial exploration. To calibrate the run times reported in the sequel, we state for the record that all of our analyses were run on a Mac OS X (10.6.7) machine with 2.93 GHz Intel Core 2 Duo and 4 GB 1067 MHz DDR3 memory.
Simulated CNV in normal samples
We consider one of the simulated datasets described in [25] with relatively short deletions and duplications (300 comprising 5, 10, 20, 30, 40, 50 SNPs each) are inserted in the middle of 13000 SNPs long sequences, using a combination of male and female X chromosome data from the Illumina HumanHap550 array, appropriately preprocessed to avoid biases. These steps included a scrambling of SNP positions so as to avoid longrange signal fluctuation. This setting mimics the small rare CNVs possibly occurring in the genome of normal individuals. In our main analysis, therefore, we process one individual at the time, reflecting the typical level of information available to scientists in these contexts. HMM methods, like PennCNV, are expected to be the most effective in this problem; segmentation methods like CBS are closer to our own and therefore also make an interesting comparison. As repeatedly discussed, the Illumina platform produces two signals for one subject, LRR and BAF. A segmentation method that can process one signal at the time would give its best results using LRR, which carries most of the information. Given this background, we compare four methods: PennCNV, CBS on LRR, fused lasso on LRR only, and group fused lasso on LRR and mBAF. The implementations we use are those reflected in the software packages: PennCNV (version 2010May01), R package DNAcopy for CBS (version 1.24.0) [52] and our own R package Piet (version 0.1.0). Tuning parameters for PennCNV and CBS are set at the default values; the fused lasso implementation corresponds to λ_{1} = 0.1, ${\lambda}_{2}=2\times \sqrt{13000}$, and λ_{3} = 0 and the group fused lasso to λ_{1} = 0.1, λ_{2} = 0, and ${\lambda}_{3}=2\times \sqrt{13000}$. To call deletions and duplications with CBS and the two fusedlasso approaches, we use both LRR and BAF data (prior to transformation to mBAF) with the following cutoff values: r_{1} = 10 and r_{2} = 1(1.5) for duplication (deletion). Performance is evaluated by the same indexes we used in [25], the true positive rate (TPR or sensitivity) and the false discovery rate (FDR), all defined on a per SNP basis. Results are summarized in Table 1.
Not surprisingly, all algorithms perform similarly well for larger deletions/duplications, and it is mainly for variants that involve 10 or fewer SNPs that differences are visible. Algorithms that rely only on LRR (for example, CBS and fused lasso) underperform in the detection of small duplications. Comparison is particularly easy for duplications involving 10 SNPs, where the selected parameter values lead to similar FDRs in the three segmentation methods. The group fused lasso can almost entirely recover the performance of PennCNV and outperforms CBS in this context.
Out of curiosity, we analyzed all sequences simultaneously. While this represents an unrealistic amount of prior information, it allows us to evaluate the possible advantages of joint analysis. FDR practically became 0 (<0.02%) for all CNV sizes, but power increases only for CNVs including fewer than 10 SNPs.
Finally, it is useful to compare running times. Summary statistics of the per sample time are reported in Table 1. While all algorithms are rather fast, the two implementations of the fused lasso dominate.
A simulated tumor data set
To explore the challenges presented by tumor data, we rely on a data set created by [17], with the specific goal of studying the effect of contamination between normal and cancer cells. The HapMap sample NA06991, genotyped on the Illumina HumanHap550 array, was used to simulate a cancer cell line by inserting a total of 10 structure variation regions, including onecopy losses, onecopy gains, and copy neutral lossofhetrozygosity (CNLOH) (see Additional file 2: Table S1). The signal from this artificial “tumor” sample was then contaminated in silico with that of the original “normal” sample, resulting in 21 data sets, with a percentage of normal cells ranging from 0% to 100%. Note that most simulated CNV or CNLOH regions are very large (some spanning an entire chromosome), and the challenge in detection is really due to the contamination levels.
For ease of comparison, we evaluate the accuracy of calling procedures as in the original reference [17]. Sensitivity is measured for each variant region as the percentage of heterozygous SNPs that are assigned the correct copy number, and specificity is measured as the percentage of originally heterozygous SNPs in unperturbed regions that are assigned CN=2. We compare the performance of GFL to BAFsegmentation [17] and PSCN [53] representing, respectively, a version of segmentation and HMM approaches specifically developed to deal with contaminated tumor samples. Both of these algorithms have been tested with success on this simulated data set.
Following other analyses, we do not preprocess the data prior to CNV detection. BAFsegmentation and PSCN were run using recommended parameter values. For each of the diluted data sets, we applied the GFL model on each chromosome, simultaneously using both LRR and mBAF, whose standard deviations are normalized to 1. Tuning constants are set to λ_{1} = 0, ${\lambda}_{2}=0.5\times 3\times \sqrt{\text{log}\phantom{\rule{0.3em}{0ex}}N}$, and ${\lambda}_{3}=0.5\times 3\times \sqrt{\text{log}\phantom{\rule{0.3em}{0ex}}N}$ for chromosomes interrogated by N SNPs. The change points resulting from hard segmentation on LRR and mBAF are combined to make a finer segmentation of the genome. Finally, we adopt the same calling procedure described by [17]. For ease of comparison with PSCN, only analyses of simulated tumor data are reported, even if BAFsegmentation and GFL would gain from using the genotype of the normal cell in defining mBAF.
Figure 1 summarizes the sensitivity of each method, as a function of the percentage of normal cells in the sample. Sensitivity is calculated for each of the 10 regions separately. All three methods work reasonably well under a wide range of percentages of normal cell contamination. In 5 out of the 10 regions, GFL appears to lead to best results, while in the other 5, PSCN does. The CNV region involving the fewest SNPs is the hemizygous loss on Chromosome 13. In this case GFL in our hands behaved in the most stable manner. GFL outperforms the two comparison methods in terms of specificity (Figure 2). While the specificity values might appear very high in any case, this is somewhat an artifact of how we define this index. In the interest of fairness, it is relevant to note that the performance of PSCN in our hands did conform to the published performance [53]. While we tried our best to set the parameter values, we have not succeeded in replicating the authors’ original results.
PSCN, like GFL, is implemented in R with some computationally intensive subroutines coded in C. BAFsegmentation relies on the R package DNAcopy, whose core algorithms are implemented in C and Fortran. BAFsegmentation wraps these in Perl. A comparison of run times indicate that GLF and BAFsegmentation are comparable, while PSCN is fifty times slower than GFL (see Additional file 3: Table S2).
One sample assayed with multiple replicates and multiple platforms
We use the data from a study [54] assessing the performance of different array platforms and CNV calling methods to illustrate the advantages of joint analysis of multiple measurements on the same subject. DNA from four individuals was analyzed in triplicate on each of 5 platforms: Affymetrix 6.0, Illumina 1M, 660W, Omni1Quad (O1Q) and Omni2.5Quad (O2Q) (among others [54]). We use the results on the first three to define “true” copy numbers and try to reconstruct them using data from O1Q and O2Q. The nine “reference” experiments were analyzed with 4 or 5 CNV calling algorithms [54] and a CNV was identified using majority votes. Consistent evidence was required from at least 2 analysis tools, on at least 2 platforms, and in at least 2 replicates (see Additional file 4: Table S3). Here overlapping CNVs detected in two replicates/algorithms/platforms are collapsed to a single CNV.
The test experiments are based on 1,020,596 and 2,390,395 autosomal SNPs, which after quality control reduce to a total of 2,657,077 unique loci. Since our focus here is to investigate how to best analyze multiple signals on the same subject, rather than on the specific properties of any CNV calling method, we carry out all the analyses using different settings of GFL in segmentation while keeping the same CNV calling and summarizing procedures. All segmentation is done on LRR only, while calling procedure uses both LRR and BAF (with cutoff r_{1} = 10 and r_{2} = 1). Here we compare three segmentation settings to analyze these 6 experiments per subject (see Additional file 5: Table S4 for more details about tuning parameters):

1.
The signals from the three technical replicates with one platform are averaged and then segmented and subjected to calling procedure separately. The final CNV list is the union of CNV calls from the two platforms.

2.
The signals from the three technical replicates with one platform are each segmented and separately subjected to calling. A majority vote of at least two out of three is used to summarize each CNV result for each platform. The final CNV list is the union of the two platforms’ lists.

3.
The signals from the three technical replicates of both platforms (6 LRR sequences) are segmented jointly. Calling is still done on each replicate separately, and the same majority vote is used to summarize the CNV result for each platform. Again, the final CNV list is the union of the two platforms’ results.
To benchmark the result of joint analysis, we use MPCBS [20], a segmentation method, specifically designed for multiplatform CNV analysis. The segments output from MPCBS are subjected to the same calling, majority voting, and summarizing procedures.
Table 2 presents our results. Averaging from different technical replicates leads to loss of power, while joint analysis of all the signals leads to the most effective performance. GFL joint analysis leads to results comparable to those of MPCBS, but it is at least 30 times faster than MPCBS.
Multiple related samples assayed with the same platform
In the context of a study of the genetic basis of bipolar disorder, the Illumina Omni2.5Quad chip was used to genotype 455 individuals from 11 Columbian and 13 Costa Rican pedigrees. We use this data set to explore the advantages of a joint segmentation of related individuals. In the absence of a reference evaluation of CNV status in these samples, we rely on two indirect methods to assess the quality of the predicted CNVs. We used the collection of CNVs observed in HapMap Phase III [5] to compile a list of 426 copy number polymorphisms (selecting all those CNVs with frequency ≥0.05 in pooled samples from 11 populations) and assumed that if we identify in our sample a CNV corresponding to one of these regions, we should consider it a true positive. For the purposes of this analysis we considered a detected CNV to correspond to one identified in HapMap if there was any overlap between the two regions.
Another indirect measure of the quality of CNV calls derives from the number of Mendelian errors encountered in the pedigrees when we consider the CNV as a segregating site. De novo CNVs are certainly a possibility, and in their case Mendelian errors are to be expected. However, when the CNV in question is a common one (already identified in HapMap), it is reasonable to expect that it segregates in the pedigrees as any regular polymorphism. We selected a very common deletion on Chromosome 8 (HapMap reports overall frequency >0.4 in 11 populations) and compared different CNV calling procedures on the basis of how many Mendelian errors they generate.
As mentioned before, PennCNV represents a stateoftheart HMM method for the analysis of normal samples and, therefore, we included it in our comparisons. However, the parameters of the underlying HMM algorithm had not been tuned on the Omni2.5Quad at this time, resulting in substandard performance. Segmentation methods are less dependent on parameter optimization; hence, GFL analysis of LRR and BAF one subject at a time can provide a better indication of the potential of singlesample methods. We considered two multiplesample algorithms: GFL and MSSCAN [16], both applied on LRR with the group structure defined by pedigree membership. (While a triomode is available in PennCNV [55], this does not adapt to the structure of our families.) A final qualification is in order. While the authors of MSSCAN kindly shared with us a betaversion of their software, we did not find it to be robust. Indeed, we were unable to use it to segment the entire genome. However, we successfully used it to segment Chromosome 8, so that we could include MSSCAN in the comparison based on Mendelian errors.
Prior to analysis, the data was normalized using the GCcontent correction implemented in PennCNV [29]. For individual analysis, the GFL parameters were λ_{1} = 0.1, λ_{2} = 0, and ${\lambda}_{3}=2\times \sqrt{\text{log}\phantom{\rule{0.3em}{0ex}}N}$, where N is the number of SNPs deployed on each chromosome; for pedigree analysis, the GFL parameters were λ_{1} = 0.1, ${\lambda}_{2}=0.5\times 2\times \sqrt{\text{log}\phantom{\rule{0.3em}{0ex}}N}$, and ${\lambda}_{3}=0.5\times 2\times \sqrt{0.3M}\times \sqrt{\text{log}\phantom{\rule{0.3em}{0ex}}N}$, where M is the number of individuals in each pedigree. For MSSCAN, CNV size is constrained to be fewer than 200 SNPs, and the maximum number of change points is set at 50. The calling procedure with r_{1} = 10 and r_{2} = 1 was applied to both the GFL and MSSCAN results.
Table 3 summarizes the total number of copy number polymorphisms (CNPs) identified in our sample by different approaches and their overlap with known CNPs from HapMap. For the purpose of this comparison we considered a variant to be a CNP when its frequency in our sample was at least 10%. All analysis modes of GFL show a higher percent of overlap with the HapMap list than the PennCNV list. It is also clear that GFL pedigree analysis achieves a larger overlap with the HapMap data than the GFL individual analysis. The time cost per sample is reasonable and scales well with the increment of sample size.
Table 4 summarizes the results of our investigation of a 154kb CNP region on Chromosome 8p (from 39,351,896 to 39,506,122 on NCBI Build 36 coordinate). All methods but PennCNV show detected deletions only; this coincides with the observation from HapMap data. We used option Mistyping of Mendel (version 11.0) [56, 57] to detect Mendelian errors. Joint segmentation methods discover more hemizygous deletions than individual analysis, resulting in fewer Mendelian errors. MSSCAN discovers the largest number of hemizygous deletions. Figure 3 shows four families derived from a large pedigree where 3 out of 4 Mendelian errors are removed by joint analysis.
Conclusions
We have presented a segmentation method based on penalized estimation that is capable of processing multiple signals jointly. We have shown how this leads to improvements in the analysis of normal samples (where segmentation can be applied to both total intensity and allelic proportions), tumor samples (where we are able to deal with contamination effectively), measurements from multiple platforms, and related individuals. Given that copy number detection is such an active area of research, it is impossible to compare one method to all other available methods. However, for each of the situations we analyzed, we tried to select state of the art alternative approaches. In comparison to these, the algorithm we present performs well. Its accuracy is always comparable to that of the most effective competitor and its computation times are better contained. Given its versatility and speed, GFL is, in our opinion, particularly useful for initial screening.
There are of course many aspects of CNV detection, ranging from normalization and signal transformation to FDR control of detected CNV, that we have not analyzed in this paper. There are also a number of improvements to our approach that appear promising, but at this stage are left for further work. For example, it is easy to modify algorithms so that the penalization parameters are location dependent and incorporate prior information on known copy number polymorphisms. It will be more challenging to develop theory and methods to select the values of these regularization parameters in a dataadaptive fashion.
Finally, while our scientific motivation has been the study of copy number variations, the joint segmentation algorithm we present is not restricted to specific characteristics of these data types, and we expect it will be applied in other contexts.
Implementation and availablity
We have implemented the segmentation routine, which is our core contribution, in an R package (Piet) available at Rforge [58]. To demonstrate a visualization of the CNV results on Chromosome 8 in the bipolar disorder study, we refer readers to Additional file 6: Figure S1.
Author’s contributions
ZZ, KL, and CS conceived this study and participated in model and algorithm development. ZZ performed the statistical analysis and wrote the R Piet implementation. All authors participated in writing the final manuscript. All authors read and approved the final manuscript.
Abbreviations
 BAF:

B allele frequency
 CN:

Copy number
 CNV:

Copy number variant
 CNP:

Copy number polymorphism
 CNLOH:

Copy neutral loss of heterozygosity
 GFL:

Generalized fused lasso
 HMM:

Hidden Markov model
 LRR:

Log R ratio
 MML:

Majorizationminimization.
References
 1.
Newton MA, Lee Y: Inferring the location and effect of tumor suppressor genes by instabilityselection modeling of allelicloss data. Biometrics 2000, 56: 1088–1097. 10.1111/j.0006341X.2000.01088.x
 2.
Iafrate AJ, Feuk L, Rivera MN, Listewnik ML, Donahoe PK, Qi Y, Scherer SW, Lee C: Detection of largescale variation in the human genome. Nat Gene 2004, 36: 949–951. 10.1038/ng1416
 3.
Sebat J, Lakshmi B, Troge J, Alexander J, Young J, Lundin P, Maner S, Massa H, Walker M, Chi M, Navin N, Lucito R, Healy J, Hicks J, Ye K, Reiner A, Gilliam TC, Trask B, Patterson N, Zetterberg A, Wigler M: Largescale copy number polymorphism in the human genome. Science 2004, 305: 525–528. 10.1126/science.1098918
 4.
Pinto D, Redon R, Feuk L, Gokcumen O, Zhang Y, Aerts J, Andrews TD, Barnes C, Campbell P, Fitzgerald T, Hu M, Ihm CH, Kristiansson K, MacArthur DG, MacDonald JR, Onyiah I, Pang AWC, Robson S, Stirrups K, Valsesia A, Walter K, Wei J, The Wellcome Trust Case Control Consortium, TylerSmith C, Carter NP, Lee C, Scherer SW, Hurles ME, Conrad DF: Origins and functional impact of copy number variation in the human genome. Nature 2009, 464: 704–712.
 5.
Jakobsson M, Scholz SW, Scheet P, Gibbs JR, VanLiere JM, Fung HC, Szpiech ZA, Degnan JH, Wang K, Guerreiro R, Bras JM, Schymick JC, Hernandez DG, Traynor BJ, SimonSanchez J, Matarin M, Britton A, van de Leemput J, Rafferty I, Bucan M, Cann HM, Hardy JA, Rosenberg NA, Singleton AB: Genotype, haplotype and copynumber variation in worldwide human populations. Nature 2008, 451: 998–1003. 10.1038/nature06742
 6.
Mills RE, Walter K, Stewart C, Handsaker RE, Chen K, Alkan C, Abyzov A, Yoon SC, Ye K, Cheetham RK, Chinwalla A, Conrad DF, Fu Y, Grubert F, Hajirasouliha I, Hormozdiari F, Iakoucheva LM, Iqbal Z, Kang S, Kidd JM, Konkel MK, Korn J, Khurana E, Kural D, Lam HYK, Leng J, Li R, Li Y, Lin CY, Luo R, et al.: Mapping copy number variation by populationscale genome sequencing. Nature 2011, 470: 59–65. 10.1038/nature09708
 7.
Redon R, Ishikawa S, Fitch KR, Feuk L, Perry GH, Andrews TD, Fiegler H, Shapero MH, Carson AR, Chen W, Cho EK, Dallaire S, Freeman JL, Gonzalez JR, Gratacos M, Huang J, Kalaitzopoulos D, Komura D, MacDonald JR, Marshall CR, Mei R, Montgomery L, Nishimura K, Okamura K, Shen F, Somerville MJ, Tchinda J, Valsesia A, Woodwark C, Yang F, et al.: Global variation in copy number in the human genome. Nature 2006, 444: 444–454. 10.1038/nature05329
 8.
Pinto D, Pagnamenta AT, Klei L, Anney R, Merico D, Regan R, Conroy J, Magalhaes TR, Correia C, Abrahams BS, Almeida J, Bacchelli E, Bader GD, Bailey AJ, Baird G, Battaglia A, Berney T, Bolshakova N, Bölte S, Bolton PF, Bourgeron T, Brennan S, Brian J, Bryson SE, Carson AR, Casallo G, Casey J, Chung BHY, Cochrane L, Corsello C, et al.: Functional impact of global rare copy number variation in autism spectrum disorders. Nature 2010, 466: 368–372. 10.1038/nature09146
 9.
Stefansson H, Rujescu D, Cichon S, Pietiläinen OPH, Ingason A, Steinberg S, Fossdal R, Sigurdsson E, Sigmundsson T, BuizerVoskamp JE, Hansen T, Jakobsen KD, Muglia P, Francks C, Matthews PM, Gylfason A, Halldorsson BV, Gudbjartsson D, Thorgeirsson TE, Sigurdsson A, Jonasdottir A, Jonasdottir A, Bjornsson A, Mattiasdottir S, Blondal T, Haraldsson M, Magnusdottir BB, Giegling I, Möller HJ, Hartmann A, et al.: Large recurrent microdeletions associated with schizophrenia. Nature 2008, 455: 232–236. 10.1038/nature07229
 10.
Colella S, Yau C, Taylor JM, Mirza G, Butler H, Clouston P, Bassett AS, Seller A, Holmes CC, Ragoussis J: QuantiSNP: An objective Bayes hiddenMarkov model to detect and accurately map copy number variation using SNP genotyping data. Nucleic Acids Res 2007, 35: 2013–2025. 10.1093/nar/gkm076
 11.
Scharpf RB, Parmigiani G, Pevsner J, Ruczinski I: Hidden Markov models for the assessment of chromosomal alterations using high throughput SNP arrays. The Ann Appl Stat 2008, 2: 687–713. 10.1214/07AOAS155
 12.
Sun W, Wright FA, Tang Z, Nordgard SH, Van Loo P, Yu T, Kristensen VN, Perou CM: Integrated study of copy number states and genotype calls using highdensity SNP arrays. Nucleic Acids Res 2009, 37: 5365–5377. 10.1093/nar/gkp493
 13.
Wang K, Li M, Hadley D, Liu R, Glessner J, Grant SFA, Hakonarson H, Bucan M: PennCNV: An integrated hidden Markov model designed for highresolution copy number variation detection in wholegenome SNP genotyping data. Genome Res 2007, 17: 1665–1674. 10.1101/gr.6861907
 14.
Yau C, Papaspiliopoulos O, Roberts GO, Holmes C: Bayesian nonparametric hidden Markov models with applications in genomics. J R Stat Soc: Ser B 2011, 73: 37–57. 10.1111/j.14679868.2010.00756.x
 15.
Olshen AB, Venkatraman ES, Lucito R, Wigler M: Circular binary segmentation for the analysis of arraybased DNA copy number data. Biostatistics 2004, 5: 557–572. 10.1093/biostatistics/kxh008
 16.
Zhang NR, Siegmund DO, Ji H, Li JZ: Detecting simultaneous changepoints in multiple sequences. Biometrika 2010, 97: 631–645. 10.1093/biomet/asq025
 17.
Staaf J, Lindgren D, VallonChristersson J, Isaksson A, Göransson H, Juliusson G, Rosenquist R, Höglund M, Borg A, Ringnér M: Segmentationbased detection of allelic imbalance and lossofheterozygosity in cancer cells using whole genome SNP arrays. Genome Biol 2008, 9: R136. 10.1186/gb200899r136
 18.
Wang H, Veldink JH, Blauw H, van den Berg L H, Ophoff RA, Sabatti C: Markov models for inferring copy number variations from genotype data on Illumina platforms. Human Heredity 2009, 68: 1–22. 10.1159/000210445
 19.
Siegmund DO, Yakir B, Zhang NR: Detecting simultaneous intervals in aligned sequences. The Ann Appl Stat 2011, 5: 645–668. 10.1214/10AOAS400
 20.
Zhang NR, Senbabaoglu Y, Li JZ: Joint estimation of DNA copy number from multiple platforms. Bioinfomatics 2010, 26: 153–160. 10.1093/bioinformatics/btp653
 21.
PiqueRegi R, MonsoVarona J, Ortega A, Asgharzadeh S: Bayesian detection of recurrent copy number alterations across multiple array samples. In. IEEE Int Workshop on Genomic Signal Process Stat (GENSiPS) 2008, 1–4.
 22.
Efron B, Zhang NR: False discovery rates and copy number variation. Biometrika 2011, 98: 251–271. 10.1093/biomet/asr018
 23.
Nowak G, Hastie T, Pollack JR, Tibshirani R: A fused lasso latent feature model for analyzing multisample aCGH data. Biostatistics 2011, 12: 776–791. 10.1093/biostatistics/kxr012
 24.
Tibshirani R, Wang P: Spatial smoothing and hot spot detection for CGH data using the Fused Lasso. Biostatistics 2008, 9: 18–29.
 25.
Zhang Z, Lange K, Ophoff R, Sabatti C: Reconstructing DNA copy number by penalized estimation and imputation. The Ann Appl Stat 2010, 4: 1749–1773. 10.1214/10AOAS357
 26.
Bengtsson H, Neuvial P, Speed T: TumorBoost: Normalization of allelespecific tumor copy numbers from a single pair of tumornormal genotyping microarrays. BMC Bioinf 2010, 11: 245. 10.1186/1471210511245
 27.
Bengtsson H, Wirapati P, Speed TP: A singlearray preprocessing method for estimating fullresolution raw copy numbers from all Affymetrix genotyping arrays including GenomeWideSNP 5 & 6. Bioinformatics 2009, 25: 2149–2156. 10.1093/bioinformatics/btp371
 28.
Carvalho B, Bengtsson H, Speed TP, Irizarry RA: Exploration, normalization, and genotype calls of highdensity oligonucleotide SNP array data. Biostatistics 2007, 8: 485–499.
 29.
Diskin SJ, Li M, Hou C, Yang S, Glessner J, Hakonarson H, Bucan M, Maris JM, Wang K: Adjustment of genomic waves in signal intensities from wholegenome SNP genotyping platforms. Nucleic Acids Res 2008, 36: e126. 10.1093/nar/gkn556
 30.
Scharpf RB, Irizarry RAME, Ritchie Carvalho B: Using the R package crlmm for genotyping and copy number estimation. J Stat Software 2011, 40: 1–32.
 31.
Scharpf RB, Ruczinski I, Carvalho B, Doan B, Chakravarti A, Irizarry RA: A multilevel model to address batch effects in copy number estimation using SNP arrays. Biostatistics 2011, 12: 33–50. 10.1093/biostatistics/kxq043
 32.
Ostrovnaya I, Olshen AB, Seshan VE, Orlow I, Albertson DG, Begg CB: A metastasis or a second independent cancer? Evaluating the clonal origin of tumors using array copy number data. Stat Med 2010, 29: 1608–1621.
 33.
Yuan M, Lin Y: Model selection and estimation in regression with grouped variables. J R Stat Soc: Ser B 2006, 68: 49–67. 10.1111/j.14679868.2005.00532.x
 34.
Zhou H, Sehl ME, Sinsheimer JS, Lange K: Association screening of common and rare genetic variants by penalized regression. Bioinformatics 2010, 26: 2375–2382. 10.1093/bioinformatics/btq448
 35.
Bleakley K, Vert JP: The group fused Lasso for multiple changepoint detection. Arxiv preprint arXiv:1106.4199 2011.
 36.
Friedman J, Hastie T, Tibshirani R: A note on the group lasso and a sparse group lasso. Arxiv preprint arXiv:1001.0736 2010.
 37.
Rudin LI, Osher S, Fatemi E: Nonlinear total variation based noise removal algorithms. Physica D: Nonlinear Phenom 1992, 60: 259–268. 10.1016/01672789(92)90242F
 38.
Friedman J, Hastie T, Höfling H, Tibshirani R: Pathwise coordinate optimization. The Ann Appl Stat 2007, 1: 302–332. 10.1214/07AOAS131
 39.
Wu TT, Lange K: Coordinate descent algorithm for lasso penalized regression. The Ann Appl Stat 2008, 2: 224–244.
 40.
Hoefling H: A path algorithm for the fused lasso signal approximator. J Comput Graphical Stat 2010, 19: 984–1006. 10.1198/jcgs.2010.09208
 41.
Tibshirani RJ, Taylor J: The solution path of the generalized lasso. The Ann Stat 2011, 39: 1335–1371. 10.1214/11AOS878
 42.
Zhou H, Lange K: A path algorithm for constrained estimation. Arxiv preprint arXiv:1103.3738 2011.
 43.
Lange K: Optimization. New York: Springer; 2004.
 44.
Conte SD, deBoor C: Elementary Numerical Analysis. New York: McGrawHill; 1972.
 45.
PiqueRegi R, MonsoVarona J, Ortega A, Seeger RC, Triche TJ, Asgharzadeh S: Sparse representation and Bayesian detection of genome copy number alterations from microarray data. Bioinformatics 2008, 24: 309–318. 10.1093/bioinformatics/btm601
 46.
Bickel PJ, Ritov Y, Tsybakov AB: Simultaneous analysis of Lasso and Dantzig selector. The Ann Stat 2009, 37: 1705–1732. 10.1214/08AOS620
 47.
Bunea F, Tsybakov A, Wegkamp M: Sparsity oracle inequalities for the Lasso. Electron J Stat 2007, 1: 169–194. 10.1214/07EJS008
 48.
Candes E, Tao T: The Dantzig selector: Statistical estimation when p is much larger than n. The Ann Stat 2007, 35: 2313–2351. 10.1214/009053606000001523
 49.
Donoho DL, Johnstone IM: Ideal spatial adaptation by wavelet shrinkage. Biometrika 1994, 81: 425–455. 10.1093/biomet/81.3.425
 50.
Rinaldo A: Properties and refinements of the fused lasso. The Ann Stat 2009, 37: 2922–2952. 10.1214/08AOS665
 51.
Zhang NR, Siegmund DO: A modified Bayes information criterion with applications to the analysis of comparative genomic hybridization data. Biometrics 2007, 63: 22–32. 10.1111/j.15410420.2006.00662.x
 52.
Venkatraman ES, Olshen AB: A faster circular binary segmentation algorithm for the analysis of array CGH data. Bioinformatics 2007, 23: 657–663. 10.1093/bioinformatics/btl646
 53.
Chen H, Xing H, Zhang NR: Estimation of parent specific DNA copy number in tumors using highdensity genotyping arrays. PLoS Comput Biol 2011, 7: e1001060. 10.1371/journal.pcbi.1001060
 54.
Pinto D, Darvishi K, Shi X, Rajan D, Rigler D, Fitzgerald T, Lionel AC, Thiruvahindrapuram B, MacDonald JR, Mills R, Prasad A, Noonan K, Gribble S, Prigmore E, Donahoe PK, Smith RS, Park JH, Hurles ME, Carter NP, Lee C, Scherer SW, Feuk L: Comprehensive assessment of arraybased platforms and calling algorithms for detection of copy number variants. Nat Biotechnol 2011, 29: 512–520. 10.1038/nbt.1852
 55.
Wang K, Chen Z, Tadesse MG, Glessner J, Grant SFA, Hakonarson H, Bucan M, Li M: Modeling genetic inheritance of copy number variations. Nucleic Acids Res 2008, 36: e138. 10.1093/nar/gkn641
 56.
Lange K, Cantor R, Horvath S, Perola M, Sabatti C, Sinsheimer J, Sobel E: Mendel version 4.0: A complete package for the exact genetic analysis of discrete traits in pedigree and population data sets. The Am J Human Genetics 2001, 69: 504. 10.1086/322739
 57.
Sobel E, Papp JC, Lange K: Detection and integration of genotyping errors in statistical genetics. The Am J Human Genet 2002, 70: 496–508. 10.1086/338920
 58.
RForge Project Piet [http://rforge.rproject.org/projects/piet/] []
Acknowledgements
The authors thank Nelson Freimer and all of the collaborators of the Bipolar Endophenotype Mapping project for authorizing use of their genotype data. We also thank Susan Service and Joseph DeYoung for assistance in data management and interpretation and Pierre Neuvial and Henrik Bengtsson for helpful discussion. CS gratefully acknowledges support from NIH/NIGMS GM053275, MH075007 and P30 1MH083268 and KL from NIH/NIGMS GM053275.
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Electronic supplementary material
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
About this article
Cite this article
Zhang, Z., Lange, K. & Sabatti, C. Reconstructing DNA copy number by joint segmentation of multiple sequences. BMC Bioinformatics 13, 205 (2012). https://doi.org/10.1186/1471210513205
Received:
Accepted:
Published:
Keywords
 Copy number variant
 Copy number polymorphism
 Fused lasso
 Group fused lasso
 MM algorithm