- Open Access
Computational tools for copy number variation (CNV) detection using next-generation sequencing data: features and perspectives
BMC Bioinformatics volume 14, Article number: S1 (2013)
Copy number variation (CNV) is a prevalent form of critical genetic variation that leads to an abnormal number of copies of large genomic regions in a cell. Microarray-based comparative genome hybridization (arrayCGH) or genotyping arrays have been standard technologies to detect large regions subject to copy number changes in genomes until most recently high-resolution sequence data can be analyzed by next-generation sequencing (NGS). During the last several years, NGS-based analysis has been widely applied to identify CNVs in both healthy and diseased individuals. Correspondingly, the strong demand for NGS-based CNV analyses has fuelled development of numerous computational methods and tools for CNV detection. In this article, we review the recent advances in computational methods pertaining to CNV detection using whole genome and whole exome sequencing data. Additionally, we discuss their strengths and weaknesses and suggest directions for future development.
Genomic variation is comprised of single nucleotide variants (SNVs), small insertions or deletions (indels), copy number variations (CNVs), and large structural variants (SVs); these variants range from single base changes to large chromosomal-level alterations . CNV refers to a type of intermediate-scale SVs with copy number changes involving a DNA fragment that is typically greater than one kilobases (Kb) and less than five megabases (Mb) . The importance of CNVs was first recognized by their prevalence in healthy individuals [3, 4]. It is estimated that approximately 12% of the genome in human populations is subject to copy number change . These pervasive CNVs in the genome are thought to have equal contributions to genetic variation in humans as another major type of variation, SNVs, which have long been considered the most abundant genetic variation in humans . So far, approximately half of the reported CNVs overlap with protein-coding regions . These gains and losses of gene copies might directly influence gene dosage within the CNV regions, which could result in a change of gene expression level. Similar to SNVs, CNVs do not necessarily have a negative effect on human health. However, among the large number of CNVs, some might have an association with, or directly involve in, diseases and phenotypes such as cancer and neuropsychiatric disorders [7, 8]. Generally, CNVs include deletions, insertions, and duplications of genomic regions. Because the lengths of CNVs vary greatly, the current computational tools usually target a specific range of CNV sizes. The readers may be aware that this review focuses on all types of CNVs including CNV events less than 1 Kb, intermediate structural variants greater than 1 Kb, and large chromosomal events over 5 Mb.
The traditional approach to identify CNVs employs cytogenetic technologies, such as karyotyping and fluorescence in situ hybridization (FISH) . In 2003, genome-wide detection of CNVs was achieved using more accurate array-based comparative genomic hybridization (arrayCGH) and single-nucleotide polymorphism (SNP) array approaches ; these approaches, however, have suffered from several inherent drawbacks, including hybridization noise, limited coverage for genome, low resolution, and difficulty in detecting novel and rare mutations [11, 12].
Over the last few years, next-generation sequencing (NGS) has evolved into a popular strategy for genotyping and has included comprehensive characterization of CNVs by generating hundreds of millions of short reads in a single run . Since 2005, several commercial platforms, including those from 454 Life Sciences, Illumina, Inc., and Life Technologies, have increased read generation throughput and base-calling accuracy unprecedentedly, enabling the sequencing of a whole human genome at a much lower cost and a faster turnaround time . Compared to array-based approaches, where probes are predefined for limited genomic regions, short reads from NGS platforms are randomly sampled from the entire genome. The advantages of the NGS approach also include higher coverage and resolution, more accurate estimation of copy numbers, more precise detection of breakpoints, and higher capability to identify novel CNVs [1, 14]. Taking these advantages into account, a diverse set of tools has been developed to detect CNVs based on different features that can be extracted from NGS data.
To provide guidelines for the rapidly growing number of CNV studies using NGS, this review presents a systematic investigation of the 37 currently available tools, which differ in computational strategies to pinpoint CNVs in whole genome sequencing (WGS) data. In addition, 11 in silico tools specific to whole exome sequencing (WES) data are discussed separately due to the challenges unique to CNV detection in WES data. Based on the discussion of the functions and limitations of current CNV-calling tools, and perspectives for future development, we mainly focus on: (i) the key features for CNV calling tools using NGS data, (ii) the key factors to consider before pipeline design, and (iii) developing combinatorial approach for more accurate CNV calling.
Five strategies for CNV detection through NGS data
Many SV-detecting methods can be applied to CNV identification. So far, the NGS based CNV detection methods can be categorized into five different strategies, including: (1) paired-end mapping (PEM), (2) split read (SR), (3) read depth (RD), (4) de novo assembly of a genome (AS), and (5) combination of the above approaches (CB) (Figure 1). Indeed, different strategies have their own advantages and limitations. Though there has been great progress in each category, none of the methods could comprehensively detect all types of CNVs. As summarized in Tables 1, 2, 3, 4, there are 6 PEM-based tools, 4 SR-based tools, 26 RD-based tools, 3 AS-based tools, and 9 tools for combinatorial approaches. Most PEM-, SR-, and CB-based tools are not specific to CNV detection but rather for SV identification, while the majority of RD- and AS-based tools are developed for the detection of CNVs instead of SVs. From the point of input materials, a majority (28) of the 48 tools listed in Tables 1, 2, 3, 4 accept SAM or BAM files as input. Specifically, most RD-based tools start from SAM/BAM files with depth information, while PEM- and SR-based tools can accept FASTQ files since they may not use read depth information. Among these tools, 32 were implemented using R and/or C/C++ computer languages, as they are convenient in incorporating statistical models, improving computational efficiency, and distributing across multiple operating systems. Note that most of these tools are standalone instead of online tools, except for visualization.
Paired-end mapping approach
The feasibility of using NGS data to detect SVs/CNVs was first achieved by PEM methods . To date, six tools were developed specifically based on PEM (Table 1). In addition, at least 8 CB tools incorporate PEM algorithms to improve accuracy (Table 4). Notably, PEM is only applicable to paired-end reads but not single-end reads . In paired-end sequencing, the DNA fragments from the same library preparation protocol have a specific distribution of the insert size. PEM identifies SVs/CNVs from discordantly mapped paired-reads whose distances are significantly different from the predetermined average insert size (Figure 1a). PEM methods can efficiently identify not only insertions and deletions but also mobile element insertions, inversions, and tandem duplications. However, PEM is not applicable to insertion events larger than the average insert size of the genome library . Another limitation of PEM is its inability to detect CNVs in low-complexity regions with segmental duplication.
Two different strategies have been used in PEM-based tools to detect SVs/CNVs, namely the clustering approach and the model-based approach. The difference lies in that the clustering approach employs a predefined distance to identify discordant reads, while the model-based approach adopts a probability test to discover the unusual distance between read pairs in comparison to the distance distribution in genome. For the clustering approach, a cluster often includes at least two read pairs (Figure 1a), which is the minimal requirement to ensure the accuracy of the predication of breakpoints and the SV/CNV sizes . The tool BreakDancer provides both clustering-based and model-based approaches to detect various SVs : its module BreakDancerMax is clustering-based while the other module, BreakDancerMini, uses a model-based approach to detect smaller insertions and deletions ranging from 10- to 100-base pairs (bps). PEMer is another tool that uses a clustering-based strategy to detect SVs . It implements various cluster sizes to guide read clustering and uses a customized cutoff to detect discordant read pairs. The advantage of PEMer is that the read clusters obtained from different platforms and different parameterizations can be merged together.
Both BreakDancer and PEMer are based on the hard cluster method. That is, each read can only be assigned to one cluster, and the reads that can be mapped to multiple genomic coordinates are discarded, even if they are mapped with high quality. Because of this limitation, it is not feasible to detect SVs/CNVs in repetitive genomic regions. To overcome this, VariationHunter  was developed by introducing a soft clustering strategy, which allows short reads in multiple clusters, aiming to improve sensitivity. It proposes two algorithms to implement soft clustering of the discordant reads: VariationHunter-SC (Set Cover) and VariationHunter-Pr (Probabilistic). Using maximum parsimony, VariationHunter-SC can minimize the total number of genomic clusters by mapping each paired-end read to a particular SV. VariationHunter-Pr was developed to solve the problems present when the same paired end reads are mapped to multiple locations with equal support. Additionally, VariationHunter was extended to detect transposon insertion events from NGS data recently .
Aside from the single sample-based PEM algorithm, a multiple samples-based PEM algorithm was also implemented in commonLAW , a software that compares all the samples to a reference genome simultaneously to obtain more accurate SVs. Similar to VariationHunter-SC, commonLAW implements a maximum parsimony approach to detect SVs/CNVs. In other words, commonLAW works like a multiple alignment tool if we consider VariationHunter as a pairwise alignment method in this analogy.
Split read-based approach
SR methods start from read pairs in which one read from each pair is aligned to the reference genome uniquely while the other one fails to map or only partially maps to the genome (Figure 1b). Those unmapped or partially mapped reads potentially provide accurate breaking points at the single base pair level for SVs/CNVs. SR methods split the incompletely mapped reads into multiple fragments. The first and last fragments of each split read are then aligned to the reference genome independently. This remapping step therefore provides the precise start and end positions of the insertion/deletion events. However, the SR-based approach heavily relies on the length of reads and is only applicable to the unique regions in the reference genome.
Pindel is the first SR-based method to identify breakpoints of large deletions (1 bp - 10 Kb) and medium-sized insertions (1 - 20 bps) . Starting with unmapped or partially mapped reads from read pairs, this tool utilizes a string match approach to search unique substrings in minimum (the 5' end of the input reads) and maximum locations (the 3' end of the input reads) in which both could be completely mapped to the genome. Based on the minimum and maximum location, the reads are split into either two (deletion) or three (short intra-read insertion fragment in the middle) segments. Another SR-based tool, Alignment with Gap Excision (AGE), identifies SV breakpoints with base pair accuracy using more strict alignments with gap excision . Using a local alignment algorithm similar to Smith-Waterman , AGE can simultaneously align two sequences accurately. Since this tool requires a predefined SV/CNV region as input to guide local alignment, it provides a complementary strategy to other tools in identifying SVs/CNVs. Similarly, the SR-based tool split-read identification, calibrated (SRiC), adopts gapped alignment to detect SVs . Lastly, another SR-based tool, SLOPE, was developed for targeted sequencing data of a limited number of genomic regions of interest  (Table 1).
Read depth-based approach
Mainly due to the accumulation of high-coverage NGS data, RD-based methods have recently become a major approach to estimate copy numbers. The underlying hypothesis of RD-based methods is that the depth of coverage in a genomic region is correlated with the copy number of the region, e.g., a gain of copy number should have a higher intensity than expected . Compared to PEM and SR-based tools, RD-based methods can detect the exact copy numbers, which the former approaches are lacking because PEM/SR methods only use the position information. Moreover, RD-based methods can detect large insertions and CNVs in complex genomic region classes, which are difficult to detect using PEM and SR methods .
Currently, WGS and WES are the two major NGS platforms for DNA-sequencing (DNA-seq). WGS can determine the full spectrum of variants in the whole genome. WES is a more effective, targeted sequencing of protein coding regions in the genome, often with high coverage. In this section, we review the RD-based tools for WGS and WES separately, as they utilize different technologies and produce different data. However, both platforms generate similar raw data in the format of short reads and involve the mapping of short reads to the reference genome. The mapping results are then used to calculate read depth in both WGS and WES.
Generally, RD-based tools can be classified into three categories depending on the study design: single samples, paired case/control samples, and a large population of samples. When there is only a single sample, it is not possible to borrow information from other samples or the matched control. Therefore, a widely applied strategy in such cases is to estimate the read depth distribution using mathematical models and to discover regions with abnormal depth departing from the overall distribution. The copy numbers reported in such cases are more likely to be absolute numbers of copies. When there are pairs of matched samples, CNV detection can be leveraged on the matched control, which could serve as a "reference" genome. The copy numbers in such cases are reported more like relative copies compared to the reference control, instead of absolute copies. Finally, to detect CNVs from a population of samples, tools are developed using the overall mean of the read depth from multiple samples to discover the discordant copy numbers in each sample.
Basically, RD-based methods follow a four-step procedure to discover CNVs: mapping, normalization, estimation of copy number, and segmentation (Figure 1c). In the mapping step, short reads are aligned to the reference genome, and the read depth is calculated according to the number of mapped reads in predefined windows. The second step focuses on normalization and correction of potential biases in read depths mainly caused by GC contents and repeat genomic regions, among others. With the normalized read depths, the third step is to estimate the accurate copy number along the chromosome to determine the gain or loss. Lastly, the genomic regions with a similar copy number are merged to detect discordant copy number regions . Notably, after mapping and normalization, the read depth data from NGS experiments are similar to the probe log ratios from arrayCGH data in the mathematical view. Therefore, some classical algorithms to detect CNV regions from arrayCGH data can be reused and modified to work on NGS read depth data.
Tools for whole genome sequencing data
Generally, RD-based tools define non-overlapping genomic windows, calculate read depths for these windows, and estimate copy numbers for each of them. Deeper read depth can not only increase the statistical power of CNV detection, but also help to infer the best size of a sliding window . A moving window with a fixed width is the simplest way to divide a chromosome . However, some genomic regions with a small read count may create a non-uniform fluctuant, which may result in high false positive detection. In CNAnorm, the authors suggested choosing the best size for sliding windows so that the number of reads mapped to each window is between 30 and 180 for WGS data . Based on a user-defined false discovery rate (FDR), ReadDepth and mrCaNaVar can automatically set an appropriate size for a sliding window according to the mean number of reads in each window. RDXplorer proposed a novel algorithm using significance testing called EWT (Event-Wise Testing), which provides a high resolution for CNV detection using 100-bp windows . A Z-score is calculated based on the number of reads mapped in each 100-bp window, according to a two-tailed normal distribution. Using a probabilistic approach, EWT was implemented to search unusual depth intervals with deletion and insertion events using corrected Z-scores. The tool SegSeq requires a predefined amount of reads mapped in a genomic window . Similar to SegSeq, recursive Smith-Waterman-seq (rSW-seq) was developed for CNV detection in matched tumor-control samples . Different from SegSeq, rSW-seq does not rely on predefined genomic windows. It first sorts reads according their genomics regions. Next, positive and negative weight values are assigned to tumor and control samples, respectively. The cumulative changes of assigned weight values are used to determine breakpoints that indicate copy number gains or losses.
Another key challenge for RD-based CNV detection is an efficient algorithm to segment the detected CNV regions after choosing the appropriate sliding window. Earlier methods such as mrCaNaVar and CNV-seq are implemented in a model-free style [31, 35, 36]. After obtaining the read depth, mrCaNaVar directly localizes segments with significantly different read depths to the overall read depth mean. The segments with three times the standard deviation over the mean of read depth are determined as duplication events. Similar criteria are applied to deletions. However, this model-free approach tends to have a high FDR even after correcting for GC-content and mapping biases.
Most of the methods use statistical models, such as Circular Binary Segmentation (CBS), Mean Shift-Based (MSB), Shifting Level Model (SLM), Expectation Maximization (EM), and Hidden Markov Model (HMM) for CNV detection (Table 2). These models assume that there is a sequence of segments (gain or loss) along each chromosome that could be detected using probability distribution functions with statistical measurements, such as the mean and variances of read depths. Compared with mode-free approach, approaches using mathematical models to detect CNVs generate more reliable results after filtering false positive regions.
The CBS method was originally developed for arrayCGH data to convert the noisy intensities into segments with equal copy numbers. Essentially, the algorithm recursively localizes the breakpoints by changing genomic positions until the chromosomes are divided into segments with equal copy numbers that are significantly different from the copy numbers from their adjacent genomic regions . The first tool to introduce CBS in NGS data is SegSeq , which uses a case (e.g., cancer) versus control dataset to detect CNVs. At first, the copy number ratio of the two matched samples is calculated for each genomic window. Using the CBS algorithm, SegSeq identifies the boundaries of DNA fragments with changed copy numbers. ReadDepth is another RD-based CNV calling tool that uses the CBS approach to segment chromosomes . However, it does not require paired samples; rather, it can be applied to a single sample. Notably, ReadDepth is able to discover epigenetic changes due to its capacity to process both regular and bisulfite-treated reads .
MSB  is another widely applied algorithm that was originally developed for CNV segmentation in arrayCGH data. In the MSB approach, the adjacent genomic windows with similar read depths are merged together along chromosomes. The breakpoints are reported when read depths of a sliding window are significantly discordant with the depths of the merged windows. Two software tools for NGS data, CNVnator  and BIC-seq , implement the MSB approach to choose the window size for read depth signal partitioning. CNVnator works on single individual samples to implement MSB. The key feature of CNVnator is its capability to detect CNVs in various sizes ranging from a few hundred bases to megabases. In contrast, BIC-seq, which also implements the MSB approach, works on matched pair samples instead of single samples .
CNVeM is a recently developed tool to determine CNVs from individual samples with all possible mapping information . It utilizes mrsFAST to map reads, allowing up to 2 mismatches . Those ambiguously mapped reads within each 300-bp window are used to estimate copy numbers using an EM algorithm. Instead of searching CNVs in a genomic region, the copy number of each base is calculated based on its mapped reads. Therefore, this algorithm can predict breakpoints in base pair resolution. Due to its ambiguous mapping strategy, CNVeM is also useful to distinguish the gain or loss for human paralog regions.
CNAseg is a representative tool utilizing HMM in the segmentation step to identify contiguous genomic regions with similar read depths . The key feature of CNAseg is its powerful framework to control the false positive rate using read depth variability between different flow cells from case and control samples. JointSLM is another tool clustering the small abnormal genomic regions using an HMM approach . However, JointSLM does not rely on single or paired samples but builds a statistical model based on multiple samples. Taking advantage of the simultaneous analyses of multiple samples, JointSLM works better than the EWT algorithm implemented in RDXplorer  in terms of detecting small CNV regions as short as 500 bp. In addition, higher sensitivity can be achieved by JointSLM with more samples .
JointSLM is not the only tool to utilize multiple samples to detect CNVs. Two additional tools, cn.MOPS and correlation matrix diagonal segmentation (CMDS), are proposed to increase statistical power and decrease computational burden based on a multiple samples approach. cn.MOPS proposed a data processing pipeline using a mixture of Poisson models to reduce the FDR in CNV detection . By modelling read depths across multiple samples in each genomic region, this tool can remove the effect of read depth variation along chromosomes. Another multiple sample-based tool, CMDS, was developed to identify CNVs using a between-chromosomal-site correlation analysis . CMDS initially takes all copy numbers from the same genomic regions from multiple samples as input. The correlation coefficients between two different genomic regions across multiple samples are calculated. The diagonal transformation and statistical tests are applied on the correlation coefficients to identify focal events (CNVs). Though CNVnator also claims its ability to discover CNVs from families and populations, it calls CNVs in each sample before the cross-sample comparison . Compared with other tools, multiple sample-based tools often gain high sensitivity and a lower false positive rate .
The different RD-based tools have their own limitations and advantages. In a recent comparison of different RD-based tools , EWT-based methods (e.g., RDxplorer) and SLM-based methods (e.g., JointSLM) showed the best performance with both high sensitivity and specificity. For large windows over 50 bp, the CBS-based tool (ReadDepth) and MSB-based tool (CNVnator) performed well, but their accuracy was decreased when dealing with smaller windows less than 10 bp. The HMM based tool, CNASeg, could obtain better CNV calling on high coverage data but performs worse on low coverage data. In terms of the size of detected CNVs, RDxplorer could detect CNVs with a length of ~500 bp. However, ReadDepth and CNVnator can identify larger events ranging between 2-5 Kb.
Tools for CNV detection through WES data
Due to investigators' interest in coding gene regions and its lower cost, WES is widely applied to Mendelian as well as other complex diseases . WES has the potential to rapidly detect single nucleotide mutations and CNVs in human coding regions. The inherent limitation of WES is that it produced reads only covering 1-2% of the human genome based on the currently available exome extraction kits. The full spectrum of CNVs and breakpoints may not be completely characterized. In addition, many large CNVs and cross-chromosome events may not be detected. Nevertheless, CNV detection based on WES data may give a quick insight into CNV patterns for a specific disease or phenotype. The analysis tools reviewed below may help establish rational pipelines to identify various CNVs in matched case-control pairs, multiple disease samples, and common CNVs in large populations.
In contrast to WGS, WES data have higher depth for targeted regions, which is ideal for more accurate CNVs using an RD-based calling approach. However, due to differing capture efficiency, the depth from different genomic regions may vary substantially and should be considered in the downstream analysis of CNV calling. First, due to inconsistent capture efficiency, there might be regions that are poorly sequenced, which requires pre-processing for WES data. Second, the assumption of normal distribution may no longer be valid due to the biases regarding read depth distribution. Third, due to the discontinuation of genomic regions, most CNV breakpoints could not be detected. Finally, the widely applied segmentation algorithms to merge windows in WGS may not be applicable in WES data due to the non-continuous distributions of the reads.
To keep pace with the flood of WES data, at least 11 RD-based CNV calling tools have been developed to address the unique challenges of WES data (Table 3). Similar to the methods for WGS data, the methods for WES data may also rely on matched case and control samples or single samples to detect discordant CNV profiles . ExomeCNV focuses on regions whose coverage is higher than a predefined minimum depth and assumes Gaussian distribution to characterize the read depth from WES. Similarly, VarScan 2 requires a minimum coverage of 3× for each captured region from WES data to reduce potential false positives . Based on the relative CNVs between the case and control samples, VarScan 2 adopts a CBS algorithm implemented by DNAcopy  for segmentation. Unlike directly modelling the log-ratios of the coverage between case and control samples for each region, the tool PropSeq implements a regression model to represent the read depth of a given region in the case sample as a function of its counterpart in the control sample .
Recently, in the work of ExomeDepth, the authors argued that the Gaussian assumptions may not completely characterize the read count from WES data due to the technical noise from library preparation, capture, and sequencing. Rather, ExomeDepth shows that the beta-binomial model could fit WES data better than the Gaussian model. In addition, because the ExomeDepth does not rely on the CNVs in the reference sample, it is useful in the detection of rare CNVs . Another flexible tool, Control-FREEC, is able to call CNVs from WGS and WES data with or without control samples . An additional convenience of Control-FREEC is its acceptance of almost all available formats currently used in NGS data rather than being limited to SAM/BAM files.
All the above CNV detection methods are designed to detect CNVs from NGS data for small-scale samples, paired or not, and are especially prevalent for cancer and normal pairs. Another scenario in WES data is separate, large-scale sequencing of a disease population and a control population. Similar to the trends in CNV calling for WGS data, more and more tools utilize multiple samples with WES data to increase sensitivity and reduce the FDR. These tools, such as CoNIFER (copy number inference from exome reads), XHMM (exome hidden Markov model), and ExoCNVTest, fit particularly well for high-resolution and large population-based studies to detect rare and common CNVs. CoNIFER was the first developed to deal with rare CNVs from multiple samples . It starts by calculating the average signal intensity for each captured region in each of the samples. Using the singular value decomposition (SVD) approach, CoNIFER normalized absolute copy numbers from each sample to discover rare CNVs. At almost the same time as CoNIFER was released, another group reported the tool XHMM, which adopts a similar concept to normalize the raw read depth data but uses principal-component analysis (PCA) . Following the recalibration of read depth, XHMM implements an HMM model to find the significant discordant CNVs. Both tools utilize a large sample size to detect CNVs while reducing potential technical noises from exome capture reactions and batch effects by removing the major variances detected in SVD or PCA. The main advantages of these two methods include that they are control-free and can be applied to large disease cohorts. To date, ExoCNVTest is the only tool to specifically detect common CNVs from WES data . After normalizing the read depth, ExoCNVTest calculates the local first principal component (FPC) on the read depth for each sample. Then, the local FPCs are used to discover the global principal components, which may record the CNV information. The association between global principal components and particular genomic regions is carried out to detect the absolute copy number genotypes with case-control samples.
Different from the RD, PEM and SR approaches that first align NGS reads to a known reference genome before the detection of CNVs, the AS-based methods first reconstruct DNA fragments, i.e., contigs, from short reads by assembling overlapping reads. By comparing the assembled contigs to the reference genome, the genomic regions with discordant copy numbers are then identified (Figure 1d). This direct assembly of short reads without using a reference is called de novo assembly. Assembly can also use a reference genome as a guide to improve its computational efficiency and contig quality. AS-based methods have a minimum read coverage requirement to satisfy the algorithm and detect overlapping fragments, although high coverage will increase the complexity of short read assembly, especially for de novo AS-based tools.
The Cortex assembler was designed as a complete de novo CNV calling method . It uses de Bruijn graphs to assemble overlapped reads from multiple sequencing samples into a single graph. The nodes and edges in the graph are marked in different colors to differentiate different samples. Then, it discovers CNVs by piping all the nodes from different samples together to find bifurcation structures, in which the sequence of the sample differs from the sequence of the reference genome. The branches to separate different colors represent SVs, e.g., insertions and deletions. Another de novo AS-based tool, Magnolya, that was published recently, also estimates copy numbers from two or more samples by utilizing a Poisson mixture model .
Without relying on read alignment, the AS methods potentially provide an unbiased approach to discover novel genetic variants ranging from the single base pair level to large structure variation. However, AS methods are rarely used in CNV detection in non-human eukaryotic genomes due to the low quality of the assembled contigs and their overwhelming demand on computational resources.
The combinatorial approach
Although there has been great progress in each of the four categories of PEM, SR, RD, and AS approaches, and a growing number of tools have been developed, none of them have been able to detect the full spectrum of all types of CNVs with high sensitivity and specificity. With distinct advantages and weaknesses, these four strategies could be complementary to each other. PEM-based methods can detect all types of SVs (e.g., deletion, novel insertion, translocation, inversion, interspersed and tandem duplication) , especially for deletions of less than 1 Kb. However, PEM tools cannot accurately estimate the actual numbers of copies, and they are not applicable to insertions larger than the average insert size of the library. In contrast, RD-based methods can accurately predict copy numbers, have good performance to detect large CNVs, and are applicable to both WGS and WES data. However, they are not applicable for detection of precise breakpoints, copy neutral events like inversions or translocations, or small CNVs (e.g., <1 Kb). For the AS-based tools, their exclusive advantage lies in that they do not require a reference genome as input, and, importantly, they allow the discovery of novel mutation sequences. However, AS-based tools require extensive computation and perform poorly on repeat and duplication regions. For the SR-based tools, they could detect breakpoints at base pair resolution and performed well on deletion and small insertions. However, SR-based tools have low sensitivity in regions with low-complexity, as they rely on unique mapping information to the genome. To overcome the inherent limitations and take advantage of different types of methods, a number of tools have been developed attempting to integrate multiple approaches from different categories to increase the performance in detecting CNVs and reduce false positive discoveries.
Five tools in Table 4 are able to use PEM and RD information to identify CNVs in populations, including SVDetect , CNVer , Genome STRiP , GASVPro , and inGAP-sv . As aforementioned, combinatorial methods can take advantage of the uniqueness of multiple tools and, thus, can reduce more false positives than any method that is purely built on PEM or RD (Figure 1e). For example, PEM methods can report accurate breakpoints, but their efficiency is low when detecting large CNV regions (e.g., insertions longer than the insert size) or counting exact copy numbers. On the other hand, RD methods are advantageous to detect large CNVs but cannot report exact breakpoints. A combination of the two would enable accurate detection of CNV regions with exact breakpoints and spanning various lengths, especially for longer insertions that are undetectable by PEM. In addition, the independent identification of CNV regions helps the tools to reduce false positives.
SVDetect is the first tool to introduce read depth information to improve PEM-based detection of large SVs. Combining the breakpoint predicted by the PEM approach and the log-ratio from the read depth between the matched case and control samples, the tool makes final judgements on gain and loss events. CNVer is another tool that combines PEM and read depth information for CNV identification . Similar to SVDetect, CNVer makes use of discordant read pair mapping information to identify breakpoints and read depth signals to identify aberrant genomic fragments. Notably, an additional advantage of CNVer is that it employed an ambiguous mapping strategy to allow better use of the reads that are mapped to multiple genomic locations. In this way, higher sensitivities in repeat and duplication regions can be achieved by CNVer. Genome STRiP is used to identify SVs/CNVs in large populations such as the date generated in The 1000 Genome Project . To reduce false positives in deletion identification, the tool requires that the deletion identified by the PEM approach should also be confirmed by the lower read depth in the same genomic regions.
The GASV-Pro is another successful example of a combinatorial method. It was upgraded from a single PEM-based tool, GASV (Geometric Analysis of Structural Variants), and incorporated read depth information to combinatorially call CNVs. GASV proposed a novel geometric approach to cluster the discordant mappings from paired reads . In this approach, the breakpoint region is represented as a polygon in a plane. This tool can further identify the accurate boundary supporting the same variant using an efficient plane sweep algorithm to detect the intersections of the polygons. The advantage of GASV is that it can integrate both arrayCGH and paired-end sequencing reads to call SVs, including CNVs. Different from GASV, GASVPro implements a probabilistic model to integrate both read pairs and read depth information . Using a Markov Chain Monte Carlo (MCMC) approach, GASVPro can map all the reads to all possible locations. The breakend read depth (BeRD) information is defined as the dropped read depth at the breakpoints. Incorporating the BeRD signals, GASVPro uses the amount of discordant read pairs to determine the probability of each detected CNV. Another combinatorial tool, inGAP-SV, can not only detect CNVs by combining read pairs and read depth information but also provide a genome browser-like visualization solution for large and complex SVs . Similar to the beRD concept in GASVPro, inGAP-SV first uses the drops in read depth as signals to defined SV "hotspots." Then, a heuristic search approach is applied on these hotspots to classify SVs using the PEM approach. In summary, there are at least two key features for inGAP-SV: capacity to identify different types of CNVs and customized visualization.
NovelSeq is the first tool to combine PEM and SR approaches and was developed specifically to detect novel insertions in the genome . It starts with the unmapped reads and assembles them to form contigs. A clustering algorithm is then applied to the assembled contigs to identify potential insertion events supported by these unmapped reads. HYDRA is a tool that combines PEM and local de novo assembly to detect SVs and was originally applied in the mouse genome. It also relies on discordant mate pairs for SV detection, using both short and long reads as input . HYDRA applies a soft clustering approach to identify putative SVs and breakpoints and therefore has the potential to be very useful to identify SVs in repeat and duplication regions in the human genome.
Challenges and perspectives
Intensive efforts in recent years to develop computational tools for CNV detection have resulted in 48 tools, as described in this paper, which target specific ranges of CNVs by utilizing different computational strategies. Systematic comparison showed that only small overlapping SVs/CNVs are identified based on different tools in population-scale genome sequencing . These small overlaps imply there may be different scales of SVs/CNVs in the human genome. Recently, distinct SV patterns were observed in experiments that used two libraries with different insert sizes . This observation indicates the importance of choosing appropriate tools for a specific experimental design.
Despite improvements to NGS technologies and CNV-detecting tools, the identification of low coverage CNVs still remains a challenge. Although the RD-based approach has to correct distortions caused by NGS biases, the relationship between the read count and true copy number can be distorted by several effects. The PCR process is known to be one major cause of this distortion, where genomic fragments with a lower PCR amplification rate often result in less reads [69, 70]. Furthermore, sequencing process can also introduce system noise [55, 71]. It was reported that NGS has lower sequencing coverage in regulatory regions . In particular, the capability of exome capture in the library preparation process complicates the connection between true copy number and read count for WES data. All these factors contribute to the difficulty in detecting the low coverage copy number. For the regions with a low read count, it is hard to articulate if it is indeed due to the low copy number or if it is a consequence of the aforementioned processes. In the following section, we focus on the RD-based strategy to discuss potential noise from NGS data and how to overcome these shortcomings using specific normalization. Since most of the PEM- and SR-based tools described here are general tools to detect SVs, their limitations and strengths were extensively reviewed previously .
One well-investigated bias in RD-based methods is related to GC content, the percentage of guanine and cytosine bases in a genomic region. GC content varies markedly along the human genome and across species and has been found to influence read coverage on most sequencing platforms. In a recent study of the Illumina HiSeq and Genome Analyzer systems , for example, a positive correlation between read coverage and GC content was observed when GC percentage is within the spectrum of 24 to 47%. To correct the GC bias, a widely used strategy in existing RD-based tools is to partition the genome into windows/bins of either fixed or various sizes [28, 31, 41]. The number of reads mapped to each window are counted and then normalized based on the GC content of this windowed region. The adjusted read count is then used as an indicator of the copy number of the corresponding genomic region. In NGS applications that involve a pair of disease and normal tissues taken from the same patient and sequenced on the same platform, the effect of GC bias can be effectively cancelled out by comparing the paired genomes directly at each region [31, 33, 41], thereby taking advantage of the similarity of GC bias between the two paired samples. For instance, CNAseg implemented an algorithm locally weighted scatterplot smoothing (LOWESS) regression to adjust the GC bias for matched samples in each 10-Kb window. Moreover, Discrete Wavelet Transform (DWT) is used to smooth the signals of mapped reads, which helps to obtain more accurate estimation of relative copy numbers between case and control samples.
Another major bias that affects CNV calling when using the RD-based approach originates from read alignment. In the alignment step, a significant portion of reads are mapped to multiple positions due to a short read length and the presence of repetitive regions in the reference genome. A variety of strategies have been proposed to assign ambiguously mapped reads to the reference genome, and each suggested approach has its own pros and cons, affecting read coverage and, hence, copy number in a different way. Some methods, e.g., BIC-seq  and SegSeq , completely ignore ambiguously mapped reads. Apparently, by counting only the uniquely mapped reads, these methods are not applicable to detect CNVs in homologous genomic regions. Other methods, e.g., CNVnator , randomly assign an ambiguous read to one of all possible positions to which this read is mapped by the aligner. Though capable of identifying copy numbers in repetitive genomic regions, this strategy suffers from false positive detection as a consequence of the random placement of ambiguously mapped reads. One possible direction to improve mapping sensitivity is to employ a soft clustering approach, which allows multiple good mappings . Despite these improvements on mapping algorithms, all of the methods described here are still problematic in repeat regions. Allowing mismatches in repeat regions may increase sensitivity in the future.
To assist researchers in selecting the proper NGS tool(s) for their CNV studies, here, we summarize our discussion of SV-detecting tools collected in this paper. Several tools, such as JointSLM and ExoCNVTest, are applicable to detect common CNVs shared among multiple samples, while CoNIFER and XHMM are good for rare CNV identification from a population of WES samples. Additionally, Pindel is an effective tool for the detection of small deletions of lengths less than 300 bp, while RD-based tools are more suitable to characterize large CNV events. As for CNV breakpoints, SVseq is designed to identify deletion with exact breakpoint information from low coverage NGS data . NovelSeq is specific to locate novel insertion. The tool inGAP-SV facilitates researchers to visualize the detected interesting CNV regions.
Compared with CNV detection using only single tool, the combination of different software has proved effective in improving CNV prediction accuracy. For example, a study that combined the results of BreakDancer and VariationHunter demonstrated higher sensitivity and specificity to detect SVs/CNVs . This result implies that the combination of different PEM tools can help detect high confidence CNVs. In Table 4, we show that PEM-based tools are often combined with RD and AS approaches to improve both sensitivity and specificity for CNV detection using WGS data. Unlike WGS data, the discontinuous reads from WES limit the application of PEM-based methods. For example, the insert size for pair-end reads in WES may not be long enough to detect CNV using PEM-based methods. However, the SR-based tool SLOPE has good performance to pinpoint translocation events by focusing on small, targeted genomic regions. By recognizing the successful application of a combinatorial approach to detect CNVs from WGS data, the similar strategy to combine RD and SR features may provide more reliable CNV predictions from WES data . In summary, incorporating various signatures to support the same CNV can dramatically improve both the statistical power and resolution of CNV size and breakpoints compared to each different strategy individually.
Competing for higher sequencing throughput, data accuracy and longer read lengths, today's sequencing technologies will continue to advance rapidly. One prominent development is the creating of third generation sequencing (TGS) technologies, which promise to provide dramatically longer read lengths. The first commercial TGS sequencer launched by PacBio (PacBio RS single molecule sequencing) can now produce reads of an average length 1300 bp, significantly longer than that of any NGS platforms today. Longer reads will greatly ease read alignment and CNV detection in repetitive regions of genome [75, 76]. These longer reads will significantly reduce mapping errors due to incorrect sequencing. The increased size of the short read will also strengthen the statistical power of the RD methods. In addition, all these improvements on mapping quality will benefit PEM methods to avoid false positives caused by chimera in the genome. For the SR methods, these longer reads may also ameliorate their inherent limitation of only detecting insertion events shorter than the average read length. Additionally, longer reads may improve assembly quality when implementing AS methods. The development of high resolution CNV detection tools is an iterative process of integrating new technologies' features and novel computational algorithms. Though these processes require committed efforts from industry-academy collaborative work, the fruitful understanding of human CNVs in healthy and disease conditions will be a worthy endeavor, which may not only allow discrimination of driver mutations for pathogenesis of Mendelian and complex diseases but also help to develop personalized medicine in the future.
de novo assembly
Binary version of a SAM file
Circular binary segmentation
Comparative genome hybridization
Copy number variation
Nucleotide sequence with its corresponding quality scores in text format
Fluorescence in situ hybridization
Hidden Markov model
Insertion and deletion
Mean shift-based method
R language for statistical analysis
Sequence Alignment/Map file format
Shifting level model
Alkan C, Coe BP, Eichler EE: Genome structural variation discovery and genotyping. Nat Rev Genet. 2011, 12: 363-376. 10.1038/nrg2958.
Freeman JL, Perry GH, Feuk L, Redon R, McCarroll SA, Altshuler DM, Aburatani H, Jones KW, Tyler-Smith C, Hurles ME: Copy number variation: new insights in genome diversity. Genome Res. 2006, 16: 949-961. 10.1101/gr.3677206.
Sebat J, Lakshmi B, Troge J, Alexander J, Young J, Lundin P, Maner S, Massa H, Walker M, Chi M: Large-scale copy number polymorphism in the human genome. Science. 2004, 305: 525-528. 10.1126/science.1098918.
Iafrate AJ, Feuk L, Rivera MN, Listewnik ML, Donahoe PK, Qi Y, Scherer SW, Lee C: Detection of large-scale variation in the human genome. Nat Genet. 2004, 36: 949-951. 10.1038/ng1416.
Redon R, Ishikawa S, Fitch KR, Feuk L, Perry GH, Andrews TD, Fiegler H, Shapero MH, Carson AR, Chen W: Global variation in copy number in the human genome. Nature. 2006, 444: 444-454. 10.1038/nature05329.
Fiegler H, Redon R, Andrews D, Scott C, Andrews R, Carder C, Clark R, Dovey O, Ellis P, Feuk L: Accurate and reliable high-throughput detection of copy number variation in the human genome. Genome Res. 2006, 16: 1566-1574. 10.1101/gr.5630906.
Beckmann JS, Estivill X, Antonarakis SE: Copy number variants and genetic traits: closer to the resolution of phenotypic to genotypic variability. Nat Rev Genet. 2007, 8: 639-646. 10.1038/nrg2149.
Beroukhim R, Mermel CH, Porter D, Wei G, Raychaudhuri S, Donovan J, Barretina J, Boehm JS, Dobson J, Urashima M: The landscape of somatic copy-number alteration across human cancers. Nature. 2010, 463: 899-905. 10.1038/nature08822.
Buysse K, Delle Chiaie B, Van Coster R, Loeys B, De Paepe A, Mortier G, Speleman F, Menten B: Challenges for CNV interpretation in clinical molecular karyotyping: lessons learned from a 1001 sample experience. Eur J Med Genet. 2009, 52: 398-403. 10.1016/j.ejmg.2009.09.002.
Carter NP: Methods and strategies for analyzing copy number variation using DNA microarrays. Nat Genet. 2007, 39: S16-S21. 10.1038/ng2028.
Snijders AM, Nowak N, Segraves R, Blackwood S, Brown N, Conroy J, Hamilton G, Hindle AK, Huey B, Kimura K: Assembly of microarrays for genome-wide measurement of DNA copy number. Nat Genet. 2001, 29: 263-264. 10.1038/ng754.
Shendure J, Ji H: Next-generation DNA sequencing. Nat Biotechnol. 2008, 26: 1135-1145. 10.1038/nbt1486.
Metzker ML: Sequencing technologies - the next generation. Nat Rev Genet. 2010, 11: 31-46. 10.1038/nrg2626.
Meyerson M, Gabriel S, Getz G: Advances in understanding cancer genomes through second-generation sequencing. Nat Rev Genet. 2010, 11: 685-696. 10.1038/nrg2841.
Korbel JO, Urban AE, Affourtit JP, Godwin B, Grubert F, Simons JF, Kim PM, Palejev D, Carriero NJ, Du L: Paired-end mapping reveals extensive structural variation in the human genome. Science. 2007, 318: 420-426. 10.1126/science.1149504.
Medvedev P, Stanciu M, Brudno M: Computational methods for discovering structural variation with next-generation sequencing. Nat Methods. 2009, 6: S13-20. 10.1038/nmeth.1374.
Chen K, Wallis JW, McLellan MD, Larson DE, Kalicki JM, Pohl CS, McGrath SD, Wendl MC, Zhang QY, Locke DP: BreakDancer: an algorithm for high-resolution mapping of genomic structural variation. Nat Methods. 2009, 6: 677-681. 10.1038/nmeth.1363.
Korbel JO, Abyzov A, Mu XJ, Carriero N, Cayting P, Zhang ZD, Snyder M, Gerstein MB: PEMer: a computational framework with simulation-based error models for inferring genomic structural variants from massive paired-end sequencing data. Genome Biol. 2009, 10: R23-10.1186/gb-2009-10-2-r23.
Hormozdiari F, Alkan C, Eichler EE, Sahinalp SC: Combinatorial algorithms for structural variation detection in high-throughput sequenced genomes. Genome Res. 2009, 19: 1270-1278. 10.1101/gr.088633.108.
Hormozdiari F, Hajirasouliha I, Dao P, Hach F, Yorukoglu D, Alkan C, Eichler EE, Sahinalp SC: Next-generation VariationHunter: combinatorial algorithms for transposon insertion discovery. Bioinformatics. 2010, 26: i350-357. 10.1093/bioinformatics/btq216.
Hormozdiari F, Hajirasouliha I, McPherson A, Eichler EE, Sahinalp SC: Simultaneous structural variation discovery among multiple paired-end sequenced genomes. Genome Res. 2011, 21: 2203-2212. 10.1101/gr.120501.111.
Ye K, Schulz MH, Long Q, Apweiler R, Ning Z: Pindel: a pattern growth approach to detect break points of large deletions and medium sized insertions from paired-end short reads. Bioinformatics. 2009, 25: 2865-2871. 10.1093/bioinformatics/btp394.
Abyzov A, Gerstein M: AGE: defining breakpoints of genomic structural variants at single-nucleotide resolution, through optimal alignments with gap excision. Bioinformatics. 2011, 27: 595-603. 10.1093/bioinformatics/btq713.
Smith TF, Waterman MS: Identification of common molecular subsequences. J Mol Biol. 1981, 147: 195-197. 10.1016/0022-2836(81)90087-5.
Zhang ZD, Du J, Lam H, Abyzov A, Urban AE, Snyder M, Gerstein M: Identification of genomic indels and structural variations using split reads. BMC Genomics. 2011, 12: 375-10.1186/1471-2164-12-375.
Abel HJ, Duncavage EJ, Becker N, Armstrong JR, Magrini VJ, Pfeifer JD: SLOPE: a quick and accurate method for locating non-SNP structural variation from targeted next-generation sequence data. Bioinformatics. 2010, 26: 2684-2688. 10.1093/bioinformatics/btq528.
Teo SM, Pawitan Y, Ku CS, Chia KS, Salim A: Statistical challenges associated with detecting copy number variations with next-generation sequencing. Bioinformatics. 2012, 28: 2711-2718. 10.1093/bioinformatics/bts535.
Yoon S, Xuan Z, Makarov V, Ye K, Sebat J: Sensitive and accurate detection of copy number variants using read depth of coverage. Genome Res. 2009, 19: 1586-1592. 10.1101/gr.092981.109.
Magi A, Tattini L, Pippucci T, Torricelli F, Benelli M: Read count approach for DNA copy number variants detection. Bioinformatics. 2012, 28: 470-478. 10.1093/bioinformatics/btr707.
Plagnol V, Curtis J, Epstein M, Mok KY, Stebbings E, Grigoriadou S, Wood NW, Hambleton S, Burns SO, Thrasher AJ: A robust model for read count data in exome sequencing experiments and implications for copy number variant calling. Bioinformatics. 2012, 28: 2747-2754. 10.1093/bioinformatics/bts526.
Xie C, Tammi MT: CNV-seq, a new method to detect copy number variation using high-throughput sequencing. BMC Bioinformatics. 2009, 10: 80-10.1186/1471-2105-10-80.
Gusnanto A, Wood HM, Pawitan Y, Rabbitts P, Berri S: Correcting for cancer genome size and tumour cell content enables better estimation of copy number alterations from next-generation sequence data. Bioinformatics. 2012, 28: 40-47. 10.1093/bioinformatics/btr593.
Chiang DY, Getz G, Jaffe DB, O'Kelly MJ, Zhao X, Carter SL, Russ C, Nusbaum C, Meyerson M, Lander ES: High-resolution mapping of copy-number alterations with massively parallel sequencing. Nat Methods. 2009, 6: 99-103. 10.1038/nmeth.1276.
Kim TM, Luquette LJ, Xi R, Park PJ: rSW-seq: algorithm for detection of copy number alterations in deep sequencing data. BMC Bioinformatics. 2010, 11: 432-10.1186/1471-2105-11-432.
Alkan C, Kidd JM, Marques-Bonet T, Aksay G, Antonacci F, Hormozdiari F, Kitzman JO, Baker C, Malig M, Mutlu O: Personalized copy number and segmental duplication maps using next-generation sequencing. Nat Genet. 2009, 41: 1061-1067. 10.1038/ng.437.
Campbell PJ, Stephens PJ, Pleasance ED, O'Meara S, Li H, Santarius T, Stebbings LA, Leroy C, Edkins S, Hardy C: Identification of somatically acquired rearrangements in cancer using genome-wide massively parallel paired-end sequencing. Nat Genet. 2008, 40: 722-729. 10.1038/ng.128.
Olshen AB, Venkatraman ES, Lucito R, Wigler M: Circular binary segmentation for the analysis of array-based DNA copy number data. Biostatistics. 2004, 5: 557-572. 10.1093/biostatistics/kxh008.
Miller CA, Hampton O, Coarfa C, Milosavljevic A: ReadDepth: a parallel R package for detecting copy number alterations from short sequencing reads. PLoS One. 2011, 6: e16327-10.1371/journal.pone.0016327.
Wang LY, Abyzov A, Korbel JO, Snyder M, Gerstein M: MSB: a mean-shift-based approach for the analysis of structural variation in the genome. Genome Res. 2009, 19: 106-117.
Abyzov A, Urban AE, Snyder M, Gerstein M: CNVnator: an approach to discover, genotype, and characterize typical and atypical CNVs from family and population genome sequencing. Genome Res. 2011, 21: 974-984. 10.1101/gr.114876.110.
Xi R, Hadjipanayis AG, Luquette LJ, Kim TM, Lee E, Zhang J, Johnson MD, Muzny DM, Wheeler DA, Gibbs RA: Copy number variation detection in whole-genome sequencing data using the Bayesian information criterion. Proc Natl Acad Sci USA. 2011, 108: E1128-1136. 10.1073/pnas.1110574108.
Wang Z, Hormozdiari F, Yang W-Y, Halperin E, Eskin E: CNVeM: Copy Number Variation Detection Using Uncertainty of Read Mapping. Research in Computational Molecular Biology. Edited by: Chor B: Springer Berlin/Heidelberg. 2012, 7262: 326-340. 10.1007/978-3-642-29627-7_34. Lecture Notes in Computer Science
Hach F, Hormozdiari F, Alkan C, Birol I, Eichler EE, Sahinalp SC: mrsFAST: a cache-oblivious algorithm for short-read mapping. Nat Methods. 2010, 7: 576-577. 10.1038/nmeth0810-576.
Ivakhno S, Royce T, Cox AJ, Evers DJ, Cheetham RK, Tavare S: CNAseg--a novel framework for identification of copy number changes in cancer from second-generation sequencing data. Bioinformatics. 2010, 26: 3051-3058. 10.1093/bioinformatics/btq587.
Magi A, Benelli M, Yoon S, Roviello F, Torricelli F: Detecting common copy number variants in high-throughput sequencing data by using JointSLM algorithm. Nucleic Acids Res. 2011, 39: e65-10.1093/nar/gkr068.
Klambauer G, Schwarzbauer K, Mayr A, Clevert DA, Mitterecker A, Bodenhofer U, Hochreiter S: cn.MOPS: mixture of Poissons for discovering copy number variations in next-generation sequencing data with a low false discovery rate. Nucleic Acids Res. 2012, 40: e69-10.1093/nar/gks003.
Zhang Q, Ding L, Larson DE, Koboldt DC, McLellan MD, Chen K, Shi X, Kraja A, Mardis ER, Wilson RK: CMDS: a population-based method for identifying recurrent DNA copy number aberrations in cancer from high-resolution data. Bioinformatics. 2010, 26: 464-469. 10.1093/bioinformatics/btp708.
Ng SB, Buckingham KJ, Lee C, Bigham AW, Tabor HK, Dent KM, Huff CD, Shannon PT, Jabs EW, Nickerson DA: Exome sequencing identifies the cause of a mendelian disorder. Nat Genet. 2010, 42: 30-35. 10.1038/ng.499.
Sathirapongsasuti JF, Lee H, Horst BA, Brunner G, Cochran AJ, Binder S, Quackenbush J, Nelson SF: Exome sequencing-based copy-number variation and loss of heterozygosity detection: ExomeCNV. Bioinformatics. 2011, 27: 2648-2654. 10.1093/bioinformatics/btr462.
Koboldt DC, Zhang Q, Larson DE, Shen D, McLellan MD, Lin L, Miller CA, Mardis ER, Ding L, Wilson RK: VarScan 2: somatic mutation and copy number alteration discovery in cancer by exome sequencing. Genome Res. 2012, 22: 568-576. 10.1101/gr.129684.111.
Seshan VE OA: DNAcopy: A package for analyzing DNA copy data. 2010, [http://www.bioconductor.org/packages/2.12/bioc/html/DNAcopy.html]
Rigaill GJ, Cadot S, Kluin RJ, Xue Z, Bernards R, Majewski IJ, Wessels LF: A regression model for estimating DNA copy number applied to capture sequencing data. Bioinformatics. 2012, 28: 2357-2365. 10.1093/bioinformatics/bts448.
Boeva V, Zinovyev A, Bleakley K, Vert JP, Janoueix-Lerosey I, Delattre O, Barillot E: Control-free calling of copy number alterations in deep-sequencing data using GC-content normalization. Bioinformatics. 2011, 27: 268-269. 10.1093/bioinformatics/btq635.
Krumm N, Sudmant PH, Ko A, O'Roak BJ, Malig M, Coe BP, Quinlan AR, Nickerson DA, Eichler EE: Copy number variation detection and genotyping from exome sequence data. Genome Res. 2012, 22: 1525-1532. 10.1101/gr.138115.112.
Fromer M, Moran JL, Chambert K, Banks E, Bergen SE, Ruderfer DM, Handsaker RE, McCarroll SA, O'Donovan MC, Owen MJ: Discovery and Statistical Genotyping of Copy-Number Variation from Whole-Exome Sequencing Depth. Am J Hum Genet. 2012, 91: 597-607. 10.1016/j.ajhg.2012.08.005.
Coin LJ, Cao D, Ren J, Zuo X, Sun L, Yang S, Zhang X, Cui Y, Li Y, Jin X, Wang J: An exome sequencing pipeline for identifying and genotyping common CNVs associated with disease with application to psoriasis. Bioinformatics. 2012, 28: i370-i374. 10.1093/bioinformatics/bts379.
Iqbal Z, Caccamo M, Turner I, Flicek P, McVean G: De novo assembly and genotyping of variants using colored de Bruijn graphs. Nat Genet. 2012, 44: 226-232. 10.1038/ng.1028.
Nijkamp JF, van den Broek MA, Geertman JM, Reinders MJ, Daran JM, de Ridder D: De novo detection of copy number variation by co-assembly. Bioinformatics. 2012
Mills RE, Walter K, Stewart C, Handsaker RE, Chen K, Alkan C, Abyzov A, Yoon SC, Ye K, Cheetham RK: Mapping copy number variation by population-scale genome sequencing. Nature. 2011, 470: 59-65. 10.1038/nature09708.
Zeitouni B, Boeva V, Janoueix-Lerosey I, Loeillet S, Legoix-ne P, Nicolas A, Delattre O, Barillot E: SVDetect: a tool to identify genomic structural variations from paired-end and mate-pair sequencing data. Bioinformatics. 2010, 26: 1895-1896. 10.1093/bioinformatics/btq293.
Medvedev P, Fiume M, Dzamba M, Smith T, Brudno M: Detecting copy number variation with mated short reads. Genome Res. 2010, 20: 1613-1622. 10.1101/gr.106344.110.
Handsaker RE, Korn JM, Nemesh J, McCarroll SA: Discovery and genotyping of genome structural polymorphism by sequencing on a population scale. Nat Genet. 2011, 43: 269-276. 10.1038/ng.768.
Sindi SS, Onal S, Peng LC, Wu HT, Raphael BJ: An integrative probabilistic model for identification of structural variation in sequencing data. Genome Biol. 2012, 13: R22-10.1186/gb-2012-13-3-r22.
Qi J, Zhao F: inGAP-sv: a novel scheme to identify and visualize structural variation from paired end mapping data. Nucleic Acids Res. 2011, 39: W567-575. 10.1093/nar/gkr506.
Sindi S, Helman E, Bashir A, Raphael BJ: A geometric approach for classification and comparison of structural variants. Bioinformatics. 2009, 25: i222-230. 10.1093/bioinformatics/btp208.
Hajirasouliha I, Hormozdiari F, Alkan C, Kidd JM, Birol I, Eichler EE, Sahinalp SC: Detection and characterization of novel sequence insertions using paired-end next-generation sequencing. Bioinformatics. 2010, 26: 1277-1283. 10.1093/bioinformatics/btq152.
Quinlan AR, Clark RA, Sokolova S, Leibowitz ML, Zhang Y, Hurles ME, Mell JC, Hall IM: Genome-wide mapping and assembly of structural variant breakpoints in the mouse genome. Genome Res. 2010, 20: 623-635. 10.1101/gr.102970.109.
Bentley DR, Balasubramanian S, Swerdlow HP, Smith GP, Milton J, Brown CG, Hall KP, Evers DJ, Barnes CL, Bignell HR: Accurate whole human genome sequencing using reversible terminator chemistry. Nature. 2008, 456: 53-59. 10.1038/nature07517.
Mamanova L, Andrews RM, James KD, Sheridan EM, Ellis PD, Langford CF, Ost TW, Collins JE, Turner DJ: FRT-seq: amplification-free, strand-specific transcriptome sequencing. Nat Methods. 2010, 7: 130-132. 10.1038/nmeth.1417.
Wang Q, Qu JH, Cheng XX, Kang YJ, Wan L, Qian MP, Deng MH: A study of biases of DNA copy number estimation based on PICR model. Front Math China. 2011, 6: 1203-1216. 10.1007/s11464-011-0125-x.
Wang W, Wei Z, Lam TW, Wang J: Next generation sequencing has lower sequence coverage and poorer SNP-detection capability in the regulatory regions. Sci Rep. 2011, 1: 55-
Minoche AE, Dohm JC, Himmelbauer H: Evaluation of genomic high-throughput sequencing data generated on Illumina HiSeq and genome analyzer systems. Genome Biol. 2011, 12: R112-10.1186/gb-2011-12-11-r112.
Zhang J, Wu Y: SVseq: an approach for detecting exact breakpoints of deletions with low-coverage sequence data. Bioinformatics. 2011, 27: 3228-3234. 10.1093/bioinformatics/btr563.
Nord AS, Lee M, King MC, Walsh T: Accurate and exact CNV identification from targeted high-throughput sequence data. BMC Genomics. 2011, 12: 184-10.1186/1471-2164-12-184.
Glenn TC: Field guide to next-generation DNA sequencers. Mol Ecol Resour. 2011, 11: 759-769. 10.1111/j.1755-0998.2011.03024.x.
Schadt EE, Turner S, Kasarskis A: A window into third-generation sequencing. Hum Mol Genet. 2010, 19: R227-240. 10.1093/hmg/ddq416.
Li J, Lupat R, Amarasinghe KC, Thompson ER, Doyle MA, Ryland GL, Tothill RW, Halgamuge SK, Campbell IG, Gorringe KL: CONTRA: copy number analysis for targeted resequencing. Bioinformatics. 2012, 28: 1307-1313. 10.1093/bioinformatics/bts146.
Ramachandran A, Micsinai M, Pe'er I: CONDEX: Copy number detection in exome sequences. Bioinformatics and Biomedicine Workshops (BIBMW), 2011 IEEE International Conference on: 12-15 November 2011. 2011, 87-93.
Deng X: SeqGene: a comprehensive software solution for mining exome- and transcriptome- sequencing data. BMC Bioinformatics. 2011, 12: 267-10.1186/1471-2105-12-267.
We appreciate Ms. Rebecca Hiller Posey for proofreading and polishing an earlier draft of the manuscript. This work was supported in part by the National Institutes of Health grants (R01LM011177, R21HG006037, P50CA095103,and P30CA68485), The Robert J. Kleberg, Jr. and Helen C. Kleberg Foundation (to Z.Z.), and Ingram Professorship Funds (to Z.Z.). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
The publication costs for this article were funded by the corresponding author.
This article has been published as part of BMC Bioinformatics Volume 14 Supplement 11, 2013: Selected articles from The Second Workshop on Data Mining of Next-Generation Sequencing in conjunction with the 2012 IEEE International Conference on Bioinformatics and Biomedicine. The full contents of the supplement are available online at http://www.biomedcentral.com/bmcbioinformatics/supplements/14/S11.
The authors declare that they have no competing interests.
MZ, PJ and ZZ conceived of the study. MZ primarily wrote the manuscript. QW1, QW2, PJ and ZZ participated in manuscript writing. All authors read and approved the final manuscript.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Zhao, M., Wang, Q., Wang, Q. et al. Computational tools for copy number variation (CNV) detection using next-generation sequencing data: features and perspectives. BMC Bioinformatics 14 (Suppl 11), S1 (2013). https://doi.org/10.1186/1471-2105-14-S11-S1
- Copy Number Variation
- Whole Exome Sequencing
- Whole Genome Sequencing Data
- Copy Number Variation Region
- Circular Binary Segmentation