 Methodology Article
 Open Access
 Published:
Reconstructing cancer karyotypes from short read data: the half empty and half full glass
BMC Bioinformatics volume 18, Article number: 488 (2017)
Abstract
Background
During cancer progression genomes undergo point mutations as well as larger segmental changes. The latter include, among others, segmental deletions duplications, translocations and inversions.The result is a highly complex, patientspecific cancer karyotype. Using highthroughput technologies of deep sequencing and microarrays it is possible to interrogate a cancer genome and produce chromosomal copy number profiles and a list of breakpoints (“jumps”) relative to the normal genome. This information is very detailed but local, and does not give the overall picture of the cancer genome. One of the basic challenges in cancer genome research is to use such information to infer the cancer karyotype.
We present here an algorithmic approach, based on graph theory and integer linear programming, that receives segmental copy number and breakpoint data as input and produces a cancer karyotype that is most concordant with them. We used simulations to evaluate the utility of our approach, and applied it to real data.
Results
By using a simulation model, we were able to estimate the correctness and robustness of the algorithm in a spectrum of scenarios. Under our base scenario, designed according to observations in real data, the algorithm correctly inferred 69% of the karyotypes. However, when using less stringent correctness metrics that account for incomplete and noisy data, 87% of the reconstructed karyotypes were correct. Furthermore, in scenarios where the data were very clean and complete, accuracy rose to 90%–100%. Some examples of analysis of real data, and the reconstructed karyotypes suggested by our algorithm, are also presented.
Conclusion
While reconstruction of complete, perfect karyotype based on short read data is very hard, a large fraction of the reconstruction will still be correct and can provide useful information.
Background
The current understanding of cancer suggests that it is a disease driven by somatic mutations that accumulate in the genome, within a certain tissue, during the lifetime of an individual. These mutations vary in size and effect. They can be small, e.g., single nucleotide mutations, or large structural variations caused by rearrangements such as deletions, inversions, tandem duplications and chromosomal translocations, or duplication and losses of entire chromosomes [1]. Over time these rearrangements accumulate and result in genomes less and less similar to the germline genome.
Cancer genomes are often described in the form of karyotypes. A karyotype is a high level description of the genome as a set of chromosomes and the number of copies of each. Normal karyotypes have two copies of each chromosome 1 to 22 and the sex chromosomes. In contrast, in cancer karyotypes some chromosomes may contain fragments originating from several normal chromosomes.
Types of aberration events
Most segmental changes that happen during the progression of the disease can be categorized as deletion, tandem duplication, inversion, translocation, and deletion and duplication of entire chromosomes.
A deletion is characterized by a missing segment of a chromosome, a tandem duplication happens when part of the chromosome is duplicated and thus two copies of a segment appear where normally there would only be one. An inversion occurs when a segment of a chromosome is reversed relative to its original orientation (Fig. 1).
A translocation happens when two different chromosomes “switch” end segments. Schematically, a translocation on two chromosomes (A,B) and (C,D) produces the chromosomes (A,D) and (C,B). A whole chromosomal duplication (deletion) adds (removes) a copy of a complete chromosome.
Breakpoints
The molecular mechanisms that cause somatic genome rearrangements are still the focus of investigation. The main paradigm is that a genome rearrangement occurs when one or more chromosomes break and a following joining event reassembles the fragments in a different order. A breakpoint is defined as a genomic location where the normal DNA sequence is interrupted and two nonadjacent sequence segments appear consecutively due to a joining event. A breakpoint can be considered as the most basic unit of rearrangement. The stars in Fig. 1 indicate breakpoints.
Models of genomic distance
Modeling the somatic evolution of cancer holds great value for understanding the disease process. In 1995 Hannenhalli and Pevzner proposed a method to calculate the genomic distance between two species based on the minimal number of reversals (or reversals and translocations, in the multichromosome case) required to transform the genome of one species to another [2, 3].
Braga et al. proposed another distance metric between genomes. They developed a method that calculates the distance between two genomes based on DoubleCut and Join (DCJ) operations and indels (Insertions and deletions), and utilized it to show evidence for deletion clusters in six species of Rickettsia [4]. Feijão et al. defined another metric based on Singlecut and join (SCJ) operations, and by using it they were able to recover between 60 and 90% of the topology of a phylogenic tree with 200 different genomes and with as many as 3000 genes [5, 6]. Zeira and Shamir defined a generalized model, called SCJD, allowing the operations of cut, join and whole chromosome duplication. They developed a linear time algorithm for computing the shortest sequence of operations transforming one linear genome with one copy per gene into another with two copies per gene [7].
OzeryFlato and Shamir introduced the elementary distance between two karyotypes, defined as the least number of elementary operations – breakage, fusion, duplication and deletion – transforming one into the other. They suggested a polynomial time 3approximation algorithm to find the shortest elementary distance between two karyotypes. Applying the algorithm on some 58,000 karyotypes taken from the Mitelman database [8], 99.9% of the resulting solutions matched the lower (optimal) bound [9].
Detecting chromosomal aberrations
Paired end reads
One of the main ways for inferring breakpoints in the genome, detecting structural variants and identifying rearrangements is using paired end reads produced by deep sequencing [10,11,12,13]. Paired end reads are generated by fragmenting the genomic DNA into short segments, followed by sequencing both ends of (some of) the segments (Fig. 2). Typical lengths are ~ 350 bp per fragment (also called insert) and ~ 100 bp per read (end). The unsequenced segment of the insert is called the gap (length ~ 150 bp in the example above). First, the normal sample reads from the same patient are aligned to the reference genome in order to reconstruct the individual’s reference genome and account for specific germline changes. Then the two ends of each tumor read are aligned back to individual’s reference genome. The approximate length of the insert and the relative orientation of its ends is known in advance. We expect the two ends of a fragment to be aligned to the reference genome at roughly that distance and with the correct relative orientation. An alignment is called a concordant if it meets those conditions, and discordant otherwise.
Discordant reads suggest a breakpoint in the genome. A read taken from that spot will have its two ends aligned to locations on the reference genome where those positions originally lie. The type of discordance suggests the rearrangement event that occurred (see [14]).
Detecting structural variations
A first step in analysis of paired end reads is their mapping to the reference genome. A variety of computational approaches were developed for inferring the structural variations from the discordant reads and produce a set of rearrangement events [15,16,17,18,19,20]. Other methods such as PREGO [21] take into account the concordant reads as well. BreaKmer [14] uses the misaligned reads together with the aligned concordant and discordant reads to predict rearrangements using kmer statistics. CouGaR [22] is a method for identifying largescale complex genomic rearrangements (CGRs) using both depth of coverage and discordant pairedends mapping. SVBay [23] applies a Bayesian approach to data of mapped pairedends reads to infer breakpoint locations and copy number variations and predict structural variations in a cancer genome. Recently, a new algorithm, Weaver [24], was proposed to estimate both the allelic copy number and interconnectivity of SV’s using a probabilistic graph model. Expanding on Weaver, Rajaraman et al. [25] used a graph model and an ILP formulation to further predict SV phasings and the interconnectivity of unphased SV’s with high specificity. Other algorithmic approaches infer rearrangements that are less simple and have more complex signatures [14, 26, 27].
Some methods seek to achieve higher accuracy by aggregating results from several different tools. MetaSV [28] offers an improvement of accuracy and precision in detecting different kinds of structural variants. By effectively merging the results from multiple tools, they were able to reach F1scores (harmonic mean of sensitivity and precision) of 96.2% for deletions and 84.7% for insertions. SomaticSeq [29] detects single nucleotide variants (SNVs) and small insertions and deletions (indels), using machine learning algorithms to incorporate the results from five somatic mutation callers. The authors report an F1 score of 90%.
Copy number variations
Duplications and deletions change the copy number (CN) of different segments of the DNA sequence, i.e. the number of times a segment is present in the karyotype. A normal (human) cell line has 22 diploid chromosomes (ignoring the sex chromosomes XX or XY) and so the CN of the entire karyotype is 2. A gain or a loss of an entire chromosome will decrease or increase the CN of that chromosome, respectively. A fraction of a chromosome can also be deleted or duplicated. The resulting segment or chromosome is said to have undergone a copy number variation (CNV).
Large CNVs can be detected by traditional methods like Fluorescence insitu Hybridization (FISH) [30]. Higher resolution detection of CNVs can be achieved by Array Comparative Genomic Hybridization (aCGH) [31]. With the advent of nextgeneration sequencing (NGS), several methods have been developed to infer CNV’s using DNA sequences [21, 32, 33]. NGS based methods have the potential to greatly increase the resolution of CNV analysis, but they present many computational challenges and different methods may still vary widely in the results they produce on the same DNA sequence [34].
Graph models for rearrangements
Graph theory has been highly instrumental in the area of genomic rearrangements. For example, de Bruijn graphs are used for genome assembly problems [35], and breakpoint graphs are used in reconstructing rearranged genomes across species [2, 36]. More recently, similar methods were adapted for cancer genomes [9, 37]. The breakpoint graph, introduced by Pevzner and Bafna in 1993 to represent the relation between two permutations of the same set of elements [38], remains today one of the key models in the study of genomic rearrangements. Greenman et al. expanded on the breakpoint graph and introduced a construction that is essentially equivalent called the allelic graph and its counterpart the somatic graph [39].
Oesper et al. proposed a construction that expands on the breakpoint graph, called interval adjacency graph [21]. The interval adjacency graph is constructed directly from CN and breakpoint data. The discordant reads are used to infer breakpoint locations on the DNA sequence and partition it to intervals accordingly. A full description of the graph appears under Methods.
Using the interval adjacency graph it is possible to infer rearranged sequences that agree with the data. Oesper et al. showed that an Eulerian path on the graph alternating between interval edges and reference / variant edges corresponds to a rearranged sequence of the chromosome. They developed an algorithm called PREGO to determine the most likely sequence of a rearranged karyotype. Using simulations they showed their algorithm can deduce the correct multiplicity of more than 80% of the variant edges, even with high noise and when the sample is heterogeneous. Furthermore, they applied PREGO to five ovarian cancer genomes and were able to identify numerous rearrangements and structural variants, some of which were consistent with known mechanisms. PREGO combines CN and adjacency information from paired end reads to infer multiplicity of different segments in the cancer genome. However except in simple cases, the underlying karyotype cannot be uniquely resolved, as many reconstructions will be consistent with the data.
Methods
We propose here a novel method that receives as input discordant pairedend reads and genomic CNs obtained from sequencing a cancer genome, and reconstructs a karyotype that is in most agreement with the input. The outline of our approach is as follows. We use the two data types together to construct a bridge graph, akin to the adjacency graph proposed by Oesper et al. [21]. An integer linear programming (ILP) optimization problem is formulated and then solved on the graph. The solution is a valid karyotype of the rearranged genome that is most concordant with the observed data. We also present the solution graphically.
The adjacency and bridge graphs
In our problem setup there is a normal (or reference) genome, whose contents is known, and an unknown target genome that should be reconstructed. A breakpoint is a point along the reference genome involved in a structural change event in the target genome.
Let C be the set of chromosomes in the reference karyotype. The breakpoints partition each chromosome c ∈ C into a set of k ^{c} intervals \( {I}_c=\left\{{I}_1^c,{I}_2^c\dots {I}_{k^c}^c\right\}, \) such that each \( {I}_{k^c}^c \) is an interval between consecutive breakpoints, or between a breakpoint and a chromosome end. The intervals are numbered in increasing order along c, so that c is equal to the concatenation of the intervals \( {I}_1^c,{I}_2^c\dots {I}_{k^c}^c \). We call the start and end points of interval I the tail and head of I and denote them by t _{ I } and h _{ I } respectively. Hence, I = [t _{ I }, h _{ I }], and −I = [h _{ I }, t _{ I }] is the interval I reversed. An extremity is a tail or a head of an interval. The set of all intervals \( \mathcal{I}={\cup}_{c\in C}{I}_j^c \) constitutes the set of the basic building blocks of the reference and target genomes. The length of interval I _{ j } (in bases) is denoted by l _{ j }, and L = ∑ l _{ i } is the total length of all intervals.
The target genome can be represented by a set of chromosomes, where each chromosome is a sequence of intervals, some possibly reversed (Fig. 3). A bridge is a pair of extremities that are not adjacent on the reference genome but are adjacent in the target genome. Bridges can be detected based on the pairedend read data of the target genome (Fig. 4). The support level of bridge b _{ i } is the number of pairedend reads that support it, denoted μ _{ i }. The total support score for all bridges is denoted\( \mu ={\sum}_{b_i}{\mu}_i \).
Each interval I _{ i } ∈ I has a CN N _{ i } ≥ 0 indicating the number of times it appears in the target genome. The set of CNs of all intervals is called the copy number profile of the target. That profile can be derived from deep sequencing data or from array CGH data. In perfect data, N _{ i } is exactly the number of copies of the interval in the target genome. In practice, the CNs are real valued estimates based on mean coverage of each interval.
Let us first reiterate the definition of the interval adjacency graph, introduced in [21]. The input is (1) the reference genome represented as a sequence of intervals for each chromosome. These intervals form the set \( \mathcal{I}=\left\{{I}_1,\dots, {I}_n\right\} \); interval I _{ j } has length l _{ j }. (2) The CN profile of the intervals: Interval I _{ j } has CN N _{ j }. (3) The set of bridges \( {\left\{{a}_i,{b}_i\right\}}_{i=1}^m \) and the support μ _{ i } for each bridge. Each a _{ i } and b _{ i } is an extremity of an interval in \( \mathcal{I} \). We define a weighted undirected graph G(V, E, w) whose vertices are the interval extremities. For each interval I _{ i } = [t _{ i }, h _{ i }], the graph contains an interval edge e _{ I }(t _{ i }, h _{ i }) ∈ E _{ I } connecting its two extremities, of weight N _{ i }. For each two intervals I _{ i }, I _{ i + 1} that are adjacent on the reference genome, a reference edge e _{ R }(h _{ i }, t _{ i + 1}) ∈ E _{ R } connects the head of I _{ i } to the tail of I _{ i + 1}. Reference edges are unweighted. Each bridge is represented by a bridge (or variant) edge e _{ V }(a _{ i }, b _{ j }) ∈ E _{ V } connecting the two extremities a _{ i } and b _{ j }, with weight μ _{ i }. In total, the edge set of the graph is E = E _{ I } ∪ E _{ R } ∪ E _{ V }. We denote by S ⊆ V the set of vertices that represent telomere nodes, i.e. the nodes representing start and end points of each reference chromosome, hence \( S={\cup}_{c\in C}\left\{{t}_1^c,{h}_{k^c}^c\right\} \) includes the heads of all starting intervals and the tails of all ending intervals in each chromosome’s partition.
A bridge graph is an interval adjacency graph with two minor changes: (1) bridge edges are assigned weights. The weight w(e) of the bridge e(u, v) is its support score, namely the number of paired end reads supporting that bridge. Hence, in a bridge graph both bridge and interval edges have weights. (2) We transform each undirected edge e(u, v) in the interval adjacency graph into two directed edges e _{→} : u → v, e _{←} : v → u. The original undirected edge is referred to as a connection to distinguish it from the directed edges, and E = E _{→} ∪ E _{←} is the set of edges in the graph. An example of a bridge graph is given in Fig. 4.
Reconstructing the rearranged karyotype
Given the bridge graph G(V, E, w), we wish to find paths in G that correspond to rearranged chromosomes. Suppose first that the input data are complete and errorless. Recall that S ⊆ V is the set of vertices that represent telomere nodes, i.e. the nodes representing the start and end points of each chromosome. A valid path p is a path through G beginning and ending at s _{1}, s _{2} ∈ S that alternately traverses interval and noninterval (i.e. reference/bridge) edges, and where the number of times each interval connection e _{ i } is traversed (in either direction), denoted f _{ p }(e _{ i }), is less than or equal to the CN of interval i, N _{ i }.
The requirement for an alternating path is because a traversal of an interval edge corresponds to traversing a segment from the reference genome, while a traversal of a reference/bridge edge is equivalent to a transition between segments. Therefore, such an alternating path represents a sequence of segments from the reference genome. Note that f _{ p }(e _{ i }) = f _{ p }(e _{ i→}) + f _{ p }(e _{ i←}) for every connection e. A set of such paths P = {p _{1}, p _{2}…p _{ n }} where for each interval connection e _{ i }, ∑_{ p ∈ P } f _{ p }(e _{ i }) = N _{ i } corresponds to a set of rearranged chromosomes, or a valid karyotype.
The restriction that the path alternates between interval and noninterval edges means that at each nontelomeric node v ∉ S, every traversal on an interval edge going into v must be followed by a traversal on a reference\bridge edge going out of v, and viceversa. Telomeric nodes are excluded from this constraint as by definition they are the start or end of a path.
As detailed above, each connection between nodes u, v is composed of two antiparallel directed edges. For each node v ∈ V we denote \( {E}_{I_{\leftarrow }}(v),{E}_{I\to }(v),{E}_{R\leftarrow }(v), \) E _{ R→}(v), E _{ B←}(v), E _{ B→}(v) as the set of interval, reference and bridge edges that go in and out of v respectively. As above, we denote by f _{ p }(e) the number of times a connection e is traversed in path p and f _{ P }(e) = ∑_{ p ∈ P } f _{ p }(e) is the total number of times a connection e is traversed in P. Additionally, for a set of connections E, f _{ P }(E) = ∑_{ e ∈ E } f _{ P }(e) is the total number of times all connections in E are traversed in P. The constraints for a valid set of paths P, representing a rearranged karyotype, can be therefore formulated as:

(1)
\( {\displaystyle \begin{array}{l}{f}_P\left({E}_{I_{\to }}(v)\right)={f}_P\left({E}_{R_{\leftarrow }}(v)\right)+{f}_P\left({E}_{V_{\leftarrow }}(v)\right)\\ {}{\forall}_{v\notin S}\end{array}} \)

(2)
\( {\displaystyle \begin{array}{l}{f}_P\left({E}_{I_{\leftarrow }}(v)\right)={f}_P\left({E}_{R_{\to }}(v)\right)+{f}_P\left({E}_{V_{\to }}(v)\right)\\ {}{\forall}_{v\notin S}\end{array}} \)

(3)
\( {\displaystyle \begin{array}{l}{f}_P(e)\in {\mathrm{\mathbb{N}}}^0\\ {}{\forall}_{e\notin E}\end{array}} \)
Scoring candidate solutions
Recall that the interval and bridge edges have weights, representing the measured CN of the intervals and the support score for the bridges, respectively. These values are in practice noisy. Given a bridge graph G(V, E, w) and a valid set of paths P representing a rearranged karyotype, we define a discordance score of P, denoted d _{ G }(P), which measures how much P is in agreement with the data in G, as follows:
The first sum measures the disagreement of P with the CN profile. It is the sum over all interval edges e ∈ E _{ I } of the absolute difference between f _{ P }(e) and the input weight w(e), normalized by l _{ e }. We normalize the weights of the intervals by their lengths since longer genomic intervals are expected to have more accurate CN values, and hence should be penalized more for disagreement. Dividing by L guarantees that the range of the first sum is [0, 1] if the absolute difference values are ≤ 1.
The second sum the disagreement of P with the bridge data. The more bridges P is utilizing, the more concordant it is with the bridge data. To reflect this, a penalty is given for each bridge edge e ∈ E _{ V } that is not used in P. The bigger the support score for a bridge is, the bigger the penalty if it is not used, and so the penalties are normalized by w(e). Dividing by \( \mu ={\sum}_{e\in {E}_V}w(e) \) guarantees that the range of the second sum is [0, 1]. To avoid summing over e ∉ P, we can rewrite the second term as\( \upalpha\;{\sum}_{\mathrm{e}\in {\mathrm{E}}_{\mathrm{V}}}\frac{\mathrm{w}\left(\mathrm{e}\right)}{\upmu}\left(1\min \left(1,{\mathrm{f}}_{\mathrm{P}}\left(\mathrm{e}\right)\right)\right) \).
The parameter α determines the relative weight the algorithm gives to pairedend reads data, i.e. how much it tries to utilize bridge edges in the solution. Using the algorithm on real tumor data, we set α = 0.5.
The ILP formulation
We wish to find a rearranged karyotype that is most consistent with the data, i.e., it corresponds to a valid set of paths and has smallest possible discordance score. This problem can be formulated as an ILP on the bridge graph G(V, E, w), as we now show.
For each connection e _{ i } ∈ E we define two variables x _{ i→}, x _{ i←}. The variables represent the number of times each edge is traversed in a path, and so f _{ P }(e _{ i }) = x _{ i→} + x _{ i←} . Each variable is noted x ^{I}, x ^{B} or x ^{R} for interval, bridge or reference edges respectively. Using these variables we can formulate the problem as follows.
Minimize:
Subject to:

(1)
∀_{ i } x _{ i } ∈ ℕ^{0}

(2)
\( {\forall}_{v\notin S}\ {\sum}_{e_i\in {E}_{I_{\to }}(v)}{x}_{i\to}^I={\sum}_{e_i\in {E}_{R_{\leftarrow }}(v)}{x}_{i\leftarrow}^R+{\sum}_{e_i\in {E}_{V\leftarrow }(v)}{x}_{i\leftarrow}^B \)

(3)
\( {\forall}_{v\notin S}{\sum}_{e_i\in {E}_{I_{\leftarrow }}(v)}{x}_{i\leftarrow}^I={\sum}_{e_i\in {E}_{R_{\to }}(v)}{x}_{i\to}^R+{\sum}_{e_i\in {E}_{V\to }(v)}{x}_{i\to}^B \)
Constraint set (1) guarantees an integral nonnegative solution. Constraints (2) and (3) are the valid path constraints. Note that telomeric nodes in S are not constrained.
Tools
The core of the algorithm was implemented in java using the ILP solver package CPLEX, distributed by IBM [40] and was run on UNIX. The simulations module and the rest of the algorithm was implemented in python version 2.7 on Windows. The code is available in https://github.com/ShamirLab/Karyotypereconstruction. A typical run of a single karyotype on a standard PC takes around 1 s.
Results and discussion
Simulations
To assess the performance of our algorithm, we simulated tumor karyotypes and applied the algorithm to them. To evaluate the quality of each reconstructed karyotype, it was compared to the correct karyotype, and summary statistics were computed. An overview of the simulation algorithm is as follows:

1.
Start with a normal diploid karyotype H with C chromosomes

2.
Perform N operations resulting in karyotype T ^{′}

3.
Compute the exact (noiseless) CN profile and the bridges in T ^{′}

4.
Add noise to the CN data and generate support values for the bridges
We start with a normal diploid karyotype H with a prescribed number of chromosomes. For simplicity, each chromosome is represented by a sequence of 300 atomic segments of equal size, which are its basic units. We perform a series of operations on the karyotype by applying deletions, inversions, tandem duplications and translocations. The types and the positions of the rearrangements are drawn uniformly at random. The span of operations that affect a single chromosomes (deletions, duplications and inversions) was limited to 30 atomic segments. This limit was set in order to avoid rapid erasure of large chromosomal segments by deletions. The total number of operations applied varies and determines the complexity of the resulting tumor karyotype T.
By comparing H and T, breakpoints are detected and each normal chromosome is partitioned into segments. Each segment has a CN (the number of occurrences of that segment in T). Each two consecutive segments in T that are not consecutive (and/or not in the same relative orientation) in H constitute a bridge. The clean (noiseless) data can thus be summarized as an integervalued CN profile and the set of all bridges formed.
To simulate noisy scenarios, the CN profile and the bridge information is modified as follows. Normally distributed noise x is added to the CN of each segment independently, where x~N(0, ϵ). The support for each bridge (corresponding to the number of discordant reads supporting it) is drawn independently from an exponential distribution Exp(λ) (The exponential distribution was chosen based on empirical data with λ = 0.1866. See below). To simulate the possibility of bridges being completely missed, each bridge has probability p to completely be omitted from the final set of bridges.
In summary, the simulation program receives the following parameters (the default values appear in parentheses):

C  The number of chromosomes (default: 5).

N  The number of structural and numerical operations applied (default: 5).

ϵ  The standard deviation of the noise in the CN profile data (default: 0.28)

p – The probability to completely miss a bridge (default: 0.05).
In the base scenario, all parameters were at their default values. These parameters correspond to those computed on a tumor sample of medium complexity and a realistic level of noise (see section real tumor analysis below). Other scenarios were explored by changing the value of one of the parameters above while keeping the rest at their default levels.
The use of 300 equalsize atomic segments per chromosome is done for convenience. The atomic segments determine where a breakpoint can occur and thus constitute the smallest identifiable unit on the genome. In real genomes they represent a single base pair (assuming there is no restriction on the location of breakpoints), of which there are ~ 3^{∗}10^{9}. For a model simulating up to 30 operations, this number of units is unnecessarily too large, as intervals will typically span millions of bases. This model disallows very short segments, which are in reality harder to detect and whose CN measurements are inaccurate.
Solution quality measures
We used five different measures for the level of correctness of a solution. Let T be the simulated (true) karyotype, let T ^{∗} be the simulated noisy karyotype, and let S be the karyotype produced by the algorithm:

1.
Is S equivalent to T? We say that S is equivalent to T if they have the same CN profile and both use the same bridges. Most equivalent karyotypes only differ in chromosomal orientation, and thus represent the same solution. We call such a solution correct (Additional file 1: Figure S1).

2.
Do S and T have the same CN profile? The CN of an interval is determined by many reads (or probes) and so is expected to be more robust than bridge information, determined by a few paired end reads. This criterion tests if S and T match in their CN profile. We call this criterion Equal Copy Number (ECN).

3.
Does S have an equal or better score than T? When noise level is high, T and T ^{∗} may differ substantially, and a solution closer to T ^{∗} than to T does not indicate a failure of the algorithm but rather that the noise level is too high. Here the score is the ILP objective function value. We call this criterion Equal or Better Score (EBS).

4.
Is S equivalent to T excluding missing bridges? T ^{∗} may not include all the bridges found in T, and in that case S can never be equivalent to T. However, we consider S to be correct for all observed bridges if it has the correct CN profile for all segments that are unaffected by a missed bridge, and is using all the bridges from T that are included in T ^{∗} (Additional file 2: Figure S2). We call this metric Equivalent for Observed Bridges (EOB).

5.
What fraction of the intervals has the correct CN? This score is the percentage of intervals, weighted by length, that have the same CN in S and T. Unlike criteria 1–4, which are binary, this criterion measures the extent of correctness of a solution, and thus is more sensitive and accounts also for partiallycorrect solutions. We call it the CN score.
Base scenario
Ten thousand karyotypes were generated for the base scenario, and the algorithm was applied with bridge support weight α = 0.1. To assess the distribution of each success rate criterion, the karyotypes were divided into 100 batches of 100 karyotypes each. Mean scores were captured for each batch and the variation of the mean was computed. The performance is summarized in Fig. 5.
The algorithm correctly identified a median of 62% of the karyotypes, with an IQR of 5%. For an additional 13% of the cases, the solution had an equal CN profile as the correct solution, with a median of 75% and an IQR of 6%. A median of 82% (IQR 4%) of all karyotypes resulted in a solution with a score equal or better than the correct one. When disregarding missing bridges, the algorithm correctly identified a median 84% of karyotypes (IQR 5%). For the CN score of all the 10,000 simulations, the mean was 0.97, and the median was 1 (IQR 0.07%). See Additional file 3: Figure S12 for a display of the distributions over all 10,000 simulations.
The effect of separate parameters
The effect of separate parameters was tested by simulations in which one parameter was altered, while keeping the other parameters at their value in the base scenario. One hundred simulated karyotypes were generated for each value and the percentage of solutions falling into the categories of correct, ECN, EBS and EOB was evaluated.
Bridge support weight
We first tested the effect of α, the relative weight assigned the bridges, on the performance, for 0 ≤ α ≤ 2. There is a noticeable improvement when α > 0, and little effect for the range of 0 < α ≤ 0.1. For larger values of α there is a small but noticeable negative effect. (Additional file 4: Table S1).
Noise in copy number measurements
We tested the algorithm for different levels of CN noise ϵ under the base scenario. The results are shown in Fig. 6. As expected, a higher level of noise makes it harder for the algorithm to find the correct solution. For ϵ ≤ 0.3 the performance of the algorithm is quite good, and for ϵ ≥ 0.4 the results begin to deteriorate. As expected, at high noise levels the majority of the solutions have better score than the true one.
The number of operations
We tested the algorithm on karyotypes that underwent 1 ≤ N ≤ 30 structural and numerical operations, under the base scenario. The results are shown in Fig. 7. As expected, more operations make the problem harder and the success rates decrease. For example, the fraction of perfectly solved cases drops from 88% with one operation to less than 10% with 30 operations. The CN score drops more slowly, as CN of long fragments can still be reasonably inferred even if their order is incorrect.
Other parameters
When testing the effect of other parameters, the results met our expectations – karyotypes with less chromosomes (Additional file 5: Figure S3) or a single copy of each chromosome instead of diploid (Additional file 5: Figure S4) yielded better results. Results were also better when the probability of missing a bridge was lower (Additional file 5: Figure S5).
We also looked at cancer heterogeneity situations. Different cells of the same tumor can have different karyotypes, having taken different evolutionary paths [41,42,43,44,45]. Most cancer data today is still based on DNA from numerous cells, providing measurements from a mixture of genomes. Can the karyotype be reconstructed out of the heterogeneous mixture? When simulating data mixture of normal and a cancer karyotype results only dropped mildly with the relative abundance of normal data (Additional file 5: Figure S8). However, when mixing two distinct cancer karyotypes, performance dropped rapidly with the heterogeneity (Additional file 5: Figure S6, Additional file 6: Table S3).
Finally, we simulated karyotypes by selecting operations with frequencies as reported in [46] rounded to multiples of 10%. There was little difference in the success rates between the uniform distribution and the uneven one (Additional file 5: Figure S7, Additional file 7).
Real tumor analysis
We applied the algorithm on data extracted from real samples. Malhotra et al. [46] examined whole genome sequencing data of 64 different tumor samples, and reported for each sample a CN profile and a set of bridges with their support. We first filtered from the data very small segments and the corresponding breakpoints (see Additional file 8 for more details). Often the set of normal chromosomes that are involved in rearrangements and CN changes in a tumor can be partitioned into independent groups of chromosomes (i.e., no two segments in different groups are connected by a bridge). In our graph representation, each such group is a connected component, which can be analyzed separately by the algorithm. The 64 tumor samples in [46] constituted together 570 such components, and each was analyzed separately. The mean number of chromosomes (out of the 23) per component was 1.72 and its standard deviation was 2.36.
Noise estimation
We first wanted to assess the noise level in the actual data affecting the reported CN values. Since CN in noiseless data should be integer, we estimated the noise d _{ i } for the reported CN c _{ i } as c _{ i } − [c _{ i }], where [x] is the nearest integer value to x. The CN data include 22,321 CN segments. A scatter plot of the standard deviation of the noise level vs. the number of bridges in each component can be seen in Fig. 8. As expected, the mean noise level across the data was 0, showing that the noise is unbiased towards neither negative nor positive values. The standard deviation was 0.28, a value that we used as our default simulation scenario. Note that this estimate is a lower bound, since some measured CN values may actually differ from the real ones by more than 0.5.
In addition to CNs, the data include bridges and for each bridge an integer value, its support. The expected average support can be derived from the read depth and the insert size (see Additional file 9 for more details) and was found to be 10.7. The observed mean support score across all the data was 10.8. Figure 9 shows the distribution of the support scores across the data. The distribution closely resembles an exponential distribution with λ = 0.1866. For that reason, that was the value used in our simulation model (see Additional file 10: Figure S9).
The GBM10 sample
We analyzed in detail three components of bridge graphs obtained from real data. Additional file 11: Table S2 shows information about them. Each has undergone 7–8 rearrangements, involving 1–4 chromosomes. For each component, the ILP algorithm outputs a directed weighted graph with a weight function that minimizes the distance and that can be broken into a set of paths P = {p _{1}, …p _{ n }}, starting and ending at a telomere nodes, and alternating interval and noninterval edges. Another script translates the solution of the ILP solver to a dot language representation [47] that can then be visualized using a graph visualization tool such as GraphViz [48].
Figure 10a shows the graph corresponding to the component of chromosomes 4 and X in tumor sample GBM 10 (Glioblastoma multiforme). The resulting karyotype produced by our algorithm for this example is shown in Fig. 10b. This graph can be broken into four different paths, representing both copies of the rearranged chromosomes 4 and X (Fig. 10c).
The other two examples are described in Additional file 12: Figures S10 and S11.
Conclusions
In this work, the problem of inferring a tumor karyotypes from short paired end read data was investigated. A novel algorithm based on graph theory and ILP was introduced to solve the problem, and simulations were performed in order to evaluate the utility of such an approach. Some examples of analysis of real data were also presented.
To accurately estimate the correctness and robustness of the algorithm, validation against a data set of verified karyotypes is needed. However, a comprehensive set of sequenced tumor samples with CN profiles and pairedend reads data, matched with entire reconstructed karyotypes, is currently not available. Data sets that currently exist either do not include a fully reconstructed karyotype, or include karyotypes of a very low resolution [8]. We therefore used simulations to test and measure the success of our algorithm in a spectrum of scenarios, as well as to point out potential pitfalls.
The analysis of simulated data suggests that the most meaningful factors affecting the accuracy of solutions produced by our method are the noise and completeness levels of the data. We tested the algorithm in a scenario, designed to mimic parameter values observed in real data. Under these conditions, the algorithm correctly inferred 69% of the karyotypes. However, the success rate increased to 79% when considering solutions that are correct relative to the noisy input, and when accounting for unreported bridges, 87% of the tested cases were correct (Fig. 5).
Furthermore, in scenarios where there is almost no noise, or when no bridges are unreported, the results are much better: accuracy was 90% and 100%, respectively (Fig. 6, Additional file 5: Figure S5). This strongly suggests that our method is limited mostly by the completeness and accuracy of the measured data. It suggests that more accurate sequencing technologies are needed in order to increase the chance to solve the karyotype reconstruction problem correctly.
Our method was relatively robust when applied on data taken from tumor cells contaminated by healthy tissue (Additional file 5: Figure S8). A sample that includes reads taken from a mixture of different tumor cells poses a bigger challenge, and the resulting karyotype is incorrect more often than it is correct (Additional file 5: Figure S6).
Depending on one’s perspective, the results can be viewed as good or bad news. On the one hand, full, perfect reconstruction is not attained in over 30% of the cases. On the other hand, even in those imperfect cases, most of the reconstruction details are correct, as quantified by the other, less stringent, measurement criteria (Fig. 5). Biological research has a great tradition of building up from incomplete data, the most obvious example being the human genome whose yetincomplete versions have kept evolving for the past 15 years. It may be the case that the imperfect reconstruction of cancer karyotypes can still produce valuable conclusions and findings.
Limitations
Using simulations allows us to gain better understanding of the capabilities and limitations of our algorithm, but it requires us to make assumptions about the mechanisms driving genomic rearrangements in tumor cells and about the statistical properties of the read data. Both types of assumptions limit the generality of conclusions we can draw.
Firstly, our model defines a limited set of possible rearrangements (deletion, duplication, inversion and chromosomal translocation) and assumes that they occur with equal probabilities. Furthermore, our simulation of rearrangement events (except translocations) limits the genomic range they can span (see Methods) and assumes that events are equally likely to occur in any position on the genome. While these assumptions are very far from the real process of mutating cancer cells, they do provide a mechanism that can generate any rearranged karyotype. Our method proved robust when adjusting the frequency of each type of rearrangement to that observed in the data obtained from [46] (Additional file 5: Figure S7), but other possible rearrangement mechanisms and their effect on the performance of the algorithm were not explored.
A second problem arises when attempting to create very complicated karyotypes using a large number of rearrangements. Stephens et al. [49] suggested that in some cases a single catastrophic event called chromothripsis occurs, in which a section of the chromosome is shattered into a large number of small fragments and then reassembled, creating a karyotype that is much more complex [49]. While all possible karyotypes can be generated using our model, very complex ones are unlikely. Note that once a deletion operation has been performed, the deleted segment cannot reappear and will therefore be absent from the final karyotype. When performing a large number of rearrangements on a chromosome, deletions will occur and sometime remove segments that were rearranged by a previous operation, essentially reducing the complexity of the resulting final karyotype. We tested our method on karyotypes that have undergone a maximum of 30 operations (Fig. 7), but a modified simulation model needs to be used in order to generate more complex karyotypes. Currently our results reflect more faithfully the ability of the algorithm on relatively simple karyotypes, which constituted the majority in the real data that we used.
A third type of limitation is due to the noise model assumptions. While we tried to borrow values of noise as estimated from the real data (see Real Tumor Analysis), there are other parameters that affect the noise and thus the quality of the analysis, including incorrectly mapped reads due to sequencing errors, nonuniquely mappable reads, insert length variance, breakpoints that fall within a read (and not in the gap), nonuniform read coverage, etc. These are all left to future work.
One of the limitations of our algorithm is its inability to “predict” bridges that were not observed in the data. The algorithm looks for a path on the graph corresponding to a karyotype that best fits the observed CN profile, yet it overlooks potential paths that can be constructed by bridging two unconnected interval edges – essentially predicting a bridge. This implies that data produced using sensitive methods, even with higher rates of false positives, might be preferable over data with false negatives. In addition, while our target function penalizes a solution for not using a reported bridge (see Additional file 4: Table S1), another measure of quality, which was not explored in this work, is the percent of bridges used by a solution.
Future directions
One important aspect of the technology in detecting bridges is the insert size. A bridge will usually be detected only when the two reads of a PER are on the two different sides of it (see Additional file 9). Therefore, the larger the read length and insert  the higher the bridge coverage. This implies that sequencing techniques with longer inserts can dramatically change the performance of the algorithm. Several such techniques are forthcoming, and some methods for detecting structural variations were already developed for them [28, 29, 22]. Note however that very short rearrangements that span less base pairs than the length of the read may be missed altogether. It is worth noting that our method receives as input only the bridges spanning breakpoints and a CN profile, and is indifferent to the method used to infer the breakpoints. Breakpoints can be detected using other methods apart from PER, such as split reads (softclipped reads) [50, 51]. Analysis of the effect this makes on the performance of our method can be done in the future.
A possible extension to our method can be the addition of weights to the reference edges. Recall that reference edges represent a connection between two segments that is expected according to the reference genome. Unlike interval edges or bridge edges, reference edges are weightless in our model. One metric that can be used to establish a confidence score for a reference edge is the number of PERs whose ends span the two segments bordering the reference connection.
A performance comparison of our algorithm to PREGO [21] is desirable, but not straightforward as the algorithms differ in input and output. PREGO expects as input the number of pairedend reads supporting each edge on the interval graph, from which a CN profile is inferred. The output of PREGO is a list of integer weights for the edges on the graph corresponding to one or more possible rearrangements. Our method, on the other hand, expects a computed CN profile and a list of bridges, and outputs a reconstructed karyotype. Such comparison is therefore left for future work.
Abbreviations
 CN:

Copy Number
 CNV:

Copy Number Variations
 EBS:

Equal or Better Score
 ECN:

Equal Copy Number
 EOB:

Equal for Observed Bridges
 ILP:

Integer Linear Programming
 PER:

PairedEnd Reads
References
 1.
Vogelstein B, et al. Cancer Genome Lanscapes. Science. 2013;339(80):1546–58.
 2.
Hannenhalli S, & Pevzner PA. Transforming men into mice (polynomial algorithm for genomicdistance problem). Proc IEEE 36th Annu Found Comput Sci. (1995). doi:10.1109/SFCS.1995.492588
 3.
Sridhar Hannenhalli PP. Transforming cabbage into turnip (polynomial algorithm for sorting signed permutations by reversals). JACM. 1999;46:1–27.
 4.
Braga MDV, Willing E, Stoye J. Double cut and join with insertions and deletions. J Comput Biol. 2011;18:1167–84.
 5.
Feijão P, Meidanis J. SCJ: a breakpointlike distance that simplifies several rearrangement problems. IEEE/ACM Trans Comput Biol Bioinforma. 2011;8:1318–29.
 6.
Biller P, Feijão P, Meidanis J. Rearrangementbased phylogeny using the singlecutorjoin operation. IEEE/ACM Trans Comput Biol Bioinforma. 2013;10:122–34.
 7.
Zeira R, Shamir R. in 396–409. Cham: Springer; 2015. doi:10.1007/9783319199290_34.
 8.
Mitelman F, Johansson B, & Mertens F (Eds.). Mitelman Database of Chromosome Aberrations and Gene Fusions in Cancer. (2016). at http://cgap.nci.nih.gov/Chromosomes/Mitelman
 9.
OzeryFlato M, & Shamir R. Sorting cancer karyotypes by elementary operations. Lect Notes Comput Sci. (including Subser Lect Notes Artif Intell Lect Notes Bioinformatics). 2008;5267 LNBI:211–225.
 10.
Tuzun E, et al. Finescale structural variation of the human genome. Nat Genet. 2005;37:727–32.
 11.
Korbel JO, et al. Pairedend mapping reveals extensive structural variation in the human genome. Science. 2007;318:420–6.
 12.
Kidd JM, et al. Mapping and sequencing of structural variation from eight human genomes. Nature. 2008;453:56–64.
 13.
Bentley DR, et al. Accurate whole human genome sequencing using reversible terminator chemistry. Nature. 2008;456:53–9.
 14.
Abo RP, et al. BreaKmer: detection of structural variation in targeted massively parallel sequencing data using kmers. Nucleic Acids Res. 2014;43:1–13.
 15.
Quinlan AR, et al. Genomewide mapping and assembly of structural variant breakpoints in the mouse genome. Genome Res. 2010;20:623–35.
 16.
Chen K et al. BreakDancer: An algorithm for high resolution mapping of genomic structural variation. 6, 677–681 (2013).
 17.
Korbel JO, et al. PEMer: a computational framework with simulationbased error models for inferring genomic structural variants from massive pairedend sequencing data. Genome Biol. 2009;10:R23.
 18.
Hormozdiari F, Alkan C, Eichler EE, Sahinalp SC. Combinatorial algorithms for structural variation detection in highthroughput sequenced genomes. Genome Res. 2009;19:1270–8.
 19.
Hormozdiari F, Hajirasouliha I, McPherson A, Eichler EE, Sahinalp SC. Simultaneous structural variation discovery among multiple pairedend sequenced genomes. Genome Res. 2011;21:2203–12.
 20.
Hormozdiari F, et al. Nextgeneration VariationHunter: combinatorial algorithms for transposon insertion discovery. Bioinformatics. 2010;26:i350–7.
 21.
Oesper L, Ritz A, Aerni SJ, Drebin R, Raphael BJ. Reconstructing cancer genomes from pairedend sequencing data. BMC Bioinformatics. 2012;13(Suppl 6):S10.
 22.
Dzamba M, et al. Identification of complex genomic rearrangements in cancers using CouGaR. Genome Res. 2017;27:107–17.
 23.
Iakovishina D, JanoueixLerosey I, Barillot E, Regnier M, Boeva V. SVbay: structural variant detection in cancer genomes using a Bayesian approach with correction for GCcontent and read mappability. Bioinformatics. 2016;32:984–92.
 24.
Li Y, Zhou S, Schwartz DC, Ma J. Allelespecific quantification of structural variations in cancer genomes. Cell Syst. 2016;3:21–34.
 25.
Rajaraman A, Ma J. in 224–240. Cham: Springer; 2017. doi:10.1007/9783319569703_14.
 26.
McPherson A, et al. nFuse: discovery of complex genomic rearrangements in cancer using highthroughput sequencing. Genome Res. 2012;22:2250–61.
 27.
Baca SC, et al. Punctuated evolution of prostate cancer genomes. Cell. 2013;153:666–77.
 28.
Mohiyuddin M, et al. MetaSV: an accurate and integrative structuralvariant caller for next generation sequencing. Bioinformatics. 2015;31:2741–4.
 29.
Fang LT, et al. An ensemble approach to accurately detect somatic mutations using SomaticSeq. Genome Biol. 2015;16:197.
 30.
Kallioniemi A, Visakorpi T, Karhu R, Pinkel D, Kallioniemi O. Gene copy number analysis by fluorescence in situ hybridization and comparative genomic hybridization. Methods. 1996;9:113–21.
 31.
Pinkel D, Albertson DG. Array comparative genomic hybridization and its applications in cancer. Nat Genet. 2005;37(Suppl):S11–7.
 32.
Yoon S, Xuan Z, Makarov V, Ye K, Sebat J. Sensitive and accurate detection of copy number variants using read depth of coverage. Genome Res. 2009;19:1586–92.
 33.
Abyzov A, Urban AE, Snyder M, Gerstein M. CNVnator: an approach to discover, genotype, and characterize typical and atypical CNVs from family and population genome sequencing. Genome Res. 2011;21:974–84.
 34.
Alkan C, Coe BP, Eichler EE. Genome structural variation discovery and genotyping. Nat Rev Genet. 2011;12:363–76.
 35.
Pevzner PA, Tang H, Waterman MS. An Eulerian path approach to DNA fragment assembly. Proc Natl Acad Sci U S A. 2001;98:9748–53.
 36.
Pevzner P. Computational Molecular Biology: An Algorithmic Approach. (MIT Press) 2000. at http://books.google.com/books?hl=en&lr=&id=dpgh2UxGpacC&pgis=1
 37.
Raphael BJ, Volik S, Collins C, Pevzner PA. Reconstructing tumor genome architectures. Bioinformatics. 2003;19
 38.
Bafna V, Pevzner PA. Genome rearrangements and sorting by reversals. SIAM J Comput. 1994;25:272–89.
 39.
Greenman CD, et al. Estimation of rearrangement phylogeny for cancer genomes. Genome Res. 2012;22:346–61.
 40.
IBM. IBM ILOG CPLEX V12.1. (2009). at ftp://public.dhe.ibm.com/software/websphere/ilog/docs/optimization/cplex/ps_usrmancplex.pdf
 41.
Almendro V, et al. Inference of tumor evolution during chemotherapy by computational modeling and in situ analysis of genetic and phenotypic cellular diversity. Cell Rep. 2014;6:514–27.
 42.
de Bruin EC, Taylor TB, Swanton C. Intratumor heterogeneity: lessons from microbial evolution and clinical implications. Genome Med. 2013;5:101.
 43.
Klein CA. Selection and adaptation during metastatic cancer progression. Nature. 2013;501:365–72.
 44.
Bedard PL, Hansen AR, Ratain MJ, Siu LL. Tumour heterogeneity in the clinic. Nature. 2013;501:355–64.
 45.
Ding L, Raphael BJ, Chen F, & Wendl MC. Advances for studying Clonal evolution in cancer. Cancer Lett 340, 212–219 (2013).
 46.
Malhotra A, et al. Breakpoint profiling of 64 cancer genomes reveals numerous complex rearrangements spawned by homologyindependent mechanisms. Genome Res. 2013;23:762–76.
 47.
Emden G, Eleftherios K, & Stephen N. Drawing graphs with dot. (2006). at http://www.graphviz.org/Documentation/dotguide.pdf
 48.
John E, Emden G, Eleftherios K, Stephen N, & Gordon W. in Graph Drawing Software (eds. Jünger, M. & Mutzel, P.) 127–148 (Springer Berlin Heidelberg, 2004). doi:10.1007/9783642186387_6
 49.
Stephens PJ, et al. Massive genomic rearrangement acquired in a single catastrophic event during cancer development. Cell. 2011;144:27–40.
 50.
Karakoc E, et al. Detection of structural variants and indels within exome data. Nat Methods. 2012;9:176–8.
 51.
Wang J, et al. CREST maps somatic structural variation in cancer genomes with basepair resolution. Nat Methods. 2011;8:652–4.
Acknowledgments
We would like to thank Nir Atias for help with the IBM CPLEX package and Roy Kasher for help with software design and optimization.
Funding
This study was supported in part by the Israel Science Foundation (grant 317/13) and by the Bella Walter Memorial Fund of the Israel Cancer Association.
Availability of data and materials
All data analyzed in this research was reported by Malhotra et al. in [46] and is available as supplemental material in http://genome.cshlp.org/content/23/5/762/suppl/DC1. The code developed in this study is available in https://github.com/ShamirLab/Karyotypereconstruction.
Author information
Affiliations
Contributions
RE and RS designed the study. RE prepared the data, developed the tools used to simulate and analyze data, and produced the results. RE and RS analyzed the results. RE and RS wrote the manuscript. Both authors read and approved the final manuscript.
Corresponding author
Correspondence to Ron Shamir.
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional files
Additional file 1: Figure S1.
An example of graphs that represent equivalent yet not identical solutions. (DOCX 60 kb)
Additional file 2: Figure S2.
Edges affected by removing a bridge. (DOCX 66 kb)
Additional file 3: Figures S12.
Boxplot of the distribution for different success measures over 10,000 simulations. (DOCX 30 kb)
Additional file 4: Table S1.
Performance of the algorithm for different values of the parameter alpha. (DOCX 19 kb)
Additional file 5: Figure S3S8.
Results of different simulated scenarios. (DOCX 33 kb)
Additional file 6: Table S3.
Operations frequencies used in the default scenario and in the alternative scenario. (DOCX 18 kb)
Additional file 7:
The effect of tumor heterogeneity. The file details the effects of tumor heterogeneity on our simulations model. (DOCX 17 kb)
Additional file 8:
Breakpoint filter. The file describes the details of our breakpoint filter. (DOCX 37 kb)
Additional file 9:
Estimation of bridge support in real data. The file details the calculations for estimating bridge support in real data. (DOCX 18 kb)
Additional file 10: Figure S9.
Histogram of bridge support scores across the data. (DOCX 25 kb)
Additional file 11: Table S2.
Components from the Malhotra data [46] studied by the algorithm. (DOCX 18 kb)
Additional file 12: Figures S10S11.
Results on real samples LUAD 6, LUSC 5. (DOCX 314 kb)
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Eitan, R., Shamir, R. Reconstructing cancer karyotypes from short read data: the half empty and half full glass. BMC Bioinformatics 18, 488 (2017). https://doi.org/10.1186/s1285901719299
Received:
Accepted:
Published:
Keywords
 Cancer
 Karyotypes
 Genome rearrangements
 Structural and numerical variations
 Deep sequencing
 Reconstruction
 Graph theory
 Integer linear programming