 Research
 Open Access
 Published:
MetaCon: unsupervised clustering of metagenomic contigs with probabilistic kmers statistics and coverage
BMC Bioinformatics volumeÂ 20, ArticleÂ number:Â 367 (2019)
Abstract
Motivation
Sequencing technologies allow the sequencing of microbial communities directly from the environment without prior culturing. Because assembly typically produces only genome fragments, also known as contigs, it is crucial to group them into putative species for further taxonomic profiling and downstreaming functional analysis. Taxonomic analysis of microbial communities requires contig clustering, a process referred to as binning, that is still one of the most challenging tasks when analyzing metagenomic data. The major problems are the lack of taxonomically related genomes in existing reference databases, the uneven abundance ratio of species, sequencing errors, and the limitations due to binning contig of different lengths.
Results
In this context we present MetaCon a novel tool for unsupervised metagenomic contig binning based on probabilistic kmers statistics and coverage. MetaCon uses a signature based on kmers statistics that accounts for the different probability of appearance of a kmer in different species, also contigs of different length are clustered in two separate phases. The effectiveness of MetaCon is demonstrated in both simulated and real datasets in comparison with stateofart binning approaches such as CONCOCT, MaxBin and MetaBAT.
Introduction
Studies in microbial ecology commonly experience a bottleneck effect due to difficulties in microbial isolation and cultivation [1]. Due to the difficulty in culturing most organisms in a laboratory, alternative methods to analyze microbial diversity are commonly used to study community structure and functionality.
One such method is the sequencing of the collective genomes (metagenomics) of all microorganisms in an environment [2]. Metagenomics is a study of the heterogeneous microbes samples (e.g. soil, water, human microbiome) directly extracted from the natural environment with the primary goal of determining the taxonomical identity of the microorganisms residing in the samples. It is an evolutionary revise, shifting focuses from the individual microbe study to a complex microbial community. As already mentioned in [3, 4], the classical genomicbased approaches require the prior clone and culturing for the further investigation. However, not all bacteria can be cultured. The advent of metagenomics succeeded to bypass this difficulty.
To further investigate the taxonomic structure of microbial samples, assembled sequence fragments, also known as contigs, need be grouped into bin that ultimately represent genomes. Contig binning serves as the key step toward taxonomic profiling and downstream functional analysis. Therefore, accurate binning of the contigs is an essential problem in metagenomic studies.
Grouping contigs into bins of putative species is one of the hurdles faced when analyzing metagenomic data. Typically, one of a few issues are encountered including: struggling to differentiate related microorganisms, repetitive sequence regions within or across genomes, sequencing errors, and strainlevel variation within the same species, decreasing accuracy for contigs below a size threshold, or excluding low coverage and low abundance organisms [5, 6].
Despite extensive studies, accurate binning of contigs remains challenging [7]. One category is referencebased (supervised), that is, reference databases are needed for the assignment from contigs or reads to meaningful taxons. The classification is either based on homology, or genomic signatures such as oligonucleotide composition patterns and taxonomic clades. Among the most important methods we can recall: Megan [8], Kraken [9], Clark [10], SKraken [11], and MetaPhlan [12].
Referencebased methods require to index a database of target genomes, e.g. the NCBI/RefSeq databases of bacterial genomes, that is used to classify. These methods are usually very demanding, requiring computing capabilities with large amounts of RAM and disk space. Yet, query sequences originating from the genomes of most microbes in an environmental sample lack taxonomically related sequences in existing reference databases. Most bacteria found in environmental samples are unknown and cannot be cultured and separated in the laboratory [13]. For these reasons, when using referencebased methods the number of unassigned contigs can be very high [14, 15].
The other category of methods is referencefree (unsupervised), where studies extract features from contigs to infer bins based on sequence composition [16â€“18], abundance [19], or hybrids of both sequence composition and abundance [5, 20â€“22]. Therefore, these approaches can be applied to bin contigs from incomplete or uncultivated genomes. Some hybrid binning tools, such as CONCOCT [5], MaxBin2.0 [20] and GroopM [21], are designed to bin contigs based on multiple related metagenomic samples. Among these methods, GroopM [21] is advantageous in its visualized and interactive pipeline. On one hand, it is flexible, allowing users to merge and split bins, on the other hand, in the absence of expert intervention, the automatic binning results of GroopM is not as satisfactory as CONCOCT [5]. CONCOCT [5] makes use of the Gaussian mixture model (GMM) to cluster contigs into bins. MetaBAT [22] calculates integrated distance for pairwise contigs and then clusters contigs iteratively by modified Kmedoids algorithm. MaxBin [20] compares the distributions of distances between and within the same genomes.
The composition of DNA, in terms of its constituent kmers, is known to be a feature of the genome. All the above studies are based on the assumption that the kmer frequency distributions of long fragments or whole genome sequences are unique to each genome. More precisely, contig binning using kmers composition is based on the observation that relative sequence compositions are similar across different regions of the same genome, but differ between distinct genomes.
In general, current tools, use the simple kmers counts as signature for a genome, and these counts are normalized, for ease of comparison, in a global fashion. That is all kmers counts are normalized in the same way, irrespective of the contig/species they belong to. Moreover, when the similarity of two contigs is evaluated as the distance of the corresponding kmers counts vectors, not all kmers contributed uniformly to the distance. In fact, because kmers have different probability to appear, the most probable kmers can produce a bias in the distance. In this study, we consider this important observation in order to develop a signature based on kmers statistics that accounts for the different probability of appearance of a kmer in different species. In general, the pairwise comparison of two sequences, or sets of sequences, can be performed with sophisticated similarity measures, based on kmers statistics, derived from research in alignmentfree statistics [23â€“28].
Another important aspect is that long contigs carry more information than short contigs. For this reason long contigs, being more representative, they can be effectively grouped into clusters of candidate species, whereas short contigs are often more noisy.
In this paper, we propose MetaCon a method for contig binning based on a new selfstandardized kmers statistics. The main contributions of MetaCon are the following: it learn the different kmers distributions based on the kmers probabilities in each contig; it uses the information carried by long contigs to guide the construction of clusters; it can estimate the number of species with a simple iterative procedure. A recent independent benchmark [7] has reported as the best binning methods CONCOCT [5] and MetaBat [22]. We tested MetaCon on simulated and real metagenomes and compared the accuracy of binning with CONCOCT [5], MetaBat [22] and MaxBin [20]. MetaCon performs better in terms of precision, recall and estimated number of species on both simulated and real datasets. The results of this study find that MetaCon can generate highquality genomes from metagenomics datasets via an automated process, which will enhance our ability to understand complex microbial communities.
Materials and methods
In this section we present MetaCon and our contribution to the problem of metagenomic contig binning. As we have already observed, most binning tools are based on similarity measures between contigs that are built over kmers frequency distributions.
However, when dealing with a similarity measure based on kmers counts there are two major issues. The first is that kmers might have a different probability to appear in different genomic sequences. The second is that longer contigs carry more information than short contigs, and the direct comparison of the two can be challenging.
The first problem has been extensively studied in the field of alignmentfree measures. The latter, suggest that short contigs should be treated differently. MetaCon addresses these problems by proposing a twophases binning structure in which each phase process one portion of the input dataset, Fig. 1 shows the processing pipeline of MetaCon. We will describe the main steps of the method, giving a brief explanation of the reasons why they were undertaken. In the following subsections each step will be described in details.
Contigs representation
Let us assume that we have N contigs in input that we need to group into bins. First, we construct the feature matrix, using the same notation as [5], where every row corresponds to a single contig that is represented by a (M+V) feature vector where M is the number of features for the coverage and V is the number of features for the composition matrix. This feature matrix has size Nx(M+V), and the two sets of features can be computed independently as follows. Similar to CONCOCT [5], the coverage matrix Y represents the average coverage of contigs in every data sample. More precisely, Y is a NxM matrix where Y_{cm} indicates the coverage of the cth contig in the mth sample. The composition matrix Z of size NxV represents the frequency of every kmers and its reverse complement for the input contigs.
To avoid zero values, a pseudo value is added. For the composition matrix, we add one to it (a relative small number since this matrix counts kmers frequencies), e.g., \(Z_{ij}^{\prime }=Z_{ij}+1\), while for the coverage matrix we modified as \(Y_{ij}^{\prime }=Y_{ij}+0.01\) (a negligible quantity in terms of coverage).
In order to normalize the coverage matrix, we rescale it into different ways. Firstly, across the contigs:
And it is followed by a normalization across samples, within every contig. The coverage profile matrix after this operation is indicated by Q:
Each contig x_{c} is represented by the M coverage features Q_{cm}, with 1â‰¤mâ‰¤M and 1â‰¤câ‰¤N. These normalizations of the coverage matrix have been already used in CONCOCT [5], and other methods.
In this paper we are interested in building a better feature vector for the kmer signature, which serves for the following procedure. We observe that the length of contigs plays an important role with respect to the quality of the kmer signature. Indeed, short contigs may not be a good representer for the genome as they do not carry too much information about the genome, they may not capture the different distributions of kmers as well as long contigs. Furthermore, since the clustering method (e.g., kmedoids) starts from random contigs as centroids, if it happens to be the short contigs, the clustering performance will somewhat degrade. We try to address this issue by splitting the whole dataset into two parts, based on contig lengths. Long contigs will be clustered in the first phase, whereas short contigs will be treated in the second phase.
Phase 1: selfstandardized kmers statistics
Inspired by the recent developments in the field of alignmentfree statistics we propose here a novel similarity measure based on probabilistic kmers statistics for the comparison of two contigs. The idea is to account for the different distribution of kmers counts, in different contigs, and to remove the bias generated by contigs of different length in a probabilistic framework with a selfstandardized kmers statistics. Note that this only applies on the long contigs, whereas we do nothing for the short contigs.
Letâ€™s define contig x_{c}, as a sequence of characters from the alphabet Î£={A,C,G,T}. Letâ€™s say X_{cw} is the frequency of the kmer w in the contig x_{c}. Given that contigs are sequenced from both strands of a genome, X_{cw} also includes the contribution of the reversed complement of w. If k is smaller than the logarithm of the length of contigs, k<logx_{c}, we can consider the variables X_{cw} as Binomial, in line with other studies [29, 30]. Similarly to other methods [22], MetaCon will use k=4, as described in result section, thus this approximation holds.
To account for the different probability of appearance of kmers, we standardize the variables X_{cw} in the following way. For the sequence x_{c}, let \(p_{c}^{j}(a)\) be the probability of the symbol a in position j in x_{c}. If we assume that the symbols at different positions are independent and identically distributed, we can simplify \(p_{c}^{j}\) and denote it by p_{c}. This i.i.d. model has been widely used in field of pattern statistics [31, 32]. Based on this assumption, we define the probability of a kmer w=w_{1}w_{2}...w_{k} to occur at a given position in the contig x_{c} as \(P_{cw}=\prod _{i=1}^{k}{p_{c}(w_{i})}\), that again is independent of the position of x_{c}.
Now, we recall that X_{cw} is a Binomial and that the kmer w has probability to occur P_{cw}, thus can compute mean and variance of X_{cw} as:
where L(x_{c}) is the length of the contig x_{c}. Thus, the kmers counts X_{cw} can be standardized, as a zscore, as follows:
As already observed the frequency of kmers in different genomes can greatly vary. Similarly, it is difficult to estimate the probability P_{cw}, as it does not follow the same model for different genomes. Thus we need to estimate P_{cw} directly from the contig. We define n_{c}(a), with aâˆˆ{G,T,A,C}, as the number of occurrences of the nucleotide a in the contig x_{c}. Then, we can estimate the probability of the symbol a in the contig x_{c} as,
To summarize, we start from the raw kmers counts directly obtained from matrix Z^{â€²}, for each contig we can compute the probabilities P_{cw} and build a probabilistic kmers statistics \(\widetilde {X}_{cw}\) by using formula (3). Similar to the normalization applied to the coverage features, the probabilistic kmers statistics \(\widetilde {X}_{cw}\) is columnwise normalized (normalization across contigs), as H:
Finally, the feature matrix F of long contigs is assembled as F=[QH], as the combination of the coverage profile Q and probabilistic kmers profile H. Then, the relatedness of a pair of contigs can be evaluated by L2 distance of the corresponding feature vectors. Here we use kmedoids clustering method ([33]), a variant of kmeans with the feature matrix F as input.
Phase 2: dealing with short contigs
In the first phase we filter short contigs, and we build clusters with the kmedoids algorithm by using the feature matrix F mentioned above. We process only long contigs in the first phase, because they are more informative in terms of kmers statistics. We believe that the underlying structure of every species can be well unveiled in the first stage when we get rid of the short contigs from the dataset. In fact, the clusters produced in the first phase will have high precision (see result section), because they are more distinguishable and less noisy. These clusters will be used as a basis for the assignment of short contigs in the second phase.
The second subset contains the short contigs, and we decided to assign them to the already classified clusters (output after the first phase) according to the shortest L1 distance. The profile matrix is the union of the composition and coverage matrices of the short contigs. Note that the composition matrix for short contigs is not normalized. L1 distance is an alternative method to measure the similarities between two multidimensional data by computing the absolute distance. In our case, we observed that L1 works better than L2 (Euclidean distance) in the second stage. We think that the L1 distance may somewhat amplify the differences better than L2 in the secondphase where short contigs are less representative.
An overview of MetaCon is presented in Fig. 1. Here we summarize the overall procedure.

1
Compute the composition and coverage matrices.

2
Normalize the coverage matrix.

3
Estimate the number of clusters: C.

4
Split the dataset into two subsets: long and short contigs.

5
Firstphase: Compute the probabilistic kmers signature and normalize the composition matrix of long contigs.

6
Clustering long contigs by kmedoids

(a)
Initialization: randomly select C contigs as the medoids.

(b)
Assignment step: Associate each contig to the closest medoid.

(c)
Update step: For each medoid m and each contig c associated to m swap m and c and compute the total cost of the new configuration, based on the average dissimilarity of c to all contigs associated to m. Select the medoid c with the lowest configuration cost.

(d)
Repeat steps b and c until there is no change in the assignments.

(a)

7
Secondphase: Assign the short contigs to the closest centroid by L1 distance.
Estimating the number of species
As we know, estimating the real number of clusters is one of the most challenging problem. The difficulty primarily attributes to the absence of prior knowledge of the data, in the case of metagenomics the real number of species in the dataset is not known. Moreover, there is no general criteria that may well assess the clusters when we encounter different datasets, in particular, when the number of clusters is big and the data has highdimensional. Despite some methods that are tailored for the datasets with known distribution, here instead we use an easy and intuitive method to estimate the number of species. We exhaustively iterate the kmeans by starting from a small number of clusters and gradually increase it until some criteria is met. This procedure stops when the nonempty clusters are less than 80% of the candidate number of cluster in the corresponding iteration. This iterative procedure might be computationally demanding, to speed up the computation in this paper we use an efficient library implementation [34].
Results and discussion
In order to validate our contribution, we compare it with the commonly known methods CONCOCT, MaxBin 2.0 and MetaBat. In particular, CONCOCT [5] and MetaBat [22] have been reported to be the best performing methods in a recent independent benchmark [7]. All of these tools use as input the composition and coverage matrices, as MetaCon does. MaxBin 2.0 [20] estimates the probability that a contig belongs to a bin based on expectationmaximization (EM). MetaBat [22] starts from one bin, and gradually assigns the contigs to that bin until the centroid does not change, repeatedly for several bins until no contigs are left. CONCOCT [5] applies PCA (principal component analysis) to the feature matrix (composed by coverage and composition matrices) for the sake of dimension reduction and afterward it uses a Gaussian mixture model.
Synthetic and real datasets
Before the discussion of the results, here we give a brief introduction of the datasets. In this paper, we test the methods on both synthetic and real metagenomic datasets. A complete description of the dataset construction can be found in [5], here for completeness we report a brief summary. In CONCOCT [5], the authors simulate two mocked communities of microbiomes in order to test the performance, called â€™Strainâ€™ and â€™Speciesâ€™ datasets. Both of these synthetic datasets are built on 16S rRNA samples involved in the Human Microbiome Project (HMP, [35]). The samples have gone through denoise operation and the OTUs were generated by the standard that 3% sequence differences to approximated species exist. The contigs were assembled from the reads in samples and the reads were subsequently mapped back onto contigs to get the coverage information.
For simulated and real data, a coassembly of reads from all samples was performed using Ray [36]. Ray was used to generate the coassembled contigs because it is able to handle large metagenomic dataset. Contigs were cut up into nonoverlapping fragments of 10 kilobases in order mitigate the effect of local assembly errors (Additional file 1: Figure S1 in supplementary material reports the contig length distribution).
Specifically, the â€™Strainâ€™ dataset contains 9417 contigs, which are coassembled from 64 samples, and it contains totally 20 organisms. The simulated â€™Strainâ€™ community is composed by five different Escherichia coli strains, five different Bacteroides species, five different gut bacteria and the rest from Clostridium.
The â€™Speciesâ€™ dataset has 101 different species, including 37628 contigs, coassembled from 96 samples. For the â€™Speciesâ€™ dataset, OTUs are removed when its total count is less then 20 across samples. This dataset aims at testing the ability to discriminate at specieslevel. The complete information for the datasets can be found in the supplementary material.
â€™Sharonâ€™ [37] is a real dataset, and it is generated from the microbiome samples of the premature infants. It contains 18 data samples, and due to the fact that we do not know the true species labels, we used TAXAassign [38] to annotate the contigs. It ended up with 7 species, 2599 contigs, after we filtered the contigs with ambiguous labels at species level.
Evaluation criteria
Precision and recall are commonly used to compare the performance of the binning algorithms under assessment. Precision measures the ability of the approach to build clusters composed by contigs from the same species. On the other hand, recall measures the ability of gathering all the contigs of a given species. Namely, the precision tests the correctness, and the recall tests the completeness. Therefore, when evaluating the performance of a binning method one should take into account both aspects in order to obtain a comprehensive evaluation.
Let n be the number of species in a metagenomic dataset, and C be the number of clusters returned by the algorithm. Let A_{ij} be the number of contigs from species j assigned to cluster i. Following the definitions in [39], for the precision we find the species with the maximum number of contigs in every cluster and sum them up, divided by the total number of contigs. As for the recall, we select the cluster with the maximal number of contigs from a given species, and again accumulate them, divided by the total number of contigs.
Results on Synthetic and Real Datasets
In the first experiment, we assess the ability of MetaCon to predict the number of clusters. The average result is reported in (Table 1). CONCOCT needs a maximal number of cluster in input, the other methods do not. In this first experiment, MetaCon outperforms the other methods by estimating the number of clusters close to the real number of species.
In next series of tests we evaluated the performance of MetaCon on the datasets against the other tools. MetaCon outperforms all other methods in terms of precision and recall, as shown in Figs. 2, 3 and 4. The precision and recall are above 95% for both simulated data and real data. For the â€™Strainâ€™ dataset (Fig. 2), the precision by MetaCon is about 97.5%, that is better than the other three methods; the recall is 95.8%, higher than MaxBin and MetaBat, almost identical with CONCOCT. For the â€™Speciesâ€™ dataset, shown in Fig. 3, it is challenging to bin the contigs since the number of species is large, MetaCon reaches 99.3% in terms of precision and 94.6% for the recall. Again, the comparison with other tools reveals an outcome similar to the dataset â€™Strainâ€™. For the real dataset â€™Sharonâ€™, Fig. 4, the results are in line with those of the synthetic datasets. MetaCon achieves higher precision and recall with respect to the other tools. The only notable difference is that on this datasets MetaBat has a precision similar to MetaCon but again a lower recall.
Additionally, we evaluate the quality of bins generated by different methods for Strain dataset. In order to evaluate the contamination and completeness of the bins, we filtered out the bins whose precision is less than 80%, reported in Fig. 5a, where the different shades of gray indicates the different level of recall. In Fig. 5b, we report the opposite procedure where we assess the precision of bins after filtering out bins with recall lower than 80%. For example, in Fig. 5a the number of clusters with precision greater than 80% and recall greater than 95% is 16 for MetaCon, for CONCOCT 11 and for MaxBin 4. MetaCon outperforms the other methods, firstly MetaCon has more bins left after screening in both Fig. 5a and b. Secondly, the bins produced by MetaCon mostly resides in the highlevel range of precision and recall. We think that the primary reason for the good performance of MetaCon is that the firststage builds highquality clusters, they may better represent the relative species and capture the different traits of species. In addition, the kmedoids may relieve the negative influence caused by the outliers since it considers the median value instead of the mean during the clustering process, and probably it further consolidates the structures of clusters.
Parameters: kmers size
In this section, we want to discuss how to choose the parameters k for MetaCon and show the results under different conditions. The selection of the kmers size is critical when we build our probabilistic kmers signature, if k is too small (k=2), it will result a less representative and informative feature matrix as only 16 features of composition matrix generated, that is not sufficient to differentiate between contigs from diverse species, specially, when some of them are closely related.
With this series of experiments we want to evaluate the influence of kmers size for MetaCon over the different datasets. The results of precision and recall are reported in Figs. 6 and 7. Note that the results are obtained with the correct number of clusters used as input since here we want to compare the various choice for the size of kmers. For the â€™Strainâ€™ dataset, the precision increases from 93% to 97% when k varies from 2 to 6, and the precision is identical when k equal to 4 and 6, when k equal to 8 the precision decreases. The recall of â€™Strainâ€™ dataset follows the similar trend of the precision. As for the â€™Speciesâ€™ dataset, the precision changes from 95% to 99% when k increases from 2 to 6, and it achieves 99% when k equal to 6, slightly better than with k equal to 4. The â€™Sharonâ€™ dataset keeps the precision and recall constant when we modify k as the number of species is small, probably less information is required in order to cluster the contigs with respect to the other two datasets that contain more species. Generally speaking, k equal to 4 could be a good choice, considering precision, recall and computing time. A similar value was used also in MetaBat [22].
The importance of contig length distribution
Another factor we want to address here is the importance of the length of contigs. Recall that in the first phase of MetaCon, we process only long contigs, and in the second phase we assign the short contigs. We want to evaluate the impact of this approach by showing how precision and recall varies in the two phases and to compare the results when we use all the contigs at once. The results of these experiments are reported in Tables 2 and 3. If we do not process the long and short contigs separately in twophases we can observed that the precision obtained by MetaCon is respectively 93.79% and 97.23% for the two datasets â€™Strainâ€™ and â€™Speciesâ€™. We can notice that the precision can be improved by separately processing the contigs: for the â€™Strainâ€™ dataset, it increases from 93.79% to 97.46%, and for â€™Speciesâ€™ it improves from 97.23% to 99.56%. A similar behavior is observed for the recall. For the â€™Strainâ€™ dataset the recall increases slightly from 95.23% to 95.78%, and for the â€™Speciesâ€™ dataset it improves from 90.95% to 95.04%.
In order to choose a good threshold to split short and long contigs, we experimented with different values (see Additional file 1: Figure S2 in supplementary material). Empirically we found that a good choice is to have about 20% of the contigs to be labelled as short. Based on these results we selected 2000bp as a good compromise.
Assignment of short contig: L1 vs L2 distance
In the second phase short contigs are assigned to the closest centroid by L1 distance. Here, we evaluate the effect of L1 distance in comparison with L2 distance. Figure 8 reports the precision of MetaCon for all datasets individually by L1 and L2 distance in the second phase. For this stage, the L1 distance outperforms L2 for all of the datasets. In particular, in the â€™Strainâ€™ dataset L1 boosts the performance from 89.56% to 97.46%, for â€™Sharonâ€™ dataset, the precision increases from 79.11% to 97.38% by using L1. We think, when it comes to assign the short contigs to the closest cluster centroid, L1 reveals its strength by amplifying the differences between contigs.
Conclusion
Binning metagenomic contigs remains a crucial step in metagenomic analysis. In this work we presented MetaCon, an unsupervised approach for metagenomic binning based on probabilistic kmers statistics and coverage. Our approach instead of processing the whole dataset at once as most methods, it splits the input and process them into two separate phases of MetaCon. We compared the binning performance over synthetic and real metagenomic datasets against other stateofart binning algorithms, showing that MetaCon achieves good performances in terms of precision, recall and estimating the number of species.
Abbreviations
 GMM:

Gaussian mixture model
 HMP:

Human microbiome project
 OTU:

Operational taxonomic unit
 PCA:

Principal component analysis
References
Staley JT, Konopka A. Measurement of in situ activities of nonphotosynthetic microorganisms in aquatic and terrestrial habitats. Ann Rev Microbiol. 1985; 39(1):321â€“46. https://doi.org/10.1146/annurev.mi.39.100185.001541.PMID: 3904603.
Handelsman J, Rondon MR, Brady SF, Clardy J, Goodman RM. Molecular biological access to the chemistry of unknown soil microbes: a new frontier for natural products. Chem Biol. 1998; 5(10):245â€“9. https://doi.org/10.1016/S10745521(98)901089.
Felczykowska A, Bloch SK, NejmanFaleÅ„czyk B, BaraÅ„ska S. Metagenomic approach in the investigation of new bioactive compounds in the marine environment. Acta Biochim Pol. 2012; 59(4):501â€“5.
Mande SS, Mohammed MH, Ghosh TS. Classification of metagenomic sequences: methods and challenges. Brief Bioinforma. 2012; 13(6):669â€“81. https://doi.org/10.1093/bib/bbs054. Accessed 25 Aug 2015.
Alneberg J, Brynjar SmÃ¡ri B, Ino DB, Melanie S, Joshua Q, Umer Z I, Leo L, Nicholas J L, Anders F A, Christopher Q. Binning metagenomic contigs by coverage and composition. Nat Methods. 2014; 11:1144â€“6.
Bowers RM, Clum A, Tice H, Lim J, Singh K, Ciobanu D, Ngan CY, Cheng JF, Tringe SG, Woyke T. Impact of library preparation protocols and template quantity on the metagenomic reconstruction of a mock microbial community. BMC Genomics. 2015; 16(1):856. https://doi.org/10.1186/s1286401520636.
Sczyrba A, Hofmann P, McHardy AC. Critical assessment of metagenome interpretationâ€”a benchmark of metagenomics software. Nat Methods. 2017; 14:1063â€“71.
Huson DH, Auch AF, Qi J, Schuster SC. Megan analysis of metagenomic data. Genome Res. 2007; 17. https://doi.org/10.1101/gr.5969107.
Wood D, Salzberg S. Kraken: ultrafast metagenomic sequence classification using exact alignments. Genome Biol. 2014; 15. https://doi.org/10.1186/gb2014153r46.
Ounit R, Wanamaker S, Close TJ, Lonardi S. Clark: fast and accurate classification of metagenomic and genomic sequences using discriminative kmers. BMC Genomics. 2015; 16(1):1â€“13. https://doi.org/10.1186/s1286401514192.
Qian J, Marchiori D, Comin M. Fast and sensitive classification of short metagenomic reads with skraken In: In: Peixoto N, Silveira M, Ali HH, Maciel C, van den Broek EL, editors. Biomedical Engineering Systems and Technologies. Cham: Springer: 2018. p. 212â€“26.
Segata N, Waldron L, Ballarini A, Narasimhan V, Jousson O, Huttenhower C. Metagenomic microbial community profiling using unique cladespecific marker genes. Nat Methods. 2012; 9. https://doi.org/10.1038/nmeth.2066.
Eisen JA. Environmental shotgun sequencing: its potential and challenges for studying the hidden world of microbes. PLoS Biol. 2007; 5(3):e82.
Lindgreen S, Adair KL, Gardner P. An evaluation of the accuracy and speed of metagenome analysis tools. Sci Rep. 2016; 6:19233.
Girotto S, Comin M, Pizzi C. Higher recall in metagenomic sequence classification exploiting overlapping reads. BMC Genomics. 2017; 18(10):917. https://doi.org/10.1186/s1286401742736.
Kislyuk A. Unsupervised statistical clustering of environmental shotgun sequences. BMC Bioinformatics. 2009; 10. https://doi.org/10.1186/1471210510316.
Kelley DR, Salzberg SL. Clustering metagenomic sequences with interpolated markov models. BMC Bioinformatics. 2010; 11. https://doi.org/10.1186/1471210511544.
Strous M. The binning of metagenomic contigs for microbial physiology of mixed cultures. Front Microbiol. 2012; 3. https://doi.org/10.3389/fmicb.2012.00410.
Leung HCM, Yiu SM, Yang B, Peng Y, Wang Y, Liu Z, Chen J, Qin J, Li R, Chin FYL. A robust and accurate binning algorithm for metagenomic sequences with arbitrary species abundance ratio. Bioinformatics. 2011; 27(11):1489â€“95. https://doi.org/10.1093/bioinformatics/btr186. Accessed 09 Mar 2015.
Wu YW, Simmons BA, Singer SW. Maxbin 2.0: an automated binning algorithm to recover genomes from multiple metagenomic datasets. Bioinformatics. 2016; 32(4):605â€“7. https://doi.org/10.1093/bioinformatics/btv638.
Imelfort M. Groopm: an automated tool for the recovery of population genomes from related metagenomes. PeerJ. 2014; 2. https://doi.org/10.7717/peerj.603.
Kang DD, Froula J, Egan R, Wang Z. Metabat, an efficient tool for accurately reconstructing single genomes from complex microbial communities. PeerJ. 2015; 3:1165. https://doi.org/10.7717/peerj.1165.
Kantorovitz MR, Robinson GE, Sinha S. A statistical method for alignmentfree comparison of regulatory sequences. Bioinformatics. 2007; 23. https://doi.org/10.1093/bioinformatics/btm211.
Sims GE, Jun SR, Wu GA, Kim SH. Alignmentfree genome comparison with feature frequency profiles (ffp) and optimal resolutions. Proc Nat Acad Sci. 2009; 106. https://doi.org/10.1073/pnas.0813249106.
Comin M, Antonello M. Fast entropic profiler: An information theoretic approach for the discovery of patterns in genomes. IEEE/ACM Trans Comput Biol Bioinformatics. 2014; 11(3):500â€“9. https://doi.org/10.1109/TCBB.2013.2297924.
Comin M, Verzotto D. Beyond fixedresolution alignmentfree measures for mammalian enhancers sequence comparison. IEEE/ACM Trans Comput Biol Bioinforma. 2014; 11(4):628â€“37. https://doi.org/10.1109/TCBB.2014.2306830.
Comin M, Leoni A, Schimd M. Clustering of reads with alignmentfree measures and quality values. Algoritm Mol Biol. 2015; 10(1):4.
Comin M, Antonello M. On the comparison of regulatory sequences with multiple resolution entropic profiles. BMC Bioinformatics. 2016; 17(1):130. https://doi.org/10.1186/s1285901609802.
Lippert RA, Huang H, Waterman MS. Distributional regimes for the number of kword matches between two random sequences. PNAS. 2002; 99(22):13980â€“9. https://doi.org/10.1073/pnas.202468099.http://www.pnas.org/content/99/22/13980.full.pdf.
G R, D C, F S, MS W. Alignmentfree sequence comparison (i): statistics and power. J Comput Biol. 2009; 16(12):1615â€“34.
RÃ©gnier M. A unified approach to word occurrence probabilities. Discret Appl Math. 2000; 104(1):259â€“80. https://doi.org/10.1016/S0166218X(00)001955.
Song K, Ren J, Reinert G, Deng M, Waterman MS, Sun F. New developments of alignmentfree sequence comparison: measures, statistics and nextgeneration sequencing. Brief Bioinforma. 2014; 15(3):343â€“53. https://doi.org/10.1093/bib/bbt067.
Kaufmann L, Rousseeuw P. Clustering by means of medoids In: Dodge Y, editor. Data Analysis based on the L1Norm and Related Methods. NorthHolland: Elsevier: 1987. p. 405â€“16.
Chen M. Super fast and terse kmeans clustering. 2017. https://nl.mathworks.com/matlabcentral/fileexchange/24616kmeansclustering.
Consortium HMP. Structure, function and diversity of the healthy human microbiome. Nature. 2012; 486(7402):207â€“14. https://doi.org/10.1038/nature11234. Accessed 07 June 2016.
Boisvert S, Raymond F, Godzaridis Ã‰, Laviolette F, Corbeil J. Ray meta: scalable de novo metagenome assembly and profiling. Genome Biol. 2012; 13(12):122. https://doi.org/10.1186/gb20121312r122.
Sharon I, Morowitz MJ, Thomas BC, Costello EK, Relman DA, Banfield JF. Time series community genomics analysis reveals rapid shifts in bacterial species, strains, and phage during infant gut colonization. Genome Res. 2013; 23(1):111â€“20. https://doi.org/10.1101/gr.142315.112. http://genome.cshlp.org/content/23/1/111.full.pdf+html.
Ijaz et al A. Taxaassign v4.0. 2013. http://github.com/umerijaz/taxaassign.
Vinh LV, Lang TV, Binh LT, Hoai TV. A twophase binning algorithm using lmer frequency on groups of nonoverlapping reads. Algoritm Mol Biol. 2015; 10(1):1â€“12. https://doi.org/10.1186/s1301501400304.
Acknowledgments
Not applicable.
Funding
Publication of this article did not receive sponsorship.
Availability of data and materials
The software is freely available for academic use at: http://www.dei.unipd.it/Ùª7Eciompin/main/metacon.html
About this supplement
This article has been published as part of BMC Bioinformatics, Volume 20 Supplement 9, 2019: Italian Society of Bioinformatics (BITS): Annual Meeting 2018. The full contents of the supplement are available at https://bmcbioinformatics.biomedcentral.com/articles/supplements/volume20supplement9
Author information
Authors and Affiliations
Contributions
All authors contributed to the design of the approach and to the analysis of the results. JQ implemented MetaCon software and performed the experiments. JQ and MC conceived the study and drafted the manuscript. All authors have read and approved the manuscript for publication.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Publisherâ€™s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Additional information
Authorsâ€™ information
Department of Information Engineering, University of Padova via Gradenigo 6/A, 35131 Padova  Italy
Email addresses: JQ (jia.qian@studenti.unipd.it), MC (comin@dei.unipd.it).
Additional file
Additional file 1
Supplementary Material. (PDF 205 kb)
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Qian, J., Comin, M. MetaCon: unsupervised clustering of metagenomic contigs with probabilistic kmers statistics and coverage. BMC Bioinformatics 20 (Suppl 9), 367 (2019). https://doi.org/10.1186/s1285901929044
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1285901929044
Keywords
 Metagenomics
 Unsupervised clustering
 Kmers statistics