 Software
 Open Access
 Published:
MIA: Mutual Information Analyzer, a graphic user interface program that calculates entropy, vertical and horizontal mutual information of molecular sequence sets
BMC Bioinformatics volume 16, Article number: 409 (2015)
Abstract
Background
Short and long range correlations in biological sequences are central in genomic studies of covariation. These correlations can be studied using mutual information because it measures the amount of information one random variable contains about the other. Here we present MIA (Mutual Information Analyzer) a user friendly graphic interface pipeline that calculates spectra of vertical entropy (VH), vertical mutual information (VMI) and horizontal mutual information (HMI), since currently there is no user friendly integrated platform that in a single package perform all these calculations. MIA also calculates JensenShannon Divergence (JSD) between pair of different species spectra, herein called informational distances. Thus, the resulting distance matrices can be presented by distance histograms and informational dendrograms, giving support to discrimination of closely related species.
Results
In order to test MIA we analyzed sequences from Drosophila Adh locus, because the taxonomy and evolutionary patterns of different Drosophila species are well established and the gene Adh is extensively studied. The search retrieved 959 sequences of 291 species. From the total, 450 sequences of 17 species were selected. With this dataset MIA performed all tasks in less than three hours: gathering, storing and aligning fasta files; calculating VH, VMI and HMI spectra; and calculating JSD between pair of different species spectra. For each task MIA saved tables and graphics in the local disk, easily accessible for future analysis.
Conclusions
Our tests revealed that the “informational model free” spectra may represent species signatures. Since JSD applied to Horizontal Mutual Information spectra resulted in statistically significant distances between species, we could calculate respective hierarchical clusters, herein called Informational Dendrograms (ID). When compared to phylogenetic trees all Informational Dendrograms presented similar taxonomy and species clusterization.
Background
Genic and intergenic regions in chromosomes have statistically distinct properties. Most intergenic regions behave randomly in regard to nucleotide mutations, besides some special regions like transcription regulatory sites and transposons. On the other hand, most of the genes are highly conserved, especially in regions like promoters, TATA box, exons, splice junctions etc. But, even these conserved regions, and others such as introns, can present many polymorphic single regions as well as two separated regions that present orchestrated mutations in a way as to try to conserve or improve determined phenotype [1–4]. Hereupon, the main goal of this work is to provide methods and tools in order to discriminate closely related species using informational spectrum distances. Here we searched for polymorphic regions seeking out covariation signals by using DNA sequences from Drosophila Adh locus. With these sequences we calculated entropy and mutual informational spectra from different closely related species sequences. Thereafter, via JensenShannon Divergence, we calculated distances between these spectra resulting in distance matrices capable of inferring the possibility of species discrimination.
Furthermore, in microorganisms the species definition is coarse. Attempts to measure sequence diversity by informational entropy and speciation have been proposed in Protista [5, 6]. However, these studies suggest that besides information entropy, mutual information could provide a means to access covariation, a central problem in diversifying molecules and species.
Thus, herein we present a computational pipeline called MIA (Mutual Information Analyzer) developed in Python [7] and BioPython [8]. MIA modules can be seen in Fig. 1, and it is capable of retrieving DNA sequences, and of calculating Entropy and Mutual Information spectra applying a statistical framework that allows inferences. This pipeline was developed due to the inexistence of an application able to calculate and display vertical Shannon entropy (VH), vertical mutual information (VMI), horizontal mutual information (HMI), and JensenShannon Divergence (JSD) between pairs of different species spectra, herein called informational distances.
Entropy and Mutual Information theory can be found in [9, 10], and a nice review in [11]. Many applications [12–14] are capable of calculating biological sequence parameters, but only BioEdit [15] calculates mutual information and vertical entropy. There are also some theoretical studies like in Grosse et al. [1] focused on horizontal mutual information of DNA sequences, and other studies of coevolution of proteins [3, 4, 16, 17] based on vertical mutual information. However none of these studies and applications has algorithms to calculate and display VH, VMI and HMI distributions and their informational distances.
Therefore, we began the development of MIA guided by the questions: Given some sequences, grouped in recognizable sets (species), are molecular data capable of discriminating theses sets using methods of information theory? Can we present statistical calculations that confirm or deny the results? Given a set of sequences with conserved and polymorphic residues and different lengths, how to deal with many possible alignments and their gaps?
In order to address these problems we demonstrated that Entropy and Mutual Information are good methods to deal with this complex problem, but caveats and warnings still remain: a) multiple sequence alignment (MSA) gives rise to many possible alignments herein denoted “mincut” for the minimum length alignment and “maxmer” for the maximum; b) gaps should be replaced by vertical consensus residues, differently from Weil et al. [4] who replaced them with a new character (a 21th amino acid) for their protein study, otherwise covariation between two residues will give rise to new states containing a strange fifth nucleotide; c) short strings present deviation called bias, thus we applied a bias correction for entropy, mutual information and respective standard errors as defined in Roulston [18] and also demonstrated by Steuer et al. in [11]; finally d) mutual information can be calculated residue by residue (1 x 1 positions) or 2 by 2 residues, or n by n residues  this parameter we will call NOL (number of letters of a word) and herein NOL will be equal to one (see Fig. 4b).
Implementation
MIA has the following algorithms: A1) NCBI: gathers data at NCBI and stores them in gbk file format; A2) gbk to fasta: analyzes gbk file and organizes sequences in fasta files per species; A3) Alignment: aligns sequences with Muscle [19] and in the end creates two fasta files: “mincut” cutting out columns and sequences with large gaps and “maxmer” maintaining maximum number of gaps; A4) Purging: replaces ambiguous nucleotides via IUPAC nucleotide ambiguity table, and eliminates sequences with undesirable words in their names like “synthetic”; A5) Consensus: replaces gaps by their vertical consensus nucleotide; A6) VMI: calculates and stores Vertical Entropy (VH) and Vertical Mutual Information (VMI) spectra, and displays respective histograms and heat maps; A7) HMI: calculates and stores Horizontal Mutual Information (HMI) spectra, and displays histograms; A8) JSD: calculates JensenShannon Divergence from pair of normalized spectra, storing distances and their SE in distance matrix files, and displays distance histograms; A9) HC: calculates hierarchical clusters and presents them as dendrograms, herein called informational dendrograms; A10) Entropy: simulates Shannon entropy.
Before gathering sequences it is important to analyze available Drosophila species data. There are some sites specialized in Drosophila data. They present sequence browsers, protein sequences, genes sequences and many parameters for molecular data. Three of these sites are: DPDB [20], Flybase [21] and BDGP [22].
In the first algorithm MIA is capable of searching for an organism, a gene or a word in the NCBI GBK. Here we searched for organism/genus “Drosophila” and the gene “Adh” (alcohol dehydrogenase). The resulting search retrieved 959 sequences of 291 species. From the total, 450 sequences of 17 species were selected (data gathered in March 2015). We did this task imposing an inferior limit called “number of sequence cutoff” in such a way that if this cutoff is high MIA finds a set with fewer species (there are not many genera/genes with a lot of sequences in NCBI). Otherwise, if the cutoff is low, the set will have a larger number of species, some of which with a low number of sequences. The consequence is that when calculating entropy and mutual information, a species with many sequences provides a lower standard error, while a species with fewer sequences provides a higher standard error. Therefore, in the Drosophila/Adh case we set the cutoff equal to 7 resulting in 17 species and 450 sequences.
After gathering sequences, the next step was the alignment algorithm  MIA uses Muscle for this task  and thereafter starts deleting columns and sequences “with many gaps” (which gives rise to the question – what is the “correct percentage of maximum gaps”?) replacing them by consensual vertical residues. However, deletions and replacements alter the informational distribution profiles. The human decision to set the percentage of possible gaps creates “mincut” and “maxmer” alignment sequences and their informational difference can be analyzed comparing distance matrices. Answering the question posed, “there is no correct choice” to how to deal with controlling gap deletions; only empirical tests and their results are likely ways to solve this problem in each genus/gene case.
With the aligned sequences MIA computes vertical entropy like did Adami in [9]. Thereafter MIA calculates mutual information in the horizontal direction as in Grosse et al. [1, 2, 23, 24], and in the vertical direction as in Martin et al. [16] and Hamacher et al. [3, 4]. All these methods are well explained in the Methods section that follows.
VH, HMI and VMI are calculated with and without bias correction; therefore the gain or loss of information for “mincut” versus “maxmer” with or without bias correction can be compared. Informational distances between different informational spectra are calculated via JSD method. Since JSD is not a linear function, standard errors are calculated by empirical propagation giving rise to distance matrices with SE.
ANOVA test was performed on each set of spectra for each method (VH, VMI, and HMI), in order to assess whether at least one spectrum was statistically different from the others (see Additional file 1: drosophila “summary” tab). Otherwise, all distributions would be statistically similar and we could not discriminate species. As can be noticed, this was only a first test to verify whether we could move forward.
All informational spectra were compared to spectra of shuffled and random sequences in order to analyze if they are statistically distinct. Spectra of shuffled sequence were generated using original sequences and shuffling the residues. Spectra of random sequence were created drawing nucleotides randomly, up to the same length as the original sequences. The first method preserves nucleotide contents and the latter is fully random tending to 25 % of representation to each nucleotide. In this study we will present only spectra of shuffled sequences and omit the random ones, since they presented similar results. Notwithstanding, MIA calculates and presents spectra for both methods.
Methods
Shannon Entropy was defined in 1948 [25] as weighted average of the log of state probabilities,
Equation 1 – Shannon entropy.
In Equation 1 H is the Shannon Entropy, p_{m} is the probability of the existence of a state m in k possible states, and b is the base of the logarithm. If b is equal to 2 the entropic unit is defined as a “bit”, but if b is “e” (Neper’s number) the entropic unit is defined as “nat”, entropy derived from natural log. In this paper we will work only with “nat”.
Shannon Entropy of DNA sequences can be measured aligning sequences and calculating the relative frequency of encountering each nucleotide at determined residue (site). In this particular case only 4 possible states are found, Ω = {A, C, G, T}. Maximum entropy is defined as 1 MER [9], and it is achieved only when all states are equiprobable (p = 0.25). In this case H ^{DNA}_{ max } = − ∑ ^{4}_{ i = 1} p _{ i } log_{2} p _{ i } = − 4 ∗ (1/4 ∗ log_{2}1/4) = 2 bit or 1.386 nat. However, if one of the states has frequency equal to 1 (100 %), and the others are 0, the resulting entropy is H = 0 because log 1 = 0 and this residue is said to be conserved.
We shall call Vertical Entropy (VH) the way of calculating entropy counting nucleotide frequencies in the vertical orientation, that means, is a measure of how polymorphic any residue is after aligning sequences for determined species.
In Fig. 2, we present an oversimplified alignment for 8 sequences: a) S1, S2 to S8 are 8 aligned sequences, i = {1,2,3,4} are nucleotide positions (or residues), each cell represents a nucleotide for a given sequence at a given position i; b) is the nucleotide frequency distribution per site for each residue; c) is the relative nucleotide frequency distribution per site; d) is the vertical entropic spectrum; e) are the values for each nucleotide entropic contribution and at the bottom is the vertical total entropy of each residue.
Thus, given the position i = 1, we realize that all nucleotides are “A”. For i = 2 we realize that 50 % of nucleotides are “A” and 50 % are “G”. As previously discussed, the first position has H = 0 nat and the residue is conserved, the second position has H = .693 nat and the residue is polymorphic. For i = 3, see the red arrow, we realize that this position is more polymorphic than the previous (see Fig. 2d) but less polymorphic than i = 4 that has H[4] = H_{max} = 1 MER.
Mutual Information (MI) represents the covariation between two random variables, here denoted X and Y [10, 11, 24]. Mutual Information (Equation 2) is defined by the sum of two entropies, in position i and j, minus the joint entropy H(i,j). As shown in the next two sections, MI can be applied in the Vertical direction of aligned sequences or in the Horizontal direction for one single or many sequences, aligned or not. Both calculations measure the nucleotide variability in two positions. The first position i is represented by the random variable X, and the second position j is represented by the random variable Y. Therefore, MI(X,Y) can be defined as,
Equation 2 – Mutual Information between two positions (i, j).
Another parameter for MI calculation is the size of the word, whose width is defined as number of letters (NOL). Therefore, we can analyze covariation between regions with width greater than 1. However, in this study we will calculate MI only for NOL = 1 (see Fig. 4b).
In order to calculate the Vertical Mutual Information (VMI), see Fig. 3, we need first to estimate the nucleotide frequencies for position i and j, covering all residue pairs. Since, MI(X,Y) is a symmetric function, in other words, MI(X,Y) = MI(Y,X), and MI is zero if X covaries independently of Y, covering (i,j) in such a way that i = {1, L1} and j = (i + 1, L), for j > i. VMI can also be expressed by,
Equation 3 – Vertical Mutual Information.
In Equation 3 m and n are nucleotide states = {A,C,G,T}. Since we are talking about a bidimensional relationship, the resulting spectrum is represented by a heat map.
In Fig. 3, we present an oversimplified example of aligned sequence with the intention of explaining how to calculate VMI. The blue and red arrows point to a particular pair (X,Y), but all possible pairs are calculated where: a) S1, S2 … to S8 are 8 aligned sequences and i or j = {1,2,3,4} are nucleotide positions (or residues); b) is the nucleotide relative frequency distribution for all pairs of nucleotides; c) is the vertical entropy for each residue and at the bottom the vertical entropy per residue; d) is the mutual information calculation for pair of residues; and e) on the left is the VMI bidimensional spectrum represented as a heat map, on the right side is the color scale.
Horizontal Mutual Information (HMI) has a different concept and method of calculation when compared to VMI. HMI is defined as a measure of autocovariation between two positions distant k units one from the other. Here k varies from 3 to L/2 (where L is the sequence length) with step equal to 1 in the 5′ to 3′ direction. The step is one, because we intend to calculate all residue to residue covariations in the gene. In other words, transcription and translation rules are not necessary in our study.
For each value of k all sequence is covered counting all possible pair of nucleotides (m,n) Є{AA, AC, … TT}. Here, p_{mn} represents the probability to find a pair (m,n), where m and n Є {A,C,G,T}. HMI(k) is given by Equation 4, and p_{m} (k) and p_{n} (k) are marginal probabilities given by Equation 5 and Equation 6, respectively.
Equation 4 – HMI equation for DNA sequences.
The marginal probabilities (p_{m}(k)e p_{n}(k)) can be calculated as,
Equation 5 – Marginal probability of nucleotide m is p_{m} (k)
and
Equation 6 – Marginal probability of nucleotide n is p_{n} (k).
In Fig. 4, we present an oversimplified example of aligned sequence with the intention of explaining how to calculate HMI. In this example we see: a) k = 3 and the cursor covering the sequence from the left to the right finding = {AA, AA,GG…}; b) in this study NOL = 1, but NOL can be any other integer greater than 1; c) the marginal probabilities for X; d) the marginal probabilities for Y; e) the joint frequencies and relative joint frequencies for each found pair; f) HMI spectrum; and g) calculation of HMI for this particular case.
With all calculated spectra we used JensenShannon Divergence (JSD) method to calculate the distances between all pairs of different informational species spectra, for VH, VMI and HMI. Since JSD needs two distributions to calculate a distance, we had to normalize the spectra and then apply them to this method. Therefore, we can calculate JSD to VH, VMI and HMI in order to calculate all distances between pairs of different species,
Equation 7 – JensenShannon Divergence (JSD).
Where P and Q are normalized spectra for different species. Equation 7 is the JSD equation, but the square root of JSD is indeed the distance between two distributions [26]. A distance equal to zero means that we cannot discriminate two distributions. A short distance means that both distributions are close and perhaps statistically impossible to discriminate depending on SE. Large distances means that species spectra are far allowing their discrimination, but also dependent on the SE. Thus, it can be inferred that JSD discriminates species with 95 % of probability if most of the distances do not fall in the confidence intervals (CI) of all the others – where CI ~ distance ± 2*SE.
Results
We tested our algorithms searching in NCBI, at nucleotide database, for Organism = “Drosophila” and Gene = “Adh” (Alcohol Dehydrogenase) resulting in 959 sequences of 291 species. Only species with 7 or more sequences available were accepted to minimize the vertical entropy and mutual information standard errors. The final result was 450 sequences and 17 species, with lengths between 405 and 2204 bp. After going through the first two modules we encountered the Alignment module having 3 parameters which were designed to control column and line (sequences) gap deletions. The first parameter “Maximum Vertical Gaps1” (set to 10 %) allows gaps up to this percentage and transforms the data in minimum length sequences, or “mincut”. A maximum length sequence is obtained with the parameter “Maximum Vertical Gaps2” (set to 40 %) which allows more gaps, and whose resulting sequences are called “maxmer”. The third parameter “Maximum Horizontal Gaps” (set to 40 %) cuts out all sequences presenting more than 40 % of horizontal gaps. The resulting aligned sequences can be seen calling an external program called Seaview [12].
All aligned sequences resulted in “mincut” length equal to 588 bp and “maxmer” length equal to 859 bp. After this procedure, sequences were purged/filtered and not “ACGT” nucleotides replaced with their consensus via IUPAC [27] ambiguous table. Finally, consensus algorithm substitutes all gaps by the vertical consensus nucleotide.
VH, HMI and VMI were calculated for “mincut” and “maxmer” with and without bias correction. The final results presented twelve distance matrices, twelve distance histograms and twelve hierarchical cluster dendrograms (2 for mincut/maxmer × 2 for with/without bias correction × 3 informational methods). In order to perform these informational calculations we used NOL equal to 1.
According to the phylogeny proposed by van der Linde et al. [28] the subfamily Drosophilinae (subgenus Sophophora and subgenus Drosophila) shows that D. paulistorum is close to D. willistoni and D. kikkawai is fairly close to D. melanogaster. The first two are further away from the last two, and this was the choice to present our data in the following sections. Therefore, we will focus on these four species, only to summarize the explanations.
Vertical entropy and mutual information
VH and VMI were computed for each of the two positions (i,j) as in [4] and also explained in methods. VH spectrum can be seen in Fig. 5 and VMI spectrum – a heat map  in Fig. 7. Both, VH spectra and VMI heat maps can be visually discriminated. In order to assess whether at least one distribution is statistically different from the others, ANOVA test was performed and resulted in pvalues near zero (for all maxmer/mincut versus with/without bias correction distributions). Therefore, there is at least one spectrum statistically different from all other spectra, and VH and VMI methods may be able to discriminate sets of molecular sequences.
Observing Vertical Entropy spectra (VH), in Fig. 5, the reader can visually discriminate the 4 frames with different profiles and frequency distributions. The data came from “maxmer” sequences with length equal to 859 bp, NOL = 1 and bias correction. Since the Adh locus is highly conserved (many residues with entropy equal to zero), the mean entropic value is very low and the standard error is large. On the left side of each frame we see the vertical red lines are SE(H(i)), where ‘i’ is the nucleotide position and SE(H) is the entropic standard error calculated from the polymorphism in this position over n sequences (species studies usually have different number of sequences). The horizontal red line stands for 2 standard deviations of the distribution and the black line for its mean. On the right side we see the frequency distribution graphic with 4 vertical lines: in black is the mean, in red is 1 standard deviation, in yellow is the median of the VH spectrum. Summarizing the four species: a) D. paulistorum has 12 sequences, mean(VH) = 37.7 (SD = 150.8) and median = 0 mnat; b) D. willistoni has 19 sequences, mean(VH) = 14.9 (SD = 70.8) and median = 0 mnat; c) D. kikkawai has 23 sequences, mean(VH) = 26.6 (SD =113.5) and median = 0 mnat; d) D. melanogaster has 30 sequences, mean(VH) = 96.7 (SD = 151.3) and median = 0 mnat.
Spectra of shuffled sequences simulated from original data can be seen in Fig. 6, having mean near 1 MER (maximum entropy), which produces a completely different spectrum when compared to the real data.
Observing Vertical Mutual Information (VMI) spectrum, in Fig. 7, the reader can visually discriminate four heat maps with different patterns, different locations of peaks and different maximum values (zoom the image). All data came from “maxmer” sequences with length equal to 859 bp, NOL = 1 and bias correction: Summarizing the four species: a) D. paulistorum has 12 sequences, max(VMI) = 734.8 (SE = 104.6) mnat at a discreet point in the heat map (112 × 138 bp); b) D. willistoni has 19 sequences, max(VMI) = 541.0 (SE = 115.2) mnat at a discreet point in the heat map (712 × 754 bp); c) D. kikkawai has 23 sequences, max(VMI) = 713.9 (SE = 51.6) mnat close to a region with a bumpy profile localized at 89 × 117 bp; and d) D. melanogaster has 30 sequences, max(VMI) = 835.1 (SE = 120.5) mnat in a highly bumpy profile with maximum value at 637 × 712 bp.
Spectra of shuffled sequences simulated from original data can be seen in Fig. 8, having mean = 136 mnat, a large SD = 106 mnat and median = 122 mnat, a completely different spectrum when compared to the real data. Here all mutual information is lost.
Horizontal mutual information
HMI was computed as in [1] and also explained in methods. All 17 Horizontal MI spectra cannot be visually discriminated. In order to assess whether at least one distribution is statistically different from the others, ANOVA test was performed with all spectra and resulted in pvalues near zero. Therefore, there is at least one spectrum statistically different from all other spectra, and HMI method may be able to discriminate sets of molecular sequences.
Observing HMI spectra, presented in Fig. 9, the reader is unable to visually discriminate them, and that’s why JSD (see next section) is a method to measure informational distances between spectra. All data came from “maxmer” sequences with length equal to 859 bp, NOL = 1 and bias correction. On the left side of each of 4 species frames, we see HMI versus k distance, also called HMI spectrum. Each value HMI(k) has its own mean and standard error obtained from the ensemble (all sequences from determined set). The mean of HMI(k) (horizontal black line) and its standard deviation (the red line is 2 SD) can be observed. On the right side we see the frequency distribution for HMI spectrum with 4 vertical lines: in black is the mean, in red is 1 standard deviation, and in yellow is the median. Summarizing the four species: a) D. paulistorum has 12 sequences, mean(HMI) = 7.4 (SD = 4.1) and median = 6.6 mnat; b) D. willistoni has 19 sequences, mean(HMI) = 7.9 (SD = 4.1) and median = 7.2 mnat; c) D. kikkawai has 23 sequences, mean(HMI) = 9.6 (SD = 4.5) and median = 9.0 mnat; d) D. melanogaster has 30 sequences, mean(HMI) = 9.3 (SD = 4.2) and median = 8.8 mnat. All these mean values are very low and we certified that they are greater than the superior value from the confidence interval from shuffled analyses.
Spectra of shuffled sequences simulated from original data can be seen in Fig. 10 having mean and median near zero mnat, a completely different spectrum compared to real data. Once again, all mutual information is lost. This is very important since all HMI have low mean, in a particular example, D. paulistorum has < HMI > = 7.2 mnat > > 0.063 (0.117) mnat for shuffled sequences. Therefore we can confirm that all spectra are statistically distinct compared to the spectrum of shuffled sequences. A nice discussion can be seen in [11, 29].
JSD
JSD applied to HMI spectra (JSD[HMI]) can be seen in Figs. 11 and 12, where informational distance histograms for maxmer and mincut, respectively, are displayed. Remarkable differences between distances are observed, many of them are statistically significant. Therefore, this method discriminates species with NOL = 1 and bias correction, since SE are negligible (see Additional file 1: HMI tab). However, phylogenetic studies are recommended to analyze whether closely related species are clusterized in a similar way.
A histogram calculated from spectra of shuffled sequences can be seen in Fig. 13, having very high mean (680 mnat) and low standard deviation (17 mnat), a completely different profile when compared to the real histogram. This histogram shows the loss of the capacity to discriminate species, since the informational distances are very similar as well as the standard errors.
Observing VMI, most of informational distances have small differences. Furthermore, for VMI and VH respective standard errors are very large resulting in a large confidence intervals. These results imply that JSD applied to both methods results in not statistically significant distance differences. Therefore, VH and VMI poorly discriminate species with NOL = 1 and bias correction (see Additional file 1: VMI tab and VH tab).
Hierarchical cluster
Hierarchical cluster analysis is the last algorithm and is computed based on a distance matrix calculated from JSD. In Fig. 14, we see four HMI dendrograms – here called Informational Dendrograms (ID)  obtained by applying weighted pair group with averaging method (WPGMA) for: a) maxmer with bias correction; b) maxmer without bias correction; c) mincut with bias correction; and d) mincut without bias correction. For JSD[VMI], applying bias correction, distances increase slightly and SE increases significantly (data not shown). But, for JSD[HMI], as seen in Figs. 11 and 12, standard errors are negligible and distances increase in few percentiles to 25 % when comparing mincut to maxmer. Those observed low SEs allow us to infer discrimination between species since distances are greater than zero with different values between pair of species. Changing from mincut to maxmer, with or without bias correction, some species change their positions in ID, but most of the clusters remain the same, like: “wilpaulsturtangor” (see acronym in the legend of Fig. 11), “triakikana”, “simmelayak”, “pbogopseudpersimira”, and the “americanas” or “amam_amam_tx”. Because many of these species can be discriminated, these clusters must be compared to phylogenetic trees.
Hierarchical cluster and phylogenetic gene trees
We also calculated phylogenetic trees for maxmer and mincut original sequences. We used Mega [14] for the methods Maximum Likelihood (ML) and Neighbor Joining (NJ). It was not our intention at all to compare phylogenetic trees versus mutual information dendrograms, but to observe whether ML and NJ can discriminate species for closely related species and whether the formed clusters are similar to the calculated informational dendrogram (ID) clusters.
In Fig. 15, we see the gene tree for maxmer sequences  a) ML × ID, and b) NJ × ID. ML was calculated using TN93 model, the resulted max log likelihood (LnL) was −14561 and Ts/Tv equal to 1.30. In Fig. 16, we see the gene tree for mincut sequences  a) ML × ID, and b) NJ × ID. ML was calculated using TN93 model, the resulted LnL was −8004 and Ts/Tv equal to 1.53. One reason for different LnL is different lengths for maxmer and mincut sequences.
For maxmer sequences (Fig. 15), ML and NJ could not discriminate the “americanas” species (D. americana, D. americana americana and D. americana texana). The species in group “wilpaulangorsturt” (D. willistoni, D. paulistorum, and D. sturtevanti) are fairly close and this also occurs in JSD[HMI]cluster method, except for D. Angor. This last group is close to the “americanas” cluster in NJ and ID. The species in group “simmelayak” (D. simulans, D. melanogaster and D. yakuba) are close in all three methods. The species in group “triakikana” (D. triaurium, D. kikkawai and D. ananassae) are close in all three methods with a similar topology. And finally, the species in group “pbogopseudpersimira” (D. pseudoobscurabogotana, D. pseudoobscura, D. persimilis, and D. miranda) are also close in all three methods.
Discussion
Mutual Information refers to common variation between residues/sites, here DNA sequences. It can be applied to horizontal direction (HMI) and vertical direction (VMI). JSD can be applied to pairs of mutual information spectra representing “mutual informational distances”. These distances are used to infer discrimination between species. However, JSD applied to VH refers to “information distance”, like a difference of potential as pointed by Adami [9]. These three informational methods generated well defined spectrum patterns, similar to molecular signatures.
For HMI, mincut sequences resulted in almost the same distance profile when compared to maxmer (Figs. 11 and 12). Comparing Fig. 11: a) “with bias correction” and b) “without bias correction”, we observe that the profile gets bumpier, which is a nice feature that allows better discrimination between species. The same occurs in Fig. 12a, b.
VH and VMI spectra can be visually discriminated, as molecular signatures (Figs. 5 and 7), but it was more difficult to visually compare all HMI spectra (Fig. 9). However JSD, with respective SE, allowed us to infer that HMI can discriminate species (Figs. 11 and 12), while with VH and VMI methods it was not possible.
Shuffled tests applied to all information spectra and distance histograms confirmed the all original results are statistically distinct than the shuffled one.
Afterwards we compared Phylogenetic Gene Trees to Informational Dendrograms calculated for HMI spectra. In Figs. 15 and 16, we observed that the clusters and topologies are in reasonable concordance. But, we certified that these Informational Dendrograms (JSD[HMI]clusterization) are not phylogenetic gene trees, they are only a mathematical way to cluster elements of the distance matrix.
Conclusions
MIA is a user friendly pipeline capable in retrieving, selecting, aligning and storing molecular sequences. It is also capable in calculating Shannon Vertical Entropy, Vertical and Horizontal Mutual Information and JSD between these informational spectra. MIA exports fasta files, calculated spectrum files and distance matrices in ASCII format. It displays VH, HMI and VMI spectra. VMI heat maps can be visualized in 2D and 3D (not shown here). MIA also displays informational distance histograms and informational dendrograms. It is designed to analyze possible species discrimination via any molecular sequences, but in this first version only DNA sequences were analyzed. More tests must be done in a near future like increasing NOL for the same data and also more deep analyses like: polymorphic genes (not highly conserved), sequences with larger lengths, and many simultaneous gene analyses.
Availability
MIA is freely available at https://github.com/flalix/MIA.
Abbreviations
 JSD:

JensenShannon divergence
 HC:

Hierarchical cluster
 HMI:

Horizontal mutual information
 HMIcluster:

Horizontal mutual information hierarchical cluster
 Maxmer:

Sequences with maximum gaps
 MI:

Mutual information
 MIA:

Mutual information analyzer
 Mincut:

Sequences with minimum gaps
 MSA:

Multiple sequence alignment
 NOL:

Number of letters or length of a word
 VH:

Vertical (Shannon) entropy
 VMI:

Vertical mutual information
References
 1.
Grosse I, Herzel H, Buldyrev SV, Stanley E. Species independence of mutual information in coding and noncoding DNA. Phys Rev E. 2000;61(5):5624–9.
 2.
Carpena P, BernaolaGalván P, Grosse I, Oliver JL, RománRoldán R, Stanley E. Finding Borders between Coding and Noncoding DNA Regions by an Entropic Segmentation Method. Phys Rev Lett. 1999;85(6):1342–5.
 3.
Bremm S, Schreck T, Boba P, Held S, Hamacher K. Computing and visually analyzing mutual information in molecular coevolution. BMC Bioinformatics. 2010;11(330):1–6.
 4.
Weil P, Hoffgaard F, Hamacher K. Estimating sufficient statistics in coevolutionary analysis by mutual information. Comput Biol Chem. 2009;33:440–4.
 5.
Kawashita SY, da Silva CV, Mortara RA, Burleigh BA, Briones MRS. Homology, paralogy and function of DGF1, a highly dispersed Trypanosoma cruzi specific gene family and its implications for information entropy of its encoded proteins. Mol Biochem Parasitol. 2008;165(1):19–31.
 6.
Ferreira RC, Briones MRS. Phylogenetic evidence based on Trypanosoma cruzi nuclear gene sequences and information entropy suggest that interstrain intragenic recombination is a basic mechanism underlying the allele diversity of hybrid strains. Infect Genet Evol. 2012;12(5):1064–71.
 7.
Python Community. Python Programming Language [Internet]. 2012 [cited 2012 Jan 29]. Available from: http://www.python.org/
 8.
Biopython community. Biopython [Internet]. 2012 [cited 2012 Jan 25]. Available from: http://biopython.org/wiki/Biopython
 9.
Adami C. Information theory in molecular biology. Phys Life Rev. 2004;1:3–22.
 10.
Cover TM, Thomas JA. Elements of Information Theory. second edition. John Wiley & Sons, Inc., Hoboken, New Jersey; 2006.
 11.
Steuer R, Kurths J, Daub CO, Weise J, Selbig J. The mutual information: detecting and evaluating dependencies between variables. Bioinformatics. 2002;18 Suppl 2:S231–40.
 12.
Gouy M, Guindon S, Gascuel O. SeaView Version 4: A Multiplatform Graphical User Interface for Sequence Alignment and Phylogenetic Tree Building. Mol Biololgy Evol. 2002;27(2):221–4.
 13.
Kearse M, Moir R, Wilson A, StonesHavas S, Cheung M, Sturrock S, et al. Geneious Basic: an integrated and extendable desktop software platform for the organization and analysis of sequence data. Bioinformatics. 2012;28(12):1647–9.
 14.
Tamura K, Stecher G, Peterson D, Filipski A, Kumar S. MEGA6: Molecular Evolutionary Genetics Analysis Version 6.0. Mol Biol Evol. 2013;30:2725–9.
 15.
Hall TA. BioEdit: a userfriendly biological sequence alignment editor and analysis program for Windows 95/98/NT. Nucl Acids Symp Ser. 1999;41:95–8.
 16.
Martin LC, Gloor GB, Dunn D, Wahl LM. Using information theory to search for coevolving residues in proteins. Bioinformatics. 2005;21(22):4116–24.
 17.
GouveiaOliveira R, Roque FS, Wernersson R, SicheritzPonten T, Sackett PW, Mølgaard A, et al. InterMap3D: predicting and visualizing coevolving protein residues. Bioinforma  Appl Notes. 2009;25(16):1963–5.
 18.
Roulston MS. Estimating the errors on measured entropy and mutual information. Phys D. 1999;125:285–94.
 19.
Edgar RC. MUSCLE: multiple sequence alignment with high accuracy and high throughput. Nucleic Acids Res. 2004;32(5):1792–7.
 20.
DPDB Home Page: Drosophila DNA polymorphism [Internet]. [cited 2012 Nov 4]. Available from: http://bioinf3.uab.cat/DPDB/dpdb.asp
 21.
Santos G, Schroeder AJ, Gelbart W, FlyBase Consortium. FlyBase: introduction of the Drosophila melanogaster Release 6 reference genome assembly and largescale migration of genome annotations. Nucl Acids Res. 2014;1:1–8.
 22.
Berkeley. Berkeley Drosophila Genome Project [Internet]. Available from: http://insitu.fruitfly.org/cgibin/ex/insitu.pl
 23.
Herzel H, Grosse I. Measuring correlations in symbol sequences. Phys A. 1995;216:518–42.
 24.
Herzel H, Grosse I. Correlations in DNA sequences: The role of protein coding segments. Phys Rev E. 1997;55(1):800–10.
 25.
Shannon CE. A Mathematical Theory of Communication. Bell Syst Tech J. 1948;27:379–423. 623–56.
 26.
Endres DM, Schindelin JE. A new metric for probability distributions. IEEE Trans Inf Theory. 2003;49(7):1858–60.
 27.
Johnson AD. An extended IUPAC nomenclature code for polymorphic nucleic acids. Genet Popul Anal. 2010;26(10):1386–9.
 28.
van der Linde K, Houle D, Spicer GS, Steppan SJ. A supermatrix based on molecular phylogeny of the family Drosophilidae. Genet Res. 2010;92(01):25–38.
 29.
Grosse I, BernaolaGalván P, Carpena P, RománRoldán R, Oliver J, Stanley HE. Analysis of symbolic sequences using the JensenShannon divergence. Phys Rev E. 2002;65(041905):1–16.
Acknowledgements
The authors wish to thank Francisco Bosco, who passed away in 2012, for the support and initial discussions, and also to Prof. Geoff Barton and Alexandre Ramos who read a previous version of the manuscript and helped to improve it. This work was supported by CAPES PhD fellowship to FL, a FAPESP (2013/078380) grant to MRSB and CNPq (Brazil) productivity fellowships to FA and MRSB.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
FL concept, design, programming, manuscript draft, documentation and testing. MB and FA concept, tests and discussion. All authors were involved in writing the manuscript. All authors read and approved the final manuscript and manuals.
Additional file
Additional file 1:
DrosophilaAdh data. (XLSX 153 kb)
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Lichtenstein, F., Antoneli, F. & Briones, M.R.S. MIA: Mutual Information Analyzer, a graphic user interface program that calculates entropy, vertical and horizontal mutual information of molecular sequence sets. BMC Bioinformatics 16, 409 (2015). https://doi.org/10.1186/s1285901508370
Received:
Accepted:
Published:
Keywords
 Software
 Information theory
 Entropy
 Mutual information
 DNA sequences
 Species