 Research
 Open Access
 Published:
Comparative analysis and prediction of nucleosome positioning using integrative feature representation and machine learning algorithms
BMC Bioinformatics volumeÂ 22, ArticleÂ number:Â 129 (2021)
Abstract
Background
Nucleosome plays an important role in the process of genome expression, DNA replication, DNA repair and transcription. Therefore, the research of nucleosome positioning has invariably received extensive attention. Considering the diversity of DNA sequence representation methods, we tried to integrate multiple features to analyze its effect in the process of nucleosome positioning analysis. This process can also deepen our understanding of the theoretical analysis of nucleosome positioning.
Results
Here, we not only used frequency chaos game representation (FCGR) to construct DNA sequence features, but also integrated it with other features and adopted the principal component analysis (PCA) algorithm. Simultaneously, support vector machine (SVM), extreme learning machine (ELM), extreme gradient boosting (XGBoost), multilayer perceptron (MLP) and convolutional neural networks (CNN) are used as predictors for nucleosome positioning prediction analysis, respectively. The integrated feature vector prediction quality is significantly superior to a single feature. After using principal component analysis (PCA) to reduce the feature dimension, the prediction quality of H. sapiens dataset has been significantly improved.
Conclusions
Comparative analysis and prediction on H. sapiens, C. elegans, D. melanogaster and S. cerevisiae datasets, demonstrate that the application of FCGR to nucleosome positioning is feasible, and we also found that integrative feature representation would be better.
Background
The nucleosome is the basic structural unit of eukaryotic chromatin. It is formed by the combination of histones and DNA. The core is an octamer formed by two copies of each histones H2A, H2B, H3 and H4, DNA is wound around it about 1.65 turns. Among them, the DNA wrapped around the octamer is called core DNA, which is 147 base pairs in length; the DNA sequence that connects two adjacent nucleosomes is called linker DNA, which ranges from 20 to 60 base pairs [1]. In eukaryotic cells, nucleosomes play a crucial role in the process of genome expression, DNA replication, DNA repair and transcription [2,3,4,5,6]. In addition, studies have demonstrated that abnormal histone modifications in the nucleosome structure are directly related to diseases such as tumors [7] and lupus erythematosus [8]. Therefore, the mechanism of nucleosome positioning in DNA sequence has an extremely important research value, which is also one of the hot spots in current epigenetics research.
The precise position of the nucleosome on the DNA sequence in the whole genome is called nucleosome positioning. Early experiments mainly used micrococcal nuclease to process chromatin to achieve nucleosome positioning [9]. In recent years, benefiting from the development and application of highthroughput experimental techniques, such as chromatin immunoprecipitationchip (ChIPchip), chromatin immunoprecipitation sequencing (ChIPSeq), many breakthroughs have been made in nucleosome positioning experiments. The nucleosome positioning maps of different species such as Saccharomyces cerevisiae [10, 11], Homo sapiens [12], Caenorhabditis elegans [13], Drosophila melanogaster [14], etc. have been obtained, which provides a large amount of data basis for researchers to carry out theoretical research and prediction.
Much of the research in nucleosome positioning is based on DNA sequence analysis [15, 16]. The DNA sequence consists of four nucleotides: A, T, C and G. Studies have shown that the affinity between genomic DNA sequences and histones is clearly dependent on sequence order, which indicates that the DNA sequence order does affect the position of nucleosome formation. Although some provide the support that nucleosome positioning is affected by multiple factors such as DNA sequence, ATPdependent nucleosome remodeling enzymes and transcription factors [17, 18]. Many researchers used sequence analysis methods to express nucleosome DNA sequence characteristics and then performed nucleosome positioning and recognition.
In the past decade, with the popularity of machine learning algorithms, a multitude of computational models based on DNA sequence information have been proposed. Chen et al. proposed the "iNucPhyschem" nucleosome prediction model using 12 physicochemical features of DNA, which identified the core DNA and linker DNA of the yeast genome nucleosome [19]. Later, the research group also established a biophysical model based on the deformation energy of DNA sequences to predict the sequence of nucleosomes [20]. Guo et al. used pseudo ktuple nucleotide composition to successfully express the feature vector of the DNA sequence, and used the support vector machine (SVM) classifier to train H. sapiens, C. elegans and D. melanogaster [21]. 3LS model used similar methods and combined the distribution of different numbers of nucleotide combinations in the sequence to further improve the prediction accuracy [22]. ZCMM model based on the Zcurve (zcurve) theory and the position weight matrix (PWM), the prediction performance is excellent on D. melanogaster [23].
Deep learning is also applied to nucleosome positioning and achieved good prediction quality. These deep learning models all used onehot encoding. Gangi et al. [24] constructed a deep learning model that integrates convolutional layers and long shortterm memory networks. LeNup model added the Inception module and gated convolutional network to the convolutional neural network to improve the nucleosome positioning [25].
In this work, we firstly will use frequency chaos game representation to construct DNA sequence features. This feature representation method has not been used in nucleosome positioning before. Secondly, we also integrated FCGR with other feature vectors and adopted the principal component analysis (PCA) algorithm to achieve the feature dimensionality reduction. Finally, various machine learning algorithms such as support vector machine (SVM), extreme learning machine (ELM), extreme gradient boosting (XGBoost), multilayer perceptron (MLP), and convolutional neural networks (CNN) will be used to perform comparative analysis and prediction of nucleosome positioning.
Results
Rule of performance evaluation
Cross validation is a statistical analysis method used to validate the model. The basic idea is to divide the original data into a training set and a test set. First, use the training set to train the model, and then use the test set to test the classification or prediction performance of the obtained model. In this work, we used Kfold crossvalidation to evaluate the performance of the predictor through four parameters: sensitivity (\(S_{n}\)), specificity (\(S_{p}\)), accuracy (ACC), and Mathew's correlation coefficient (MCC). The specific definition are as follows:
where TP, TN, FP and FN are the numbers of true positives, true negatives, false positives and false negatives, respectively [25]. \(S_{n}\) is the true positive rate. When \(S_{n}\)â€‰=â€‰1, it means that all core DNA of nucleosomes have been correctly predicted.\({ }S_{p}\) is true negative rate. When \(S_{p}\)â€‰=â€‰1, it means that all linker DNAs are correctly predicted. ACC reflects the ratio of the number of correctly predicted samples of each category to the total sample. MCC comprehensively evaluates the prediction results. MCCâ€‰âˆˆâ€‰[âˆ’â€‰1,1]. MCCâ€‰=â€‰â€‰âˆ’â€‰1 means that the correlation is completely opposite. MCCâ€‰=â€‰1 means that the prediction result is completely correlated with the true category. MCCâ€‰=â€‰0 means that the prediction is completely random.
Receiver operating characteristic curve (ROC curve) and area under curve (AUC) are often used to evaluate the pros and cons of a binary classifier. Area under curve (AUC) is the area under the Roc curve, usually between 0.5 and 1. As a value, AUC can be used to evaluate the quality of the classifier more intuitively. The larger the AUC value, the better. Taking into account the length of the paper, this paper only calculates the AUC value and does not draw the ROC curve one by one.
Performance of predictors
According to the characteristics of FCGR described above, the different values of K nucleotide will affect the feature expression of the DNA sequence [26]. A large K value means a high feature dimension. And generally, highdimensional features are relatively sparse, and the fitting quality may not be outstanding. Obviously, choosing an appropriate K value will have a greater impact on the classification effect of each classifier. Some studies have combined DNA sequence features [22, 23, 27, 28]. Similarly, FCGR can also use different combinations of K nucleotide values as feature vectors.
Feasibility of FCGR
In this work, we flatten the FCGR matrix into a normalized vector (1D) corresponding to the frequency of K nucleotides as the input of SVM and ELM [27]. The input of MLP and CNN models are not only singlechannel FCGR images (2D) [26, 27], but also multiple Kvalue images, the image size is 64â€‰Ã—â€‰64. For the input of multiKvalue images, we leveraged multiple channels to feed in the combination of K values when training the model, and used simple averaging to calculate the final prediction result. To find the appropriate value of K or combination, we use 10fold crossvalidation. Figure 1 shows the classification accuracy of each classifier with different K values and combinations.
For SVM, the accuracy of H. sapiens, C. elegans reaches its peak with Kâ€‰=â€‰1, 2 and 4; the accuracy of D. melanogaster was the highest with Kâ€‰=â€‰2 and 4. For ELM, the accuracy of D. melanogaster reaches an peak when Kâ€‰=â€‰2; the accuracy of H. sapiens reaches its peak when Kâ€‰=â€‰2 and 4; the classification accuracy of C. elegans is best with Kâ€‰=â€‰1, 2 and 4 like using SVM.
For MLP, the accuracy of H. sapiens and D. melanogaster reaches its peak with Kâ€‰=â€‰3, 4 and 5; the classification accuracy of C. elegans is best with Kâ€‰=â€‰3 and 4. For CNN, H. sapiens have the best classification quality when using the FCGR image with Kâ€‰=â€‰4; the accuracy of C. elegans reaches its peak with Kâ€‰=â€‰4 and 5; the accuracy of D. melanogaster reaches its peak with Kâ€‰=â€‰3, 4 and 5. Table 1 clearly shows the best prediction results for four species via 10fold crossvalidation.
For S. cerevisiae dataset, we used SVM, ELM and MLP to achieve \(S_{n}\)â€‰=â€‰\(S_{p}\)â€‰=â€‰ACCâ€‰=â€‰MCCâ€‰=â€‰AUCâ€‰=â€‰1 via 10fold crossvalidation when Kâ€‰=â€‰3 or 4. There may be room for improvement in the predicted quality of the other three datasets.
Comparison of the results with integrative features
In addition, we also integrated FCGR with other feature representations [29,30,31,32], such as DAC, TAC, DACC, TACC, PCPseDNC, PCPseTNC, and input them into SVM and ELM. Besides, we added the extreme gradient boosting (XGBoost) algorithm. The comparative analysis results are shown in Tables 2, 3 and 4 respectively.
From the results in Tables 2, 3 and 4, the combination of FCGR and DAC as feature vectors have a greater prediction quality. XGBoost performance is relatively stable, and each prediction results have little difference, especially for inputting highdimensional features. However, after some highdimensional feature vectors are input into SVM and ELM, the prediction results are relatively poor. It shows that XGBoost is more suitable for processing highdimensional features.
Comparison of the results with dimensionality reduction
Considering the high dimensionality of the integrative feature vector, it is possible that highdimensional feature vectors would bring the curse of dimensionality, which leads to overfitting of the prediction result. Therefore, we also adopted the principal component analysis (PCA) algorithm [33] to achieve feature dimensionality reduction. Then, the feature vector after dimensionality reduction is input into SVM, ELM and XGBoost respectively. In the process of using PCA to dimensionality reduction, the cumulative contribution rate of the retained principal components will directly affect the dimensionality reduction effect. Therefore, we calculated the accuracy of 95%, 93%, 90%, 88% and 85% of the contribution rate of the retained principal components respectively. Figures 2, 3 and 4 shows the classification accuracy of each classifier with different contributing rate of principal component. And the results of the optimal contribution rate of the principal components corresponding to each predictor are shown in Tables 5, 6 and 7 respectively.
From Tables 5, 6 and 7, We have noticed that the prediction quality has been improved after dimensionality reduction through PCA for H. sapiens. It is increased by 4.57%, 3.12%, 6.00%, 9.03%, 3.56% in ACC, \(S_{n}\), \(S_{p}\), MCC and AUC when we combined FCGR vectors and TAC for using with XGBoost. However, the prediction quality has been not improved significantly for C. elegans. Especially when ELM was used, its prediction quality decreased slightly. For D. melanogaster, similarly, there is no significant improvement.
Comparison with other algorithms
To verify the effectiveness of our method, we compared the prediction results of the optimal performing predictors in Tables 1, 2, 3 and 4 with other models using the same datasets. DLNN5 [24] is a deep learning model with a convolution kernel size of 5, and ZCMM [23] is based on SVM. Tables 8, 9, 10 and 11 shows that our methods perform prominently on H. sapiens and S. cerevisiae datasets. For S. cerevisiae dataset, we used SVM, ELM and MLP to achieve \(S_{n} = S_{p}\)â€‰=â€‰ACCâ€‰=â€‰MCCâ€‰=â€‰AUCâ€‰=â€‰1 via 10fold crossvalidation when Kâ€‰=â€‰3 or 4. Compared with the model that based on DNA deformation energy in the original paper [20], the prediction performance has been obviously lifted. For H. sapiens, combined FCGR vectors and TAC for using with XGBoost is higher than ZCMM in ACC, \(S_{n}\), \(S_{p}\), MCC, AUC by 10.87%, 15.58%, 5.23%, 21.25%, 8.81%, respectively; likewise, it is higher than DLNN5 in ACC, \(S_{n}\), \(S_{p}\) by 3.22%, 2.11%, 4.45%, respectively. The performance of CNN is slightly better than ZCMM and DLNN5. For C. elegans, compared with ZCMM, we use ELM to increase the evaluation indicators by 2.20%, 10.64%, 1.56%, 13.15%, 3.01% when combined FCGR vectors with Kâ€‰=â€‰1, 2 and 4. For D. melanogaster, our prediction accuracy is lower, and ZCMM's prediction accuracy (ACC) is the highest at 93.62%. Results imply that our final prediction is positive, it only performed unfavorably on the D. melanogaster dataset.
Comparison with other advanced methods
In addition to DLNN5 and ZCMM models, there are some other advanced methods for nucleosome prediction in the same dataset. LeNup model utilizes improved convolutional neural networks, which adds inception modules and gated convolutional networks [25]. 3LS is based on the linear regression model [22]. LeNup used the 20fold crossvalidation and provided comparison data with 3LS for H. sapiens, C. elegans and D. melanogaster. Therefore, we utilized the results provided by LeNup for comparative analysis in Tables 12, 13 and 14.
LeNup has the best overall prediction effect. The accuracy of C. elegans is 0.9188, and the average accuracy of other species are also over 0.88. The prediction result of our method is relatively close to it on the H. sapiens dataset. For C. elegans, ELM with FCGR performs slightly worse than 3LS, ACC, \(S_{p}\), MCC, AUC decreased by 0.29%, 3.44%, 0.51%, 1.86% respectively.
Discussion
Firstly, the results in Table 1 and Fig. 1 clearly showed that the FCGR feature of the combined K value is better than the single K value, and the SVM output better prediction results. When training CNN and MLP models, we utilized multichannel multiple Kvalue input images, and the prediction accuracy had been improved. All these indicated that FCGR feature combinations with different K values can better express sequence features, thereby improving models' prediction accuracy.
Secondly, we further integrated FCGR with other feature representations, and combined three types of machine learning algorithms to compare prediction results (Tables 2, 3, 4). Besides, we performed PCA dimensionality reduction processing on feature vectors to prevent highdimensional features from causing overfitting (Tables 5, 6, 7). Although the overall prediction quality has improved after the PCA dimensionality reduction processing with the integrated feature, superior results are obtained for using FCGR feature representation. These also further illustrated the advantages of FCGR features representation.
Here we compared the results of the proposed method with other advanced algorithms. Slightly superior results are achieved with our algorithm on H. sapiens and S. cerevisiae datasets, but there are gaps in the other two datasets. On the one hand, it explains the feasibility of our method; on the other, our work has room for improvement.
Conclusions
In this work, we used FCGR to represent the features of the DNA sequence and applied it to the nucleosome positioning. Our experiments have achieved positive results. Especially when multiple features are used in combination, the prediction quality can be improved. The advantage of this representation is that the time consumed in the process of constructing features is shortened, and the features are clear and intuitive. The quality of integrating features representation is also acceptable. Particularly after we use PCA for dimensionality reduction, the prediction quality of H. sapiens dataset has been improved. This demonstrates the feasibility of the method.
In this paper, we also tried a simple CNN model with FCGR image and got mediocre results. Since deep learning is now increasingly used in bioinformatics. In the further research of nucleosome positioning, we will try to build a more efficient deep learning prediction model to achieve prediction of DNA represented in the form of images, such as FCGR image.
Methods
Dataset descriptions
To compare the results of the predictors, the datasets of this work downloaded from two published papers [20, 21]. The first group of datasets involved H. sapiens, C. elegans and D. melanogaster from the paper by Guo et al. [21]. The length of each DNA sequence is 147 bp. The second dataset involved S. cerevisiae genome from the paper by Chen et al. [20]. The length of each DNA sequence is 150 bp. Both of the datasets contain two types of samples: nucleosomeforming sequences (positive data) and nucleosomeinhibiting sequences (negative data). And none of the sequences included hasâ€‰â‰¥â€‰80% pairwise sequence identity with any other. The details of the datasets are shown in Table 15.
DNA sequence feature representation
Except for the above mentioned, common DNA sequence representation methods include basic kmer (Kmer) [34], reverse complementary kmer (RevKmer) [35], etc. based on deoxyribonucleic acid composition, and some are based on the correlation between nucleotide physical and chemical indicators, such as dinucleotidebased autocovariance (DAC), trinucleotidebased autocovariance (TAC) [29], etc. and pseudo ktuple nucleotide composition (PseKNC) [21] based on pseudo deoxyribonucleic acid composition. These feature representation methods have specific calculation formulas and iterative functions, and some calculations are more complex and require a long time. This paper will mainly use a simple and intuitive feature representation.
Chaos game representation (CGR) is a graphical representation method of gene sequence based on chaos theory proposed by Jeffrey in 1990 [36]. The method is as follows: The four nucleotides {A, T, G, C} are located at the four vertices of the plane coordinate system, and the position of each nucleotide in the DNA sequence in the plane is \(P_{i}\). According to formula (2) draw the coordinate point of each nucleotide:
Among them, \(P_{0}\) is the given starting point, L is the length of the DNA sequence, and \(N_{i}\) represents the corresponding coordinate of the ith nucleotide, where Aâ€‰=â€‰(0,0), Tâ€‰=â€‰(1,0), Gâ€‰=â€‰(1,1), Câ€‰=â€‰(0,1). This method draws a corresponding image of a DNA sequence through the iterative function and makes the nucleotides in the sequence correspond to the points on the image one by one [36,37,38,39,40]. From Fig. 5, we can see the CGR graphical representation of the two types of sample sequences in the H. sapiens dataset.
Divide the CGR image into \(2^{K} \times 2^{K}\) subblocks and calculate the number of points appearing on each subblock, we can determine the frequency of K nucleotide combinations, and then convert the CGR image into a \(2^{K} \times 2^{K} { }\) matrix, which is called frequency chaos game representation (FCGR) [39]. For example, we divided the CGR graph of Fig. 5a into a \(2^{3} \times 2^{3}\) matrix and calculated the number of occurrences of the midpoint of each subblock, and obtain the frequency matrix shown in Table 16.
FCGR can be used not only as a numerical matrix, but also as a grayscale image. The original CGR image is divided into \(4^{K}\) subblocks. The darker the subblock, the more dots appear in the subblocks; the lighter subblocks, indicates that the number of dots in the color block is small, and the pixel value of the image is between 0 and 255 [39]. From Fig. 6, we can see the FCGR image of the sample sequence with Kâ€‰=â€‰3, 4 and 5, respectively.
Support vector machine
Support vector machine (SVM) is a commonly used twoclass classification model. Compared with other classification algorithms, it has a good classification effect and strong generalization ability on small data sets. It can also handle nonlinear classification problems through nuclear techniques. Thus, support vector machines have also been widely used in the field of bioinformatics [19, 21, 23]. Its basic idea is to map the sample from the original lowdimensional space to a highdimensional space, so that the sample can find a partitioning hyperplane with the largest interval in the feature space, and separate samples of different categories.
In this paper, we will use the python package (Scikitlearn 0.23), which can be downloaded from https://scikitlearn.org/stable/index.html. This package contains the SVM module, and the implementation is based on libsvm. We will train the SVM with the radial basis function (RBF) kernel, meanwhile two parameters will be considered: penalty parameter C and kernel coefficient Gamma. In the training process, we used the grid optimization method to determine the best values of the two parameters.
Extreme learning machine
Extreme learning machine (ELM) was proposed by GuangBin Huang. The algorithm is a new machine learning algorithm based on single hidden layer feedforward neural networks (SLFNs). Compared with traditional algorithms, ELM has a faster learning speed while maintaining learning accuracy. The core idea is to randomly select the input layer weight and hidden layer bias of the network, and get the corresponding hidden node output [41]. The network structure of ELM model is shown in Fig. 7.
The experiment reference used David Lambert's Python version of ELM resources, which can be downloaded from the ELM web portal (https://www.ntu.edu.sg/home/egbhuang/). The code can be found on https://github.com/dclambert/PythonELM.
Extreme gradient boosting
Extreme gradient boosting (XGBoost) is an open source machine learning project developed by Tianqi Chen et al. [42]. It is one of the boosting algorithms, which has the characteristics of high efficiency, flexibility, high accuracy, and strong portability. It is applied in the field of biomedicine [43].
The idea of XGBoost algorithm is to continuously add trees and perform feature splitting to complete the construction of a tree. In the whole process, each addition of a tree is learning a new function to fit the residual of the previous prediction. When the training is completed, K trees will be obtained. If we want to predict the score of a sample, according to the features of this sample, each tree will fall to a corresponding leaf node, and each leaf node corresponds to a score. Finally, we only need to add up the scores corresponding to each tree to get the predicted value of the sample.
In this experiment, we used the python package (xgboost 1.2.0), which can be downloaded from https://github.com/dmlc/xgboost.
Multilayer perceptron
Multilayer perceptron (MLP) is also called deep neural networks (DNNs) [44]. MLP is based on the extension of perception. Multiple hidden layers are introduced between the input layer and the output layer, and the neurons between the layers are fully connected. So, both the hidden layer and the output layer in MLP are fully connected layers.
For the MLP, we used the AI Studio (https://aistudio.baidu.com/aistudio/index) experimental platform and PaddlePaddle (https://www.paddlepaddle.org.cn/) deep learning framework provided by Baidu (https://www.baidu.com/) to implement the experimental model with python (https://www.python.org/). MLP has three hidden layers with Relu activation function [45], each layer contains 50 neurons, the output layer uses a softmax activation function. Besides, MLP is trained by 5 epchos, with Adamax optimizer a learning rate of 0.001. Adamax algorithm is a variant of Adam algorithm based on infinite norm, which makes the algorithm of learning rate update more stable and simple [46]. We use cross entropy as our loss function.
Convolutional neural network
Convolutional Neural Network (CNN) is a representative algorithm of deep learning. It has demonstrated extraordinary advantages in the field of computer vision and has also been widely used in bioinformatics [47, 48]. Convolutional neural networks can automatically extract features from input data. Compared with fully connected neural networks, it can simplify model complexity and effectively reduce model parameters [49]. Convolutional neural networks are applied to the general framework of image mode, mainly composed of convolutional layers, activation function, pooling layers and fully connected layers [49, 50].
Owing to the limitation of the sample data volume, during the training process, we need to prevent the overfitting problem faced by CNN, so we add a batch normalization (BN) layer [51] after the convolutional layer and add a dropout layer [52] after the fully connected layer. In our network, the convolutional layer uses a 3â€‰Ã—â€‰3 convolution kernel, the number of filters in the first layer is 64, and the second is 32. The pooling layer use the maximum pooling of 2â€‰Ã—â€‰2, with strideâ€‰=â€‰2. The first fully connected layer neurons' number is 100, and the second is 50. Then, the dropout probability of the subsequent dropout layer is 0.5. Except the softmax activation function used in the output layer, the activation function in the other layers is Relu. CNN is training by 20 epchos, with Adamax optimizer a learning rate of 0.001. The loss function is cross entropy. Like MLP, we also used the AI Studio experimental platform and PaddlePaddle deep learning framework provided by Baidu to implement the experimental model in python. The specific network structure is shown in Fig. 8.
Availability of data and materials
The datasets of this work can be downloaded from two published papers [20, 21]. The python source code used in this work are freely available at https://github.com/lliqiecho/Comparativeanalysisandpredictionofnucleosomepositioning.
Abbreviations
 ACC:

Accuracy
 AUC:

Area under curve
 CGR:

Chaos game representation
 CNN:

Convolutional neural networks
 DAC:

Dinucleotidebased autocovariance
 DACC:

Dinucleotidebased autocross covariance
 ELM:

Extreme learning machine
 FCGR:

Frequency chaos game representation
 MCC:

Mathew's correlation coefficient
 MLP:

Multilayer perceptron
 PCA:

Principal component analysis
 PCPseDNC:

Parallel correlation pseudo dinucleotide composition
 PCPseTNC:

Parallel correlation pseudo trinucleotide composition
 \({\text{S}}_{{\text{n}}}\) :

Sensitivity
 \({\text{S}}_{{\text{p}}}\) :

Specificity
 SVM:

Support vector machine
 TAC:

Trinucleotidebased autocovariance
 TACC:

Trinucleotidebased autocross covariance
 XGBoost:

Extreme gradient boosting
References
Luger K, MÃ¤der AW, Richmond RK, et al. Crystal structure of the nucleosome core particle at 2.8 A resolution. Nature. 1997;389:251â€“60.
Nocetti N, Whitehouse I. Nucleosome repositioning underlies dynamic gene expression. Genes Dev. 2016;30(6):660â€“72.
Bai L, Morozov AV. Gene regulation by nucleosome positioning. Trends Genet. 2010;26(11):476â€“83.
Schalch T, Duda S, Sargent DF, et al. Xray structure of a tetranucleosome and its implications for the chromatin fibre. Nature. 2005;436:138â€“41.
Eaton ML, Kyriaki G, Sukhyun K, et al. Conserved nucleosome positioning defines replication origins. Genes Dev. 2010;24(8):748â€“53.
Rodriguez J, Lee L, Lynch B, et al. Nucleosome occupancy as a novel chromatin parameter for replication origin functions. Genome Res. 2017;27(2):269â€“77.
Desvoyes B, Sanchez MP, RamirezParra E, et al. Impact of nucleosome dynamics and histone modifications on cell proliferation during Arabidopsis development. Heredity. 2010;105(1):80â€“91.
Bruns A, BlÃ¤ss S, Hausdorf G, et al. Nucleosomes are major T and B cell autoantigens in systemic lupus erythematosus. Arthritis Rheum Off J Am CollRheumatol. 2000;43(10):2307â€“15.
Kornberg RD. Structure of chromatin. Annu Rev Biochem. 1977;46:931â€“54.
Yuan GC, Liu YJ, Dion MF, et al. Genomescale identification of nucleosome positions in Scerevisiae. Science. 2005;309(5734):626â€“30.
Lee W, Tillo D, Bray N, et al. A high resolution atlas of nucleosome occupancy in yeast. Nat Genet. 2007;39(10):1235â€“44.
Ozsolak F, Song JS, Liu XS, et al. Highthroughput mapping of the chromatin structure of human promoters. Nat Biotechnol. 2007;25(2):244â€“8.
Valouev A, Ichikawa J, Tonthat T, et al. A highresolution, nucleosome position map of C. elegans reveals lack of universal sequencedictated positioning. Genome Res. 2008;18(7):1051â€“63.
Mavrich TN, Jiang C, Ioshikhes IP, et al. Nucleosome organization in the Drosophila genome. Nature. 2008;453(7193):358â€“62.
Xi L, FondufeMittendorf Y, Xia L, et al. Predicting nucleosome positioning using a duration Hidden Markov Model. BMC Bioinformatics. 2010;11:346.
van der Heijden T, van Vugt JJ, Logie C, et al. Sequencebased prediction of single nucleosome positioning and genomewide nucleosome occupancy. Proc Natl Acad Sci U S A. 2012;109(38):E2514â€“22.
Segal E, FondufeMittendorf Y, Chen L, et al. A genomic code for nucleosome positioning. Nature. 2006;442:772â€“8.
Struhl K, Segal E. Determinants of nucleosome positioning. Nat Struct Mol Biol. 2013;20(3):267â€“73.
Chen W, Lin H, Feng PM, et al. iNucPhysChem: a sequencebased predictor for identifying nucleosomes via physicochemical properties. PLoS ONE. 2012;7(10):e47843.
Chen W, Feng PM, Ding H, et al. Using deformation energy to analyze nucleosome positioning in genomes. Genomics. 2016;107(2â€“3):69â€“75.
Guo SH, Deng EZ, Xu LQ, et al. iNucPseKNC: a sequencebased predictor for predicting nucleosome positioning in genomes with pseudo ktuple nucleotide composition. Bioinformatics. 2014;30(11):1522â€“9.
Awazu A. Prediction of nucleosome positioning by the incorporation of frequencies and distributions of three different nucleotide segment lengths into a general pseudo ktuple nucleotide composition. Bioinformatics. 2017;33(1):42â€“8.
Cui Y, Xu Z, Li J. ZCMM: a novel method using Zcurve theory based and position weight matrix for predicting nucleosome positioning. Genes (Basel). 2019;10(10):765.
Di Gangi M, Lo Bosco G, Rizzo R. Deep learning architectures for prediction of nucleosome positioning from sequences data. BMC Bioinformatics. 2018;19(Suppl 14):418.
Zhang J, Peng W, Wang L. LeNup: learning nucleosome positioning from DNA sequences with improved convolutional neural networks. Bioinformatics. 2018;34(10):1705â€“12.
Rizzo R, Fiannaca A, La Rosa M, et al. Classification experiments of DNA sequences by using a deep neural network and chaos game representation. In: Proceedings of the 17th international conference on computer systems and technologies 2016. 2016. p. 222â€“8.
Touati R, Messaoudi I, Oueslati AE, et al. New intraclasshelitrons classification using DNAImage sequences and machine learning approaches. IRBM. 2020. https://doi.org/10.1016/j.irbm.2019.12.004.
Lu M, Liu S, Sangaiah AK, et al. Nucleosome positioning with fractal entropy increment of diversity in telemedicine. IEEE Access. 2018;6:33451â€“9.
Dong Q, Zhou S, Guan J. A new taxonomybased protein fold recognition approach based on autocrosscovariance transformation. Bioinformatics. 2009;25:2655â€“62.
Chen W, Zhang X, Brooker J, et al. PseKNCGeneral: a crossplatform package for generating various modes of pseudo nucleotide compositions. Bioinformatics. 2015;31(1):119â€“20.
Teif VB. Nucleosome positioning: resources and tools online. Brief Bioinform. 2016;17(5):745â€“57.
Liu B, Zhang D, Xu R, et al. Combining evolutionary information extracted from frequency profiles with sequencebased kernels for protein remote homology detection. Bioinformatics. 2014;30(4):472â€“9.
Lever J, Krzywinski M, Altman N. Principal component analysis. Nat Methods. 2017;14:641â€“2.
Liu B, Fang L, Wang S, et al. Identification of microRNA precursor with the degenerate Ktuple or Kmer strategy. J Theor Biol. 2015;385:153â€“9.
Noble WS, Kuehn S, Thurman R, et al. Predicting the in vivo signature of human gene regulatory sequences. Bioinformatics. 2005;21(suppl_1):i338â€“343.
Jeffrey HJ. Chaos game representation of gene structure. Nucleic Acids Res. 1990;18(8):2163â€“70.
Hoang T, Yin C, Yau SS. Splice sites detection using chaos game representation and neural network. Genomics. 2020;112(2):1847â€“52.
Joseph J, Sasikumar R. Chaos game representation for comparison of whole genomes. BMC Bioinformatics. 2006;7:243.
Almeida JS, Carrico JA, Maretzek A, et al. Analysis of genomic sequences by chaos game representation. Bioinformatics. 2001;17(5):429â€“37.
Pei S, Dong W, Chen X, et al. Fast and accurate genome comparison using genome images: The Extended Natural Vector Method. Mol Phylogenet Evol. 2019;141:106633.
Huang GB, Zhu QY, Siew CK. Extreme learning machine: theory and applications. Neurocomputing. 2006;70(1â€“3):489â€“501.
Chen T, Guestrin C. XGBoost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining. ACM. 2016. p. 785â€“94.
Li W, Yin Y, Quan X, et al. Gene expression value prediction based on XGBoost algorithm. Front Genet. 2019;10:1077.
LÃ¶chel HF, Eger D, Sperlea T, et al. Deep learning on chaos game representation for proteins. Bioinformatics. 2020;36(1):272â€“9.
Nair V, Hinton GE. Rectified linear units improve restricted Boltzmann machines. In: Proceedings of the 27th international conference on machine learning. Haifa: Omnipress; 2010. p. 807â€“14.
Kingma DP, Ba J. Adam: a method for stochastic optimization. Comp Sci. 2014.
Yin B, Balvert M, Zambrano D, et al. An image representation based convolutional network for DNA classification. In: International conference on learning representations. 2018. https://openreview.net/forum?id=HJvvRoe0W.
Anjum MM, Tahmid IA, Rahman MS. CNN model with hilbert curve representation of DNA sequence for enhancer prediction. bioRxiv. 2019; 552141.
Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems: 2012. p. 1097â€“105.
Lecun Y, Bottou Y, Bengio Y, et al. Gradientbased learning applied to document recognition. Proc IEEE. 1998;86(11):2278â€“324.
Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Proceedings of the 32nd international conference on international conference on machine learning, vol 37. 2015; p. 448â€“456.
Srivastava N, Hinton G, Krizhevsky A, et al. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. 2014;15(1):1929â€“58.
Acknowledgements
The authors would like to thank anonymous reviewers for their constructive comments, which helped us to greatly improve the manuscript.
About this supplement
This article has been published as part of BMC Bioinformatics Volume 22 Supplement 6, 2021: 19th International Conference on Bioinformatics 2020 (InCoB2020). The full contents of the supplement are available online at https://bmcbioinformatics.biomedcentral.com/articles/supplements/volume22supplement6.
Funding
This work was supported in part by Natural Science Foundation of Hunan Province of China (Grant No. 2016JJ3116), Key Foundation of Hunan Educational Committee (Grant No. 19A497). The funding bodies had no role in the design of the study and collection, analysis, and interpretation of data and in writing the manuscript. Publication costs are funded by Key Foundation of Hunan Educational Committee (Grant No. 19A497).
Author information
Authors and Affiliations
Contributions
GSH and YL directed the research. GSH and QL designed the experiments. QL ran all the experiments and wrote the paper. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Han, GS., Li, Q. & Li, Y. Comparative analysis and prediction of nucleosome positioning using integrative feature representation and machine learning algorithms. BMC Bioinformatics 22 (Suppl 6), 129 (2021). https://doi.org/10.1186/s1285902104006w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s1285902104006w
Keywords
 Nucleosome classification
 Frequency chaos game representation
 Support vector machine
 Extreme learning machine
 Extreme gradient boosting
 Convolutional neural networks