 Research article
 Open Access
 Published:
Completing sparse and disconnected proteinprotein network by deep learning
BMC Bioinformatics volume 19, Article number: 103 (2018)
Abstract
Background
Proteinprotein interaction (PPI) prediction remains a central task in systems biology to achieve a better and holistic understanding of cellular and intracellular processes. Recently, an increasing number of computational methods have shifted from pairwise prediction to network level prediction. Many of the existing network level methods predict PPIs under the assumption that the training network should be connected. However, this assumption greatly affects the prediction power and limits the application area because the current golden standard PPI networks are usually very sparse and disconnected. Therefore, how to effectively predict PPIs based on a training network that is sparse and disconnected remains a challenge.
Results
In this work, we developed a novel PPI prediction method based on deep learning neural network and regularized Laplacian kernel. We use a neural network with an autoencoderlike architecture to implicitly simulate the evolutionary processes of a PPI network. Neurons of the output layer correspond to proteins and are labeled with values (1 for interaction and 0 for otherwise) from the adjacency matrix of a sparse disconnected training PPI network. Unlike autoencoder, neurons at the input layer are given all zero input, reflecting an assumption of no a priori knowledge about PPIs, and hidden layers of smaller sizes mimic ancient interactome at different times during evolution. After the training step, an evolved PPI network whose rows are outputs of the neural network can be obtained. We then predict PPIs by applying the regularized Laplacian kernel to the transition matrix that is built upon the evolved PPI network. The results from crossvalidation experiments show that the PPI prediction accuracies for yeast data and human data measured as AUC are increased by up to 8.4 and 14.9% respectively, as compared to the baseline. Moreover, the evolved PPI network can also help us leverage complementary information from the disconnected training network and multiple heterogeneous data sources. Tested by the yeast data with six heterogeneous feature kernels, the results show our method can further improve the prediction performance by up to 2%, which is very close to an upper bound that is obtained by an Approximate Bayesian Computation based sampling method.
Conclusions
The proposed evolution deep neural network, coupled with regularized Laplacian kernel, is an effective tool in completing sparse and disconnected PPI networks and in facilitating integration of heterogeneous data sources.
Background
Studying proteinprotein interaction (PPI) can help us better understand intracellular signaling pathways, model protein complex structures and elucidate various biochemical processes. To aid discovering more denovo PPIs, many computational methods have been developed and can generally be categorized into one of the following three types: (a) pairwise biological similarity based computational approaches by sequence homology, gene coexpression, phylogenetic profiles, threedimensional structural information, etc.; [1–7]; (b) pairwise topological features based methods [8–11]; and (c) whole network structure based methods [1, 12–20].
For the pairwise biological similarity based methods, without resort to determining whether two given proteins will interact from first principles in physics and chemistry, the predictive power of those methods is greatly affected by the features being used, which may be noisy or inconsistent. To circumvent limitations of pairwise biological similarity, network structure based methods are playing an increasing role in PPI prediction since these methods can not only get the whole network structure involved and topological similarities implicitly included, but also utilize pairwise biological similarities as weights for the edges in the networks.
Along this line, variants of random walk [12–15] have been developed. Given a PPI network with N proteins, the computational cost of these methods increases by N times for allagainstall PPI prediction. In Fouss et al. [16], many kernel methods for link prediction have been systematically studied, which can measure the similarities for all node pairs and make prediction at once. Compared to the random walk, kernel methods are usually more efficient. However, neither random walk methods nor kernel methods perform very well in predicting interaction between faraway node pairs in networks [16]. Instead of utilizing network structure explicitly, many latent features based on rank reduction and spectral analysis have also been used to do prediction, such as geometric denoise methods [1, 17], multiway spectral clustering [18], matrix factorization based methods [19, 20]. Note that the objective functions in these methods should be carefully designed to ensure fast convergence and avoid being stuck in local optima. What is advantagous for these methods is that biological features and network topological features can complement each other to improve the prediction performance, such as by weighting network edges with pairwise biological similarity scores [19, 20]. However, one limitation for these methods is that, only the pairwise features for the existing edges in the PPI network are utilized, whereas from a PPI prediction perspective what is particularly useful is to incorporate pairwise features for node pairs that are not currently linked by a direct edge but may become linked. Recently, Huang et al. proposed a sampling method [21] and a linear programming method [22] to find optimal weights for multiple heterogeneous data, thereby building weighted kernel fusion for all node pairs. These methods applied regularized Laplacian kernel (RL) to the weighted kernel fusion to infer missing or new edges in the PPI network. These methods improved PPI prediction performance, especially for detecting interactions between nodes that are far apart in the training network, by using only small training networks.
However, almost all the methods discussed above need the training network to be a single connected component to measure nodepair similarities, despite of the fact that existing PPI networks are usually disconnected. Consequently, these traditional methods only keep the maximum connected component of the original PPI network as golden standard data, which is then divided as a connected training network and testing edges. That is to say, these methods cannot effectively predict interactions for proteins that are not located in the maximum connected component. Therefore, it is of great interest and utility if we can infer PPI network from a small amount of interaction edges that do not need to form a connected network.
From our previous study of network evolutionary analysis [23], we here designed a neural network based evolution model to implicitly simulate the evolution processes of PPI networks. Instead of simulating the evolution of the whole network structure with the growth of nodes and edges as models discussed in Huang et al. [23], we only focus on the edge evolution and assume all nodes are already existing. We initialize the ancient PPI network as an allzero adjacent matrix, and use the disconnected training network with interaction edges as labels. Each row of the allzero adjacent matrix and the training matrix will be used as the input and label for the neural network respectively. We then train the model to simulate the evolution process of interactions. After the training step, we use outputs of the last layer of the neural network to represent rows of the evolved contact matrix. Finally, we further apply the regularized Laplacian kernel to a transition matrix that is built upon the evolved contact matrix to infer new PPIs. The results show our method can efficiently utilize the extremely sparse and disconnected training network, and improve the prediction performances by up to 8.4% for yeast and 14.9% for human PPI data.
Methods
Problem definition
Formally, a PPI network can be represented as a graph G=(V,E) where V is the set of nodes (proteins) and E is the set of edges (interactions). G is defined by the adjacency matrix A with V×V dimension:
where i and j are two nodes in the nodes set V, and (i,j) represents an edge between i and j, (i,j)∈E. We divide the golden standard network into two parts: the training network G_{ tn }=(V,E_{ tn }), and testing set G_{ tt }=(V_{ tt },E_{ tt }), such that E=E_{ tn }∪E_{ tt }, and any edge in G can only belong to one of these two parts. The detailed process of dividing the golden standard network is shown by Algorithm 1. We set the α (the preset ratio of G_{ tn }(,E) to G(,E)) less than a small value to make the G_{ tn } extremely sparse and with a large number of disconnected components.
Figure 1 shows the flow chart of our method, which is named as evolution neural network based regularized Laplcian kernel (ENNRL) to reflect the fact that it contains two steps. The first step, ENN, uses the sparse disconnected training network of PPIs to train a deep neural network in order to obtain an “evolved” and more complete network, and this “evolved” network is then used as a transition matrix for the regularized Laplacian kernel in the second step to predict PPIs for node pairs that are not directed connected. Inspired by the structure of autoencoder [24], the architecture of the neural network is designed to “evolve” a partial PPI network, guided by its current connection, via a smaller proteome (i.e., a smaller hidden layer) at an “ancient” time, asssuming zero a priori knowledge of existing PPIs at the input. Specifically, as shown in Fig. 2, with respect to the smaller hidden layer in the middle, the input layer looks like a symmetric mirror image of the output layer. But all nodes in the input layer have zero value input, reflecting the assumption of zero a priori knowledge about interactions between proteins that these nodes represent, i.e., the input m×m adjacency matrix in Fig. 1 contains all zeros, where m=V. And the output layer nodes have labels from the training PPI adjacency matrix. Then, deep learning is adopted to drive and guide the neural network from a blank input to first “devolve” into a smaller hidden layer (representing an interactome at ancient time) and then “evolve” into the output layer, which has the training PPI network G_{ tn } as the target/label. The rationale is that, if PPI interactions in the training network can be explained (or reproduced) via a smaller ancient interactome, the such trained neural network should be also capable of generalizing and predicting unobserved de novo PPIs. To avoid exactly producing the training PPI networks, a blank adjacency matrix is used as the input.
After the training process is completed, we build the evolved PPI network/matrix EA with the outputs of neural network’s last layer. Based on EA, we build a transition matrix using Eq. (2), where EA+EA^{′} makes the transition matrix symmetric and positive semidefinite. Finally, we apply the regularized Laplacian (RL) kernel defined by Eq. (3) to the transition matrix T to get the inference matrix P, in which P_{i,j} indicates the probability of an interaction for protein i and j. For Eq. (3), L=D−T is the Laplacian matrix made of the transition matrix T and the degree matrix D, and 0<α<ρ(L)^{−1} and ρ(L) is the spectral radius of L.
Algorithm 2 describes the detailed training and prediction processes.
Evolution neural network
The structure of the evolution neural network is shown in the Fig. 2, which contains five layers including the input layer, three hidden layers and the output layer. Sigmoid is adopted as the activation function for each neuron, and layers are connected with dropouts. Dropouts can not only help us prevent overfitting, but also indicate the mutation events during the evolution processes, such as which nodes (representing proteins) at a layer (corresponding a time during evolution) may be evolved from some nodes from the previous layer, as indicated by edges and corresponding weights connecting those nodes.
For specific configuration of the neural network in our experiments, the number of neurons in the input and output layer depends on the network size m=V of specific PPI data. Each protein is represented by the corresponding row of the adjacency matrix trainingAdj of G_{ tn } that contains the interaction information for that protein with other proteins in the proteome. We train the evolution neural network by each row of the blank adjacency matrix as the input and the corresponding row of trainingAdj as the label. A typical autoencoder structure is chosen for the three hidden layers, where encoder and decoder correspond to the biological devolution and evolution processes respectively; and cross entropy is used as the loss function. Note that, the correspondence of encoder/decoder to biological devolution/evolution is at this stage more of an analogy in helping with the design of the neural network structure than a real evolution mode for PPI networks. It is also worth noting that different with the traditional autoencoder, we did not include the layerwise isomorphism pretraining to initial the weights for our neural network since the inputs are all zero vectors. The neural network is implemented by the TensorFlow library [25], deployed on Biomix cluster at Delaware Biotechnology Institute.
Data
We use yeast and human PPI networks downloaded from DIP (Release 20140117) [26] and HPRD (Release 9) [27] to train and test our method. After removing the selfinteractions, the detailed information of these two datasets are shown in the Table 1.
Results
Experiments on yeast PPI data
To show how well our model can predict PPIs from the extremely sparse training network with disconnected components, we set α, the ratio of interactions in G_{ tn } to the total edges in G, to be less than 0.25. As shown in Table 2, the G_{ tn } has only 4061 interactions, and contains 2812 disconnected components, where the minimum, average and maximum size of components are 1, 1.81 and 2152 respectively. Based on the G_{ tn }, we train our model and predict the large testing set G_{ tt } that has 18,362 interactions according to the Algorithm 2.
We then compared our ENNRL method to the control method ADJRL which applies regularized Laplacian kernel directly to the training network G_{ tn }. As shown in Fig. 3, the AUC increase from 0.8112 for the control method to 0.8358 for ENNRL. Moreover, we make the prediction task more challenging by setting the α to be less than 0.125, which makes the G_{ tn } sparser with only 1456 interactions, but 3915 disconnected components; and the maximum component in G_{ tn } only has 1006 interactions. The results in Fig. 4 shows the gap between ENNRL ROC curve and ADJRL ROC curve is obviously increased; and our ENNRL gained 8.39% improvement in AUC. If comparing Figs. 3 and 4, it is easy to see that the AUC of ADJRL decreases by 0.055 from 0.8112 in Fig. 3 to 0.7557 in Fig. 4. However, our ENN method performs stably with only 0.016 decrease in AUC. This suggests that traditional random walk methods usually need the training network to be connected; and the prediction performance largely depends on the size and density of the maximum connected component. However, when the training network becomes sparse and disconnected, the traditional random walk based methods will lose the predictive power likely because they cannot predict interactions among those disconnected components. We repeated the whole experiments up to ten times, Table 3 shows the average performance with the standard deviation. All these results show our method performs stably and effectively in overcoming the limitation of traditional random walk based methods; and the improvements are statistically significant.
Moreover, we further analyzed how our ENNRL method can effectively predict interactions to connect disconnected components, and how its intracomponent and crosscomponent predicting behaviors adaptively change with different training networks. As defined in the “Methods” section, the value P_{i,j} in the inference matrix P indicates the probability of an interaction for protein i and j. We ranked all the protein pairs by their value in the inference matrix P; ideally we can choose a optimal threshold on the value of P_{i,j} to make prediction. Since it is difficult to find the optimal threshold without prior knowledge, we used a ratio ρ instead. Specifically, for the ranked protein pairs, the top ρ∗100 percent can be considered as predicted positives G_{ pp }, and the predicted true positive G_{ ptp } is the intersection set between G_{ pp } and G_{ tt }. We then added the interactions in G_{ ptp } to the training network G_{ tn } to see how many disconnected components can become reconnected. The results are shown in the Fig. 5, the dashed lines indicate the number of disconnected components in the training networks G_{ tn }; the solid lines with markers indicate that how the number of disconnected components would change based on prediction with different ρ (The red color is for the case α=0.125, the blue color is for α=0.25). As it shows, our methods can effectively predict interactions to reconnect those disconnected components in the training networks for both cases; especially, for the training network of α=0.125. The comparison of the results of those two cases shows that the red solid line decreases significantly faster than the blue solid line. It demonstrates that, for the training network α=0.25 that has fewer but larger size disconnected components, the prediction of our method recovers more intracomponent interactions; whereas for the training network α=0.125 that has more and smaller size disconnected components, the prediction of our method can more effectively recover crosscomponent interactions, which are more difficult for the traditional random walk methods to detect. Therefore, to a large extent, this explains why the performance difference between our method and the traditional random walk method ADJRL is relatively small in Fig. 3 but more pronounced in Fig. 4, because in the latter case our method has clear advantage in detecting crosscomponent interaction.
Experiments on human PPI data
We further tested our method by the human PPI data downloaded from HPRD (Release 9) [27], which is much larger and sparser than the yeast PPI network. Similarly, we carried out two comparisons by setting the α to be less than 0.25 and 0.125 respectively to divide G in to training network G_{ tn } and testing set G_{ tt }. The detailed information about the division can be found in the Table 4.
The prediction performances in Figs. 6 and 7 show our ENNRL has obviously better ROC curves and higher AUC than that of ADJRL. Especially for the test with α=0.125, our ENNRL method obtains up to 14.9% improvement for predicting 34,779 testing interactions based on a training set G_{ tn } with only 2260 interactions but 7667 disconnected components. Similar tendency is also observed from Figs. 6 and 7. When α is decreased from 0.25 to 0.125, the AUC of ADJRL decreases by up to 0.072, while our ENNRL only decreased by 0.021. We also did ten repetitions as shown in Table 5 to demonstrate the stable performance of the ENNRL. All these results on human PPI data further indicate our ENNRL model is a promising tool to predict edges for any sparse and disconnected training network.
Moreover, similar to the experiments we did for the yeast data, we also analyzed the crosscomponent interaction prediction performance on HPRD human data. The result shown in the Fig. 8 is consistent with the result of yeast data. Our method can effectively predict interactions to connect disconnected components in both training networks (α=0.125 and α=0.25); and the red solid line decrease remarkably faster than the blue solid line. All these results further support the conclusion we made in the last section.
Optimize weights for heterogeneous feature kernels
Most recently, Huang et al. [22, 28] developed a method to infer de novo PPIs by applying regularized Laplacian kernel to a kernel fusion that based on optimally weighted heterogeneous feature kernels. To find the optimal weights, they proposed weight optimization by linear programming (WOLP) method that based on random walk over a connected training networks. Firstly, they utilized Barker algorithm and the training network to construct a transition matrix which constrains how a random walk would traverse the training network. Then the optimal kernel fusion can be obtained by adjusting the weights to minimize the elementwise difference between the transition matrix and the weighted kernels. The minimization problem is solved by linear programming.
Given a large disconnected network, although Huang et al. [22] demonstrated that the weights learned from the maximum connected component can also be used to build kernel fusion for that large disconnected network, the weights will not be optimal when the maximum connected component is very small compared to the original disconnected network. As we all know that current available golden standard PPI networks are usually disconnected and remains far from complete. Therefore, it would be of great interest if we can obtain the transition matrix directly from these disconnected components, including but to limited to the maximum connected component, and use that transition matrix to help us find the optimal weights for heterogeneous feature kernels. To verify this idea, we use the transition matrix T obtained by Eq. (2) to find the optimal weights based on the linear programming Eq. (4) [22].
We tested this method by the yeast PPI network with same setting in Table 2; and six feature kernels are included: G_{ tn }: G_{ tn } is training network with α=0.25 or 0.125 in Table 2. K_{ Jaccard } [29]: This kernel measure the similarity of protein pairs i,j in term of \(\frac {neigbors(i) \cap neighbors(j)}{neighbors(i) \cup neighbors(j)}\).
K_{ SN }: It measures the total number of neighbors of protein i and j, K_{ SN }=neighbors(i)+neighbors(j). K_{ B } [30]: It is a sequencebased kernel matrix that is generated using the BLAST [31]. K_{ E } [30]: This is a gene coexpression kernel matrix constructed entirely from microarray gene expression measurements. K_{ Pfam } [30]: Similarity measure derived from Pfam HMMs [32]. All these kernels are normalized to the scale of [ 0,1] in order to avoid bias.
Discussion
To make a comprehensive analysis, we also included prediction results based on the kernel fusion built by the approximate bayesian computation and modified differential evolution sampling (ABCDEP) method [21], and the kernel fusion built by equally weighted feature kernels EK for comparison. Similar to the comparison in [22], the ABCDEP and EK based results can serve as the upper bound and lower bound of the prediction performance. Comparisons for two settings α=0.25 and α=0.125 are shown by the Figs. 9 and 10 respectively, where ENNRL is our proposed method without integrating any heterogeneous feature kernels; ENNlpRL is a kernel fusion method and is a combination of ENNRL and the linear programming optimization method WOLP [22], which use the transition matrix T obtained by ENNRL as the target transition matrix to find the optimal weights for heterogeneous feature kernels by linear programming; ABCDEPRL is also a kernel fusion method and finding the optimal weights based on a sampling method [23]; ADJRL is the traditional random walk method; and EWRL is a baseline kernel fusion method that weights all heterogeneous feature kernels equally. Note that, the transition matrix T obtained from ENNRL is not a feature kernel and only serve as the target transition matrix for ENNlpRL to optimize weights.
As shown in Fig. 9, ENNlpRL benefits from a more complete and accurate target transition matrix provided by ENNRL, outperforms other methods and gets very close to the upper bound 0.8529 achieved by ABCDEPRL. Similarly, in Fig. 10, although the maximum component of G_{ tn } is very sparse – with 1006 proteins and only 1456 training interactions, the ENNlpRL still is enhanced from the ENNRL and gets very close to the ABCDEPRL. Therefore, all these results indicate that the transition matrix T learned by our ENN model can further improve the prediction performance for other downstream tools like WOLP in leveraging useful information from heterogeneous feature kernels.
Conclusions
In this work we developed a novel method based on deep learning neural network and regularized Laplacian kernel to predict de novo interactions for sparse and disconnected PPI networks. We built the neural network with a typical autoencoder structure to implicitly simulate the evolutionary processes of PPI networks. Based on the supervised learning using the rows of a sparse and disconnected training network as labels, we can obtain an evolved PPI network as the outputs of the deep neural network, which has an input layer identical to the output layer but with zero input value and a smaller hidden layer simulating an ancient interactome. Then we predicted PPIs by applying regularized Laplacian kernel to the transition matrix built upon that evolved PPI network. Tested on DIP yeast PPI network and HPRD human PPI network, the results show that our method exhibits competitive advantages over the traditional regularized Laplacian kernel that based on the training network only. The proposed method achieved significant improvement in PPI prediction, as measured by ROC score, over 8.39% higher than the baseline for yeast data, and 14.9% for human data. Moreover, the transition matrix learned from our evolution neural network can also help us to build optimized kernel fusion, which effectively overcome the limitation of traditional WOLP method that needs a relatively large and connected training network to obtain the optimal weights. Then we also tested it by the DIP yeast data with six feature kernels, the prediction result shows the AUC can be further improved and very close to the upper bound. Given the current golden standard PPI networks are usually disconnected and very sparse, we believe our model provides a promising tool that can effectively utilize disconnected networks to predict PPIs. In this paper, we designed the autoencoder deep learning structure analogous to the evolution process of PPI network, which, although should not be interpreted as a real evolution model of PPI networks, would nonetheless be worthwhile to explore further for the future work. Meanwhile, we also plan to investigate other deep learning models for solving PPI prediction problems.
Abbreviations
 ABCDEP:

Approximate bayesian computation and modified differential evolution sampling
 ADJRL:

Adjacency matrix based regularized Laplacian kernel
 AUC:

Area under the curve
 ENN:

Evolution neural network
 ENNRL:

Evolution neural network based regularized Laplacian kernel
 PPI:

Protein protein interaction
 RL:

Regularized Laplacian kernel
 ROC:

Receiver operating characterisitic
 WOLP:

Weight optimization by linear programming
References
 1
Kuchaiev O, Rašajski M, Higham DJ, Pržulj N. Geometric denoising of proteinprotein interaction networks. PLoS Comput Biol. 2009; 5(8):1000454.
 2
Murakami Y, Mizuguchi K. Homologybased prediction of interactions between proteins using averaged onedependence estimators. BMC Bioinformatics. 2014; 15(1):213.
 3
Salwinski L, Eisenberg D. Computational methods of analysis of protein–protein interactions. Curr Opin Struct Biol. 2003; 13(3):377–82.
 4
Craig R, Liao L. Phylogenetic tree information aids supervised learning for predicting proteinprotein interaction based on distance matrices. BMC Bioinformatics. 2007; 8(1):6.
 5
Gonzalez A, Liao L. Predicting domaindomain interaction based on domain profiles with feature selection and support vector machines. BMC Bioinformatics. 2010; 11(1):537.
 6
Zhang QC, Petrey D, Deng L, Qiang L, Shi Y, Thu CA, Bisikirska B, Lefebvre C, Accili D, Hunter T, Maniatis T, Califano A, Honig B. Structurebased prediction of proteinprotein interactions on a genomewide scale. Nature. 2012; 490(7421):556–60.
 7
Singh R, Park D, Xu J, Hosur R, Berger B. Struct2net: a web service to predict protein–protein interactions using a structurebased approach. Nucleic Acids Res. 2010; 38(suppl 2):508–15.
 8
Chen HH, Gou L, Zhang XL, Giles CL. Discovering missing links in networks using vertex similarity measures. In: Proceedings of the 27th Annual ACM Symposium on Applied Computing. SAC ’12. New York: ACM: 2012. p. 138–43.
 9
Lü L, Zhou T. Link prediction in complex networks: A survey. Physica A. 2011; 390(6):11501170.
 10
Lei C, Ruan J. A novel link prediction algorithm for reconstructing proteinprotein interaction networks by topological similarity. Bioinformatics. 2013; 29(3):355–64.
 11
Pržulj N. Proteinprotein interactions: Making sense of networks via graphtheoretic modeling. BioEssays. 2011; 33(2):115–23.
 12
Page L, Brin S, Motwani R, Winograd T. The pagerank citation ranking: Bringing order to the web. Technical Report 199966, Stanford InfoLab. Previous number = SIDLWP19990120. 1999. http://ilpubs.stanford.edu:8090/422/.
 13
Tong H, Faloutsos C, Pan JY. Random walk with restart: fast solutions and applications. Knowl Inf Syst. 2008; 14(3):327–46.
 14
Li RH, Yu JX, Liu J. Link prediction: The power of maximal entropy random walk. In: Proceedings of the 20th ACM International Conference on Information and Knowledge Management. CIKM ’11. New York: ACM: 2011. p. 1147–1156. https://doi.org/10.1145/2063576.2063741.
 15
Backstrom L, Leskovec J. Supervised random walks: Predicting and recommending links in social networks. In: Proceedings of the Fourth ACM International Conference on Web Search and Data Mining. WSDM ’11. New York: ACM: 2011. p. 635–44.
 16
Fouss F, Francoisse K, Yen L, Pirotte A, Saerens M. An experimental investigation of kernels on graphs for collaborative recommendation and semisupervised classification. Neural Netw. 2012; 31(0):53–72.
 17
Cannistraci CV, AlanisLobato G, Ravasi T. Minimum curvilinearity to enhance topological prediction of protein interactions by network embedding. Bioinformatics. 2013; 29(13):199–209.
 18
Symeonidis P, Iakovidou N, Mantas N, Manolopoulos Y. From biological to social networks: Link prediction based on multiway spectral clustering. Data Knowl Eng. 2013; 87(0):226–42.
 19
Wang H, Huang H, Ding C, Nie F. Predicting protein–protein interactions from multimodal biological data sources via nonnegative matrix trifactorization. J Comput Biol. 2013; 20(4):344–58. https://doi.org/10.1089/cmb.2012.0273.
 20
Menon AK, Elkan C. Link prediction via matrix factorization. In: Proceedings of the 2011 European Conference on Machine Learning and Knowledge Discovery in Databases  Volume Part II. ECML PKDD’11. Berlin: Springer: 2011. p. 437–52.
 21
Huang L, Liao L, Wu CH. Inference of proteinprotein interaction networks from multiple heterogeneous data. EURASIP J Bioinforma Syst Biol. 2016; 2016(1):1–9. https://doi.org/10.1186/s1363701600402.
 22
Huang L, Liao L, Wu CH. Proteinprotein interaction prediction based on multiple kernels and partial network with linear programming. BMC Syst Biol. 2016; 10(2):45. https://doi.org/10.1186/s129180160296x.
 23
Huang L, Liao L, Wu CH. Evolutionary model selection and parameter estimation for proteinprotein interaction network based on differential evolution algorithm. IEEE/ACM Trans Comput Biol Bioinforma. 2015; 12(3):622–31. https://doi.org/10.1109/TCBB.2014.2366748.
 24
Bengio Y. Learning deep architectures for ai. Found Trends Mach Learn. 2009; 2(1):1–127. https://doi.org/10.1561/2200000006.
 25
Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, Ghemawat S, Goodfellow I, Harp A, Irving G, Isard M, Jia Y, Jozefowicz R, Kaiser L, Kudlur M, Levenberg J, Mané D, Monga R, Moore S, Murray D, Olah C, Schuster M, Shlens J, Steiner B, Sutskever I, Talwar K, Tucker P, Vanhoucke V, Vasudevan V, Viégas F, Vinyals O, Warden P, Wattenberg M, Wicke M, Yu Y, Zheng X. TensorFlow: LargeScale Machine Learning on Heterogeneous Systems. Software available from tensorflow.org. 2015. http://tensorflow.org/.
 26
Salwinski L, Miller CS, Smith AJ, Pettit FK, Bowie JU, Eisenberg D. The database of interacting proteins: 2004 update. Nucleic Acids Res. 2004; 32(90001):449–51.
 27
Keshava Prasad TS, Goel R, Kandasamy K, Keerthikumar S, Kumar S, Mathivanan S, Telikicherla D, Raju R, Shafreen B, Venugopal A, Balakrishnan L, Marimuthu A, Banerjee S, Somanathan DS, Sebastian A, Rani S, Ray S, Harrys Kishore CJ, Kanth S, Ahmed M, Kashyap MK, Mohmood R, Ramachandra YL, Krishna V, Rahiman BA, Mohan S, Ranganathan P, Ramabadran S, Chaerkady R, Pandey A. Human protein reference database2009 update. Nucleic Acids Res. 2009; 37(suppl 1):767–72.
 28
Huang L, Liao L, Wu CH. Proteinprotein interaction network inference from multiple kernels with optimization based on random walk by linear programming. In: Proceedings of 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). Washington DC: IEEE computer society: 2015. p. 201–7.
 29
Jaccard P. Étude comparative de la distribution florale dans une portion des Alpes et des Jura. Bulletin del la Société Vaudoise des Sci Nat. 1901; 37:547–79.
 30
Lanckriet GRG, De Bie T, Cristianini N, Jordan MI, Noble WS. A statistical framework for genomic data fusion. Bioinformatics. 2004; 20(16):2626–635.
 31
Altschul SF, Gish W, Miller W, Myers EW, Lipman DJ. Basic local alignment search tool. J Mol Biol. 1990; 215(3):403–10.
 32
Sonnhammer ELL, Eddy SR, Durbin R. Pfam: A comprehensive database of protein domain families based on seed alignments. Proteins Struct Funct Bioinforma. 1997; 28(3):405–20.
Acknowledgements
The authors are grateful to the anonymous reviewers for their valuable comments and suggestions.
Funding
Publication of this article is funded by Delaware INBRE program, with grant from the National Institute of General Medical SciencesNIGMS (8 P20 GM10344612) from the National Institutes of Health. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
Availability of data and materials
The code for this work can downloaded from the project homepage. https://www.eecis.udel.edu/~lliao/enn/The data were downloaded from reference [26] http://dip.mbi.ucla.edu/dip/and refrence [27] http://www.hprd.org/.
Author information
Affiliations
Contributions
LH designed the algorithm and experiments, and performed all calculations and analyses. LL and CHW aided in interpretation of the data and preparation of the manuscript. LH wrote the manuscript, LL and CHW revised it. LL and CHW conceived of this study. All authors have read and approved this manuscript.
Corresponding author
Correspondence to Li Liao.
Ethics declarations
Ethics approval and consent to participate
No human, animal or plant experiments were performed in this study, and ethics committee approval was therefore not required.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Huang, L., Liao, L. & Wu, C. Completing sparse and disconnected proteinprotein network by deep learning. BMC Bioinformatics 19, 103 (2018). https://doi.org/10.1186/s1285901821127
Received:
Accepted:
Published:
Keywords
 Disconnected protein interaction network
 Neural network
 Interaction prediction
 Network evolution
 Regularized Laplacian