 Methodology article
 Open Access
 Published:
Graphbased prediction of Proteinprotein interactions with attributed signed graph embedding
BMC Bioinformatics volume 21, Article number: 323 (2020)
Abstract
Background
Proteinprotein interactions (PPIs) are central to many biological processes. Considering that the experimental methods for identifying PPIs are timeconsuming and expensive, it is important to develop automated computational methods to better predict PPIs. Various machine learning methods have been proposed, including a deep learning technique which is sequencebased that has achieved promising results. However, it only focuses on sequence information while ignoring the structural information of PPI networks. Structural information of PPI networks such as their degree, position, and neighboring nodes in a graph has been proved to be informative in PPI prediction.
Results
Facing the challenge of representing graph information, we introduce an improved graph representation learning method. Our model can study PPI prediction based on both sequence information and graph structure. Moreover, our study takes advantage of a representation learning model and employs a graphbased deep learning method for PPI prediction, which shows superiority over existing sequencebased methods. Statistically, Our method achieves stateoftheart accuracy of 99.15% on Human protein reference database (HPRD) dataset and also obtains best results on Database of Interacting Protein (DIP) Human, Drosophila, Escherichia coli (E. coli), and Caenorhabditis elegans (C. elegan) datasets.
Conclusion
Here, we introduce signed variational graph autoencoder (SVGAE), an improved graph representation learning method, to automatically learn to encode graph structure into lowdimensional embeddings. Experimental results demonstrate that our method outperforms other existing sequencebased methods on several datasets. We also prove the robustness of our model for very sparse networks and the generalization for a new dataset that consists of four datasets: HPRD, E.coli, C.elegan, and Drosophila.
Background
Proteins are versatile macromolecules and perform a vast array of vital functions within organisms, and over 80% of proteins interact with other proteins while carrying out their functions [1]. Those interactions, known as proteinprotein interactions (PPIs), are physical contacts of high specificity established between two or more protein molecules. PPI is of great importance in many cellular biological processes, including signal transduction, immune response, cell proliferation, DNA transcription, and replication. Analysis and elucidation of the PPIs provide valuable insights into the molecular mechanism and the protein functions [2]. In recent years, the rapid development of highthroughput technologies are used to detect protein interactions, such as yeast twohybrid screens (Y2H) [3], tandem affinity purification (TAP) [4] and mass spectrometric protein complex identification (MSPCI) [5], Tandem Affinity Purification and Mass Spectrometry (TAPMS) [6], affinity chromatography and CoImmunoprecipitation (CoIP) [7]. These experimental methods have contributed to exponential growth of the number of PPIs of various species, but the functional annotation of both proteins and their interactions is updated at a slow speed. Meanwhile, these data suffer from problems including high false positives, false negative rate and low coverage [8]. To be more specific, although many proteinprotein interaction links have been experimentally determined, the total number is still relatively few compared to the tremendous amount of links collected by the highthroughput technologies [9]. And these genomescale experiments are costly, with inherent bias and limited coverage. The limitations of device resolution and environmental interference during operation will inevitably lead to errors and deviations in experimental techniques [10]. Therefore, highthroughput computational methods that are useful for the study of protein functions are required for discovering PPI with high quality and accuracy [11].
Recently, many highthroughput computational methods have been proposed. On the whole, they can be divided into two groups: classic machine learning algorithms and deep learning methods. For the first group, different machine learning methods were utilized for predicting PPIs to improve the efficiency and accuracy, such as decision trees [12], kNearest Neighbor (KNN) [13], naive bayes [14], random forest [15] and support vector machine (SVM) [16–18]. These features of these methods measure physicochemical properties of the 20 canonical amino acids, and aim at summarizing full sequence information relevant to PPIs. Compared to classic machine learning methods, deep learning methods are advantaged in extracting features directly from data and capture nonlinear dependencies between abstract features. They can also fully exploit the availability of the increasing largescale and highdimension raw datasets. Therefore, deep learning methods are unprecedentedly popular in recent years and have been successfully applied in various problems [19]. For PPI prediction, Sun et al. recently proposed a stacked autoencoder (SAE) to study the sequencebased PPI prediction, which was the first to apply a deep learning algorithm to sequencebased PPI prediction and achieved promising results [20]. Du et al. proposed a method called Deep neural networks for Protein Protein Interactions prediction (DeepPPI), which employed deep learning to extract highlevel discriminative features from common protein descriptors [21]. Lei et al. put forward a novel computational method based on Multimodal Deep Polynomial Network (MDPN) to encode multiple data from protein properties for PPIs prediction [22]. Hashemifar et al. presented a convolutionbased model where feature extractions are terminated by processing data through an original randomly initialized and untrained matrix they named “random projection module" [23]. Next, a neural network based approach called Ensemble Deep Neural Networks (EnsDNN) was proposed to predict PPIs based on different representations of amino acid sequences [24]. Particularly, EnsDNN separately used auto covariance descriptor, local descriptor, and multiscale continuous and discontinuous local descriptor, to represent and explore the pattern of interactions between sequentially distant and spatially close amino acid residues. Finally, Richoux et al. compared two carefully designed deep learning models and showed pitfalls to avoid while predicting PPIs through deep learning methods [25].
However, the deep learning algorithm presented by Sun et al. [20] and most of the methods we discussed above only considered sequence data, while the network data, such as their degree, position, and neighboring nodes in the graph, has been proved to be informative in PPI prediction. For example, Licamele and Getoor looked at the shared neighborhood among proteins and calculated the clustering coefficient among the neighborhoods for the firstorder and secondorder protein relations to predict the interactions in a yeast dataset [26]. Paradesi et al. identified nine structural features for Saccharomyces cerevisiae PPI networks and used them to learn classifiers for predicting new interactions [27]. You et al. developed a robust manifold embedding technique for assessing the reliability of interactions and predicting new interactions by utilizing the topological information of PPI networks [28].
The biggest challenge to apply graphbased deep learning methods for PPI prediction is the utilization way of the network information, that is, how to represent the graph structure of PPI network in lowdimensional embeddings, which should be used as feature inputs for downstream machine learning classifier. The good news is that there has been a surge in approaches that automatically learn to encode graph structure recently, using techniques based on deep learning and nonlinear dimensionality reduction. These methods are representation learning on graphs, which can be used to analyze social networks, molecular graph structures and recommender systems. The idea behind the representation learning approach is to learn a mapping that embeds nodes, or entire graphs, as points in a lowdimensional vector space. The purpose is to optimize the mapping so that geometric relationships in the learned space could reflect the structure of the original graph [29]. Representation learning has been successfully applied to link prediction, such as predicting missing friendship links in social networks [30] and inferring affinities between users and movies [31].
In this paper, by regarding PPI network as an undirected graph, we propose signed variational graph autoencoder (SVGAE), a representation learning model that could effectively take advantage of the graph structure and naturally incorporate protein sequence information as features. Our overall framework is composed of three parts. The first part is designed to code raw protein sequences, and the second part is the essential SVGAE model used to further extract vector embedding for each protein with both graph structure and sequence information. The final part is a simple threelayer softmax classifier. Our SVGAE model is designed based on the variational graph autoencoder (VGAE) model proposed by Kipf and Welling [32], which is a framework that makes use of latent variables and is capable of learning interpretable representations for undirected graphs. To apply it to efficiently predict PPI, we primarily made three key improvements on the VGAE and greatly boosted the ultimate performance. Firstly, we modified the cost function to only consider those interactions with high confidence, which allowed us to learn accurate feature representation by focusing on highconfidence interaction information and was more robust to noise. In addition, we gave different signs to different interactions in the adjacency matrix so that the model could consider different impact of each interaction during the training process and strengthen the negative impact of the highly negative interactions. The last improvement was that we further train a neural network as the final classifier instead of using generative model to infer interactions. Since the input embedded representations already contained enough information, using the simple classifier is sufficient according to Occam’s Razor Principle [33].
Results
Our work consists of three steps: basic protein sequence coding, graphbased feature extraction model, and the final neural network classifier. Firstly, we transform raw protein sequences into fixedlength coding using the conjoint triad (CT) method. Next, we propose an improved weighted variational graph autoencoder (SVGAE) to learn embeddings for each protein based on their sequence features and local graph information. Finally, we use these embeddings as inputs to train a simple feedforward neural network as the final classifier. In this section, we firstly evaluated the performance of the proposed method for predicting five different datasets: Human protein reference database (HPRD) dataset, Database of Interacting Protein (DIP) Human, Drosophila, Escherichia coli (E. coli), and Caenorhabditis elegans (C. elegan) by using different evaluation measures. We then compared the performance of the proposed method with existing methods from previous literature are presented. Finally, we discussed the robustness and the generalization of the proposed model.
Evaluation criteria
In this paper, the performance of the proposed model was evaluated by means of the classification accuracy, specificity, sensitivity and precision, Fscore value, as defined respectively by:
where TP, TN, FP and FN represent true positive, true negative, false positive, and false negative, respectively.
Comparison with other methods
In order to demonstrate the performance of our model, we evaluated our model on five datasets as described in the “Datasets” section and compared our model to several popular methods. As indicated in Table 1, our method achieved above 98.5% accuracy on all datasets. For 2007 HPRD, Drosophila and C.elegan datasets, the Fscore values of our model are more than 99%. As shown in Figs. 1 and 2, on the 2007 HPRD dataset, our model achieved stateoftheart Fscore value of 99.15% compared to eight popular existing methods. For example, Sun’s [20] obtained prediction Fscore value of 97.16% and prediction accuracy of 97.19%. And Pan’s [34] work obtained prediction accuracy of 97.90% of latent dirichlet allocationrandom forest (LDAROF) and prediction Fscore value of 90.4% of latent dirichlet allocationsupport vector machine (LDASVM) respectively. In summary, our model achieved the best prediction capacity.
The detailed results of our method on the DIP Human, E. coli, Drosophila and C. elegans were listed in Table 2. Our model was compared against multiple baseline approaches, including: SAE [20], Lasagna [35], DeepPPI [25]. The results of SAE [20] were attained from the data provided by Guo et al. [18]. As the other methods of SVM [18], LDAROF [34], compressive samplingsupport vector machine (CSSVM) [36], and extreme learning machine (ELM) [37] were not conducted on these four datasets. Particularly, Richoux et al. [25] proposed to compare two different neural network architectures: Deep Fully Connected Network (DFC) and Deep Long Short Memory Network (DLSTM). Our model was compared with these two models respectively. In addition, since the DFC and DLSTM methods were applicable to largescale datasets which were designed to solve the problem of information leak and didn’t avoid the underfitting in the small datasets, we didn’t use their results on E. coli and C. elegans datasets.
On the DIP Human dataset, our model yielded a Fscore value of 98.78% which greatly outperformed the model named Lasagna [35] being 97.25% and the model named DFC [25] being 95.14%. For E. coli, our model achieved the Fscore value of 98.92%, which was also significantly superior to other methods. For Drosophila, our model obtained the Fscore value of 99.80% while for C.elegans the Fscore value was 99.25%. It can be seen that our model has demonstrated promising results on several datasets and has proved its potential in PPI prediction regardless of the dataset’s size, coverage, and species.
Robustness and generalization
First, we discuss the robustness of our model for very sparse datasets since the existing PPIs are limited [38]. We define coverage as the proportion of training samples to the number of total samples. The training set is selected as each protein includes coverage portion of its positive edges and highly negative edges. The less the coverage, the more sparse the training set.
As we can see from Fig. 3, the accuracy, sensitivity, specificity, and precision are all rising as the coverage increases. It can be observed from Fig. 3a that the accuracy is already above 93% even when the coverage is only 0.1, which illustrates that our model is effective even when the datasets are sparse and can be applied in real prediction of other species. Besides, as indicated in Fig. 3b, the sensitivity is up to 99% at the coverage of only 0.4. Figure 3c and d also demonstrate the robustness and validity of our model consistently.
To test the generalization, we combine HPRD, E.coli, C.elegan and Drosophila datasets into one larger dataset. This new dataset consists of four different species and contains 21881 proteins with 69550 positive samples and 69283 negative samples. We randomly split the dataset into 50% training samples and 50% test samples and this process is repeated five times. The training data was trained for 50 iterations and the average of five results denotes the final score. As we can see in Table 3, the accuracy, sensitivity, specificity and precision, Fscore value are all more than 96%. Considering the heterogeneity and sparsity of this dataset, the performance is reasonable and desirable.
Discussion
With a graph representation learning model, our method is demonstrated effective and robust in PPI prediction. However, the model still has great improvement space as follows. In our model, considering different impacts of different edges, we introduce the mechanism of confidence for each edge in the PPI network which can be reflected in the adjacency matrix A. Currently, we divide all the edges into three groups and assign constant value for each edge in the same group. In the future work, the quality of each interaction should be taken into account and assigned a specific confidence value to make the prediction model more informative.
For protein sequence coding, we used the predefined CT method. Other popular coding methods such as auto covariance (AC) and local descriptor (LD), the manually constructed or selected representation is more or less polluted or biased, which would affect the learning ability of the deep learning method. But the CT method is based on rules and the error will be smaller than these two method. Therefore, developing more precise coding methods is crucial to further improve the model in our future work. As we all known, the sequence of nucleotides that forms a gene is first translated into an amino acid sequence, following the rules encoded in the genetic code. The corresponding linear chain of amino acids becomes functional only when it adopts a threedimensional shape, the socalled tertiary, or native structure of the protein. These 3D structures of proteins provide the opportunity for in silico prediction methods. The opportunity is that if in silico methods can predict whether two given 3D structures interact, then these methods may be applied to predict interactions among the large amount of proteins with known or inferred 3D structure [39]. Our future work could integrate the text description information annotated in the database to the codings using natural language processing technique and the 3D structures of proteins to better represent the protein.
Conclusions
In this paper, we proposed SVGAE, an improved graph representation learning method, to incorporate graph information in PPI networks into PPI prediction. Then the abstract features are based on both sequence information and graph structure. Experimental results demonstrated that our method performed significantly well and outperformed other existing sequencebased methods on several datasets. We also proved the robustness of our model for very sparse networks and the generalization for different kinds of datasets. To the best of our knowledge, our method is one of the first models to apply graphbased representation learning technique, thus successfully apply a deep learning algorithm to graphbased PPI prediction. It is also anticipated that our method can be generalized to many other related bioinformatics studies. For example, we can conduct representation learning with graph and simplified molecular input line entry specification (SMILES) string features of drug molecules using SVGAE to predict drug interactions. We can construct an undirected weighted graph, where each vertex represents one drug and each edge denotes one interaction between two drugs. After obtaining the SMILES strings of drug molecules as the input features of each node, we can apply SVGAE to obtain the hidden representations of drugs and predict the interactions between them. Particularly, Drugdrug interactions (DDIs) are from DrugComb database [40] and DDIExtraction 2013 dataset [41] and molecular structures of drugs can be obtained freely from DrugBank.
Methods
Our overall framework consists of three steps as shown in Fig. 4: basic protein sequence coding, graphbased feature extraction model, and the final neural network classifier. The first step is to transform raw protein sequences into fixedlength codings in order for subsequent training. Next, we propose an improved weighted variational graph autoencoder (SVGAE) to learn embeddings for each protein based on their sequence features and local graph information, which is equivalent to the feature reduction and extraction. Finally, we use these embeddings as inputs to train a simple feedforward neural network as the final classifier. In this section, we also introduce the datasets and model settings.
Datasets
Benchmark dataset
We used Pan’s [34] dataset from http://www.csbio.sjtu.edu.cn/bioinf/LR_PPI/Data.htmas the benchmark dataset. The positive samples in the dataset are from the human protein reference database (HPRD, 2007 version), with the elimination of the selfinteractions and duplicate interactions. We finally obtained 36591 positive pairs. Based on the common assumption that two proteins in different cellular compartments do not interact, proteins used in constructing negative samples are selected by following the listed criteria: (1) Collecting human proteins annotated with “human” in the ID field only. (2) Excluding sequences annotated with ambiguous or uncertain subcellular localization terms, such as “potential”, “probable”, “probably”, “maybe”, or “by similarity". (3) Including those sequences marked with unique locations only. (4) Excluding sequences annotated with “fragments”, and eliminating sequences with less than 50 amino acid residues as they may only be fragments. (5) Proteins with unusual amino acids such as U and X were removed. The collected proteins were randomly paired with other proteins in different subcellular locations to generate negative samples. Finally, the total amount of the remaining negative samples were 36324.
Other datasets
We also used Guo’s [18] dataset to evaluate our model including: (1) Human dataset containing 9435 proteins with 37020 positive samples and 37027 negative samples. (2) E. coli dataset containing 1834 proteins with 6954 positive samples and 6954 negative samples. (3) Drosophila dataset containing 7059 proteins with 21975 positive samples and 21975 negative samples. (4) C. elegan dataset containing 2640 proteins with 4030 positive samples and 4030 negative samples. The negative samples used for training in each dataset were selected according to the same criteria presented in the above Benchmark Dataset part.
Model settings
Our model was implemented using Tensorflow in Python and took advantage of the strong computing capacity of GPU. All the simulations were carried out on a computer with 4.00GHz 8core CPU and 59GB memory. The GPU we used was NVIDIA GeForce GTX 1080 with 7GB memory. Our source code and datasets are available at https://github.com/fangyangbit/SVGAE.
The SVGAE model has two hidden layers with 96 neurons and 48 neurons of each layer respectively. The final softmax classifier has three hidden layers with 128, 64 and 32 neurons each layer and uses dropout technique during training in order to avoid overfitting. Dropout is a technique that randomly drops units (along with their connections) from the neural network during training [42]. For both parts, we initialize weights as described in [43] and train for 50 iterations using Adam algorithm with a learning rate of 0.005. We tuned the hyperparameters of our model to optimize system performance by conducting 5fold sentencelevel crossvalidation on the training set. To determine the parameter, fourfifths of the whole dataset are randomly chosen to train the classifiers with different number of hidden nodes, while the rest onefifths of the dataset are used as the validation set to compute the accuracy. For Adam optimization, we set the learning rate lr = 0.005 as suggested by Kingma et al. [44]. To alleviate the overfitting problem, the dropout rate was set to 0.5 in our model, as used by Hinton et al. [45].
In our experiments, each dataset was randomly split into 80% training set and 20% test set. The model was trained and validated using 5fold crossvalidation, and the performance of our model was evaluated by the holdout test set. In order to test the robustness of our method, this process of random selection was repeated five times. Therefore, five models were generated based on different training sets and the overall performance was the average of results on five different test sets.
Protein sequence coding
There are several existing methods for protein sequence representation such as auto AC, CT, and LD [46]. In our model, we choose CT as our coding method. The CT method was first proposed by Shen et al. to code single protein [47]. The information of protein sequences can be projected into a homogeneous vector space by counting the frequencies of each triad type. The whole process is described as follows. First of all, all amino acids are clustered into seven categories according to their dipole and side chain volumes. The classification of amino acids is shown in Table 4 and the classification principle is described in detail in the paper [47]. Amino acids within the same group likely involve synonymous mutations due to their analogous characteristics. Then, each amino acid can be substituted by its category label and thus each protein is a string of integers. Next, a window of size three is used to slide across the sequence one step at a time and count the number of occurrences of each triad type. Since we regard any three continuous amino acids as a unit and the amino acids have been catalogued into seven classes, there are 7×7×7 different combinations, that is the size of CT vector is 343. The CT representation is defined as:
where n_{i} is the number of occurrences of each triad type and q equals to 343.
Our SVGAE model
This is the core part of our overall framework. After initial coding of sequences by the CT method, we further conduct representation learning with graph and sequence features using our signed variational graph autoencoder (SVGAE) model. We will discuss this model by problem formulation, and then its inference part and generative part. The inference part is an encoder that encodes original proteins into embeddings while the generative part is a decoder that decodes the embeddings back into original proteins. The purpose of this model is to learn interpretable embedding for each protein by training encoder and decoder at the same time.
Problem formulation
We are given an undirected weighted graph \(\mathcal {G} = (\mathcal {V}, \mathcal {E})\) with \(N =  \mathcal {V} \) nodes, thus N is the number of proteins, and each vertex of \(\mathcal {G}\) represents one protein while each edge is one interaction. The adjacency matrix A of \(\mathcal {G}\) is provided. In matrix A,A_{ij} denotes whether there exists an interaction between protein i and protein j. We enforce selfloops in the graph by simply adding the identity matrix to A. The input features of each node are included in an N×R matrix X, which are sequence representations by the CT method and R equals to 343 in this case. The desired outputs of this model are latent variables z_{i}, summarized in an N×P matrix Z, which will contain the embeddings of proteins we expect to get where P is the dimension of each embedding.
The model is basically an encoderdecoder approach. First the encoder maps each node v_{i} in the graph to a lowdimensional vector embedding, z_{i}, based on the node’s position in the graph, its local neighborhood structure, and its attributes. Next, the decoder extracts the classification label A_{ij} associated with v_{i} and v_{j} (i.e., the label of interaction between protein i and j). By jointly optimizing the encoder and decoder, the model learns to compress information about graph structure into the lowdimensional embedding space. The intuition behind this encoderdecoder idea is that if we can learn to decode higdimensional graph information from encoded lowdimensional embeddings, then, in principle, these learned lowdimensional vectors should contain all information necessary for downstream machine learning tasks, for example, classification.
The encoder
The inference module is a graph convolutional networks (GCNs) encoder [48], which is a function with the goal of a mapping from the original features X to embeddings Z with the augmented information of A. In our current implementation, a simple model parameterized by a twolayer GCN is utilized:
where μ=GCN_{μ}(X,A) is the matrix of mean vectors μ_{i} and logσ=GCN_{σ}(X,A). The GCN model is defined as GCN(X,A)=AReLU(AXW_{0})W_{1}, and W_{i} are parameter matrices we need to train. In our model, GCN_{μ}(X,A) and GCN_{σ}(X,A) share W_{0} in order to reduce parameters. ReLU(·)=max(0,·) is the activation function and \(\mathcal {N}\) is the unit gaussian distribution. The intuition of GCN using GCN(X,A)=ReLU(AXW_{0}) is as follows. The multiplication with adjacency matrix A means that, for every node, we sum up all the feature vectors of all neighboring nodes and itself. In this way, GCN can effectively learn embeddings through integrating neighboring graph features.
The decoder
The generative module we define here is a simple inner product decoder:
where σ(·) is the logistic sigmoid function. We use the inner product of two embeddings z_{i} and z_{j} as the probability of these two proteins existing the interaction. As indicated in Fig. 4, the output of the decoder \(\boldsymbol {\hat {A}}\) is the approximation of adjacency matrix A and we optimize the model so as to make them as close as possible.
Training method and implementation details
In this section, we will detailedly discuss two improvements we proposed on VGAE and explain why they work in our application.
(1) cost function
As the whole interaction network is regarded as an undirected graph and each item in the adjacency matrix represents whether there exists an interaction between the two proteins, in our SVGAE model, we define the cost function as:
where KL[q(·)∥p(·)] is the KullbackLeibler divergence between q(·) and p(·). The first term is to minimize the reconstruction error of the adjacency matrix A. It should be noticed that, we only consider those interactions with high confidence, which we specify as A^{⋆} (A^{⋆}⊂A) to be reconstructed. The second term is to minimize the difference between q(ZX,A) and p(Z). The cost function is the tradeoff between how accurate our model can be and how close the distribution of embeddings can match p(Z). In this case, we assume p(Z) as a Gaussian prior and the reparameterization trick is used for training [49].
As mentioned above, we only consider the cost of interactions with high confidence while ignoring those edges of uncertain confidence. In other words, we assume that those edges of uncertain confidence are random noises to our model and even disturb the training process thus affecting the ultimate performance. Therefore, we need to construct highconfidence sets from original datasets. For positive samples, the confidence are always high since they are actually observed. As for other items in the adjacency matrix A, they are divided into two groups: the highly negative group and the uncertain group. The edges in the highly negative group are negative edges with high confidence, which are selected based on the criteria described in the “Datasets” section.
(2) signed adjacency matrix
During the training process, the adjacency matrix A plays an important role since it not only defines the cost function but also serves as a critical parameter in the GCN. The common adjacency matrix consists of only 0 and 1. However, as we discussed in the last section, different edges actually have different confidence and therefore should have different impacts on the learning process.
In order to embody the differentiated impacts, we assign positive edges positive values (1), the highly negative group negative values (1) and the uncertain group 0. By setting different signs and even different weights, we expect to reinforce existing observed interactions and in the meanwhile, strengthen the negative impact of the highly negative interactions. Detailed comparison of the model with or without signed adjacency matrix in Table S1.
Feedforward neural network classifier
Instead of directly using generative model to infer interactions, we take out embeddings z_{i} contained in the matrix Z and further train a simple neural network as the final classifier. Correspondingly, the inputs to the classifier are concatenations of embeddings of two proteins, while the output label is a binary value representing whether there exist an interaction between the two proteins.
The performance of the classifier can be remarkably good without complex neural network structures since the embeddings already contain enough information and are highly representative in the learned lowdimensional vector space.
Availability of data and materials
The code and datasets during the current study are available in the GitHub repository: https://github.com/fangyangbit/SVGAE.
Abbreviations
 PPIs:

Proteinprotein interactions
 Y2H:

Yeast twohybrid screens
 TAP:

Tandem affinity purification
 MSPCI:

Mass spectrometric protein complex identification
 TAPMS:

Tandem affinity purification and mass spectrometry
 CoIP:

Affinity chromatography and CoImmunoprecipitation
 KNN:

kNearest neighbor
 SVM:

Support vector machine
 SAE:

Stacked autoencoder
 DeepPPI:

Deep neural networks for Protein Protein Interactions prediction
 MDPN:

Multimodal deep polynomial network
 EnsDNN:

Ensemble deep neural networks
 LDAROF:

Latent dirichlet allocationrandom forest
 ELM:

Extreme learning machine
 CSSVM:

Compressive sampling support vector machine
 DFC:

Deep fully connected network
 DLSTM:

Deep long short memory network
 DIP:

Database of interacting protein
 C.elegans:

Caenorhabditis elegans
 E.coli:

Escherichia coli
 AC:

Auto covariance
 CT:

Conjoint triad
 LD:

Local descriptor
 GCN:

Graph convolutional network
 ReLU:

Rectified linear unit
 LDASVM:

Latent dirichlet allocationsupport vector machine
 VGAE:

Variational graph autoencoder
 SVGAE:

Signed variational graph autoencoder
 HPRD:

Human protein reference database
 SMILES:

Simplified molecular input line entry specification
 DDIs:

Drugdrug interactions
References
 1
Berggård T, Linse S, James P. Methods for the detection and analysis of protein–protein interactions. Proteomics. 2007; 7(16):2833–42.
 2
Wang RS, Wang Y, Wu LY, Zhang XS, Chen L. Analysis on multidomain cooperation for predicting proteinprotein interactions. BMC Bioinformatics. 2007; 8(1):391.
 3
Ito T, Chiba T, Ozawa R, Yoshida M, Hattori M, Sakaki Y. A comprehensive twohybrid analysis to explore the yeast protein interactome. Proc Natl Acad Sci. 2001; 98(8):4569–74.
 4
Gavin AC, Bösche M, Krause R, Grandi P, Marzioch M, Bauer A, Schultz J, Rick JM, Michon AM, Cruciat CM, et al.Functional organization of the yeast proteome by systematic analysis of protein complexes. Nature. 2002; 415(6868):141–7.
 5
Ho Y, Gruhler A, Heilbut A, Bader GD, Moore L, Adams SL, Millar A, Taylor P, Bennett K, Boutilier K, et al.Systematic identification of protein complexes in saccharomyces cerevisiae by mass spectrometry. Nature. 2002; 415(6868):180–3.
 6
Huang H, Alvarez S, Nusinow DA. Data on the identification of protein interactors with the evening complex and pch1 in arabidopsis using tandem affinity purification and mass spectrometry (tap–ms). Data Brief. 2016; 8:56–60.
 7
Foltman M, Sanchez Diaz A. Studying protein–protein interactions in budding yeast using coimmunoprecipitation. In: Yeast Cytokinesis. Springer: 2016. p. 239–56. https://doi.org/10.1007/9781493931453_17.
 8
Luo X, You Z, Zhou M, Li S, Leung H, Xia Y, Zhu Q. A highly efficient approach to protein interactome mapping based on collaborative filtering framework. Sci Rep. 2015; 5:7702.
 9
Han JDJ, Dupuy D, Bertin N, Cusick ME, Vidal M. Effect of sampling on topology predictions of proteinprotein interaction networks. Nat Biotechnol. 2005; 23(7):839–44.
 10
Byron O, Vestergaard B. Protein–protein interactions: A suprastructural phenomenon demanding transdisciplinary biophysical approaches. Curr Opin Struct Biol. 2015; 35:76–86.
 11
Deng M, Zhang K, Mehta S, Chen T, Sun F. Prediction of protein function using protein–protein interaction data. J Comput Biol. 2003; 10(6):947–60.
 12
Chen XW, Liu M. Prediction of protein–protein interactions using random decision forest framework. Bioinformatics. 2005; 21(24):4394–400.
 13
Browne F, Wang H, Zheng H, Azuaje F. Supervised statistical and machine learning approaches to inferring pairwise and modulebased protein interaction networks. In: Bioinformatics and Bioengineering, 2007. BIBE 2007. Proceedings of the 7th IEEE International Conference On. IEEE: 2007. p. 1365–9.
 14
Lin X, Chen X. w.Heterogeneous data integration by treeaugmented naïve bayes for protein–protein interactions prediction. Proteomics. 2013; 13(2):261–8.
 15
You ZH, Chan KC, Hu P. Predicting proteinprotein interactions from primary protein sequences using a novel multiscale local feature representation scheme and the random forest. PLoS One. 2015; 10(5):0125811.
 16
Dohkan S, Koike A, Takagi T. Improving the performance of an svmbased method for predicting proteinprotein interactions. In Silico Biol. 2006; 6(6):515–29.
 17
You ZH, Li J, Gao X, He Z, Zhu L, Lei YK, Ji Z. Detecting proteinprotein interactions with a novel matrixbased protein sequence representation and support vector machines. BioMed Res Int. 2015; 2015. https://doi.org/10.1155/2015/867516.
 18
Guo Y, Li M, Pu X, Li G, Guang X, Xiong W, Li J. Pred_ppi: a server for predicting proteinprotein interactions based on sequence data with probability assignment. BMC Res Notes. 2010; 3(1):145.
 19
Cao C, Liu F, Tan H, Song D, Shu W, Li W, Zhou Y, Bo X, Xie Z. Deep learning and its applications in biomedicine. Genomics Proteomics Bioinforma. 2018; 16(1):17–32.
 20
Sun T, Zhou B, Lai L, Pei J. Sequencebased prediction of protein protein interaction using a deeplearning algorithm. BMC Bioinformatics. 2017; 18(1):277.
 21
Du X, Sun S, Hu C, Yao Y, Yan Y, Zhang Y. Deepppi: boosting prediction of protein–protein interactions with deep neural networks. J Chem Inf Model. 2017; 57(6):1499–510.
 22
Lei H, Wen Y, Elazab A, Tan EL, Zhao Y, Lei B. Proteinprotein interactions prediction via multimodal deep polynomial network and regularized extreme learning machine. IEEE J Biomed Health Inf. 2018. https://doi.org/10.1109/jbhi.2018.2845866.
 23
Hashemifar S, Neyshabur B, Khan AA, Xu J. Predicting protein–protein interactions through sequencebased deep learning. Bioinformatics. 2018; 34(17):802–10.
 24
Zhang L, Yu G, Xia D, Wang J. Protein–protein interactions prediction based on ensemble deep neural networks. Neurocomputing. 2019; 324:10–19.
 25
Richoux F, Servantie C, Borès C, Téletchéa S. Comparing two deep learning sequencebased models for proteinprotein interaction prediction. arXiv preprint. 2019. arXiv:1901.06268.
 26
Licamele L, Getoor L. Predicting proteinprotein interactions using relational features. Proc ICML Workshop Stat Netw Anal. 2006;5. https://drum.lib.umd.edu/handle/1903/7555.
 27
Paradesi MS, Caragea D, Hsu WH. Structural prediction of proteinprotein interactions in saccharomyces cerevisiae. In: Bioinformatics and Bioengineering, 2007. BIBE 2007. Proceedings of the 7th IEEE International Conference On. IEEE: 2007. p. 1270–4. https://doi.org/10.1109/bibe.2007.4375729.
 28
You ZH, Lei YK, Gui J, Huang DS, Zhou X. Using manifold embedding for assessing and predicting protein interactions from highthroughput experimental data. Bioinformatics. 2010; 26(21):2744–51.
 29
Hamilton WL, Ying R, Leskovec J. Representation learning on graphs: Methods and applications. arXiv preprint. 2017. arXiv:1709.05584.
 30
Tang J, Qu M, Wang M, Zhang M, Yan J, Mei Q. Line: Largescale information network embedding. In: Proceedings of the 24th International Conference on World Wide Web. International World Wide Web Conferences Steering Committee: 2015. p. 1067–77. https://doi.org/https://doi.org/10.1145/2736277.2741093.
 31
Berg R. v. d., Kipf TN, Welling M. Graph convolutional matrix completion. arXiv preprint. 2017. arXiv:1706.02263.
 32
Kipf TN, Welling M. Variational graph autoencoders. arXiv preprint. 2016. arXiv:1611.07308.
 33
Heylighen F. Occam’s razor. Principia Cybernet Web. 1997. http://pespmc1.vub.ac.be/OCCAMRAZ.html.
 34
Pan XY, Zhang YN, Shen HB. Largescale prediction of human protein protein interactions from amino acid sequence based on latent topic features. J Proteome Res. 2010; 9(10):4992–5001.
 35
Chen M, Ju CJT, Zhou G, Zhang T, Chen X, Chang KW, Zaniolo C, Wang W. Lasagna: Multifaceted proteinprotein interaction prediction based on siamese residual rcnn. bioRxiv. 2018:501791. https://doi.org/10.1093/bioinformatics/btz328.
 36
Zhang YN, Pan XY, Huang Y, Shen HB. Adaptive compressive learning for prediction of protein–protein interactions from primary sequence. J Theor Biol. 2011; 283(1):44–52.
 37
You ZH, Li S, Gao X, Luo X, Ji Z. Largescale proteinprotein interactions detection by integrating big biosensing data with computational model. BioMed Res Int. 2014; 2014. https://doi.org/10.1155/2014/598129.
 38
Vyas R, Bapat S, Jain E, Karthikeyan M, Tambe S, Kulkarni BD. Building and analysis of proteinprotein interactions related to diabetes mellitus using support vector machine, biomedical text mining and network analysis. Comput Biol Chem. 2016; 65:37–44.
 39
Hue M, Riffle M, Vert JP, Noble WS. Largescale prediction of proteinprotein interactions from structures. BMC Bioinformatics. 2010; 11(1):144.
 40
Zagidullin B, Aldahdooh J, Zheng S, Wang W, Wang Y, Saad J, Malyutina A, Jafari M, Tanoli Z, Pessia A, Tang J. Drugcomb: an integrative cancer drug combination data portal. Nucleic Acids Res. 2019; 47:43–51.
 41
SeguraBedmar I, Martínez P, HerreroZazo M. SemEval2013 task 9 : Extraction of drugdrug interactions from biomedical texts (DDIExtraction 2013). In: Second Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2: Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval 2013). Atlanta, Georgia, USA: Association for Computational Linguistics: 2013. p. 341–50. https://www.aclweb.org/anthology/S132056.
 42
Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: A simple way to prevent neural networks from overfitting. J Mach Learn Res. 2014; 15(1):1929–58.
 43
Glorot X, Bengio Y. Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics. Sardinia: PMLR: 2010. p. 249–56.
 44
Kingma D, Ba J. Adam: A method for stochastic optimization. arXiv preprint. 2014. arXiv:1412.6980.
 45
Hinton GE, Srivastava N, Krizhevsky A, Sutskever I, Salakhutdinov RR. Improving neural networks by preventing coadaptation of feature detectors. arXiv preprint. 2012. arXiv:1207.0580.
 46
You ZH, Lei YK, Zhu L, Xia J, Wang B. Prediction of proteinprotein interactions from amino acid sequences with ensemble extreme learning machines and principal component analysis. BMC Bioinformatics. 2013; 14(8):10.
 47
Shen J, Zhang J, Luo X, Zhu W, Yu K, Chen K, Li Y, Jiang H. Predicting protein–protein interactions based only on sequences information. Proc Natl Acad Sci. 2007; 104(11):4337–41.
 48
Kipf TN, Welling M. Semisupervised classification with graph convolutional networks. arXiv preprint. 2016. arXiv:1609.02907.
 49
Kingma DP, Welling M. Autoencoding variational bayes. arXiv preprint. 2013. arXiv:1312.6114.
Acknowledgements
Not applicable.
Funding
This work was supported by National Key Research and Development Program of China (Grant No. 2016YFB1000902) and National Natural Science Foundation of China (Grant Nos. 61976021 and U1811262).
Author information
Affiliations
Contributions
FY, KF and DS conceived of the method. FY and KF designed the method. FY, KF and HL implemented the method and wrote the manuscript. All authors read, edited and approved of the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Additional file 1
Detailed comparison of the model with or without signed adjacency matrix. Table S1. Detailed comparison of the model with or without signed adjacency matrix.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Yang, F., Fan, K., Song, D. et al. Graphbased prediction of Proteinprotein interactions with attributed signed graph embedding. BMC Bioinformatics 21, 323 (2020). https://doi.org/10.1186/s12859020036468
Received:
Accepted:
Published:
Keywords
 Proteinprotein interaction
 Representation learning
 Network embedding
 Variational graph autoencoder