 Methodology article
 Open Access
 Published:
edge2vec: Representation learning using edge semantics for biomedical knowledge discovery
BMC Bioinformatics volume 20, Article number: 306 (2019)
Abstract
Background
Representation learning provides new and powerful graph analytical approaches and tools for the highly valued data science challenge of mining knowledge graphs. Since previous graph analytical methods have mostly focused on homogeneous graphs, an important current challenge is extending this methodology for richly heterogeneous graphs and knowledge domains. The biomedical sciences are such a domain, reflecting the complexity of biology, with entities such as genes, proteins, drugs, diseases, and phenotypes, and relationships such as gene coexpression, biochemical regulation, and biomolecular inhibition or activation. Therefore, the semantics of edges and nodes are critical for representation learning and knowledge discovery in real world biomedical problems.
Results
In this paper, we propose the edge2vec model, which represents graphs considering edge semantics. An edgetype transition matrix is trained by an ExpectationMaximization approach, and a stochastic gradient descent model is employed to learn node embedding on a heterogeneous graph via the trained transition matrix. edge2vec is validated on three biomedical domain tasks: biomedical entity classification, compoundgene bioactivity prediction, and biomedical information retrieval. Results show that by considering edgetypes into node embedding learning in heterogeneous graphs, edge2vec significantly outperforms stateoftheart models on all three tasks.
Conclusions
We propose this method for its added value relative to existing graph analytical methodology, and in the real world context of biomedical knowledge discovery applicability.
Background
Introduction
The knowledge graph (KG) has become the preferred data model for complex knowledge domains. Accordingly Wilcke et al. published: “The knowledge graph as the default data model for learning on heterogeneous knowledge” [1]. Biology and biomedical knowledge is complex and involves a plethora of entity and association types, hence is particularly suited to heterogeneous graph methodology. From such a KG, statistical knowledge can be inferred, for example, probabilistic associations between genes and phenotypic traits. In KG terms, node and edge semantics are varied and critical for precise representation of the knowledge. Methods which consider surrounding node and edge contexts support a rich and combinatorially expanding feature set. KG embedding connotes representation of entities as computable feature vectors amenable to machine learning (ML) methods [2, 3]. As both KG and ML methodology advances, the issues of embedding, representation and vectorization become crucial, as signaled by related research activity spanning computing, natural and social sciences [3]. Deep learning is a powerful approach for representation learning on large graphs and datasets. Multilayer deep neural networks entail transformations from input raw data to layered representations obviating the need for feature engineering up front. Instead a set of continuous, latent features (representations) are learned which, in the graph use case, encode localized structural topology around a given node facilitating prediction tasks based on network structure.
Previous work has focused on using neural network learning models to generate node embeddings for graphs such as DeepWalk [4], LINE [5], and node2vec [6]. However these models were designed for homogeneous networks, which means that they do not explicitly encode information related to the types of nodes and edges in a heterogeneous network. Recently, metapath2vec [7] was proposed by incorporating metapaths with node semantics for node embedding learning. However, this approach has several drawbacks: 1) domain knowledge is required to define metapaths and those mentioned in [7] are symmetric paths which are unrealistic in many applications; 2) metapath2vec does not consider edge types rather only node types; and 3) metapath2vec can only consider one metapath at one time to generate random walk, it cannot consider all the metapaths at the same time during random walk. On another related track, which might be termed biomedical data science (BMDS), previous work has employed KG embedding and ML methodology with the focus on applicability and applications such as compound target bioactivity [8, 9] and diseaseassociated gene prioritization [10]. Yet other efforts have simply employed offtheshelf ML toolkits (e.g. Scikitlearn, WEKA) and methods to address biomedical informatics prediction challenges.
To address the above problems, edge2vec was developed to consider edge semantics when generating node sequence using a random walk strategy. An edgetype transition matrix is defined to improve representation of node “context” and designed with an ExpectationMaximization (EM) model. In the maximization step, we use the transition matrix to generate node sequences based on random walk in a heterogeneous graph. In the expectation step, we use the generated node ’context’ from node embeddings as feedback to optimize the transition matrix. We also use a skipgram sampling strategy to select partial nodes for the EM approach to make the edge2vec model run on largescale networks to learn node embeddings in a more efficient way. In the end, the topologically similar nodes (with similar substructures or located near each other in the network) are with similar emebeddings; the semantically similar nodes (with same nodetypes or logistically related attributes) are with similar embeddings.
Within biomedicine, the sciences involved in drug discovery are diverse. Drug efficacy and safety depend on calibrated modulation of complex, interrelated biomolecular pathways and targets. Prediction of compoundtarget bioactivity, normally noncovalent binding, remains highchallenge and highvalue, both for generating novel drug leads and hypotheses, and for elucidating the mechanism of action for known compounds and drugs. With this rich knowledge domain as context, in this paper, we apply edge2vec on Chem2Bio2RDF [11], a highly heterogeneous graph integrating over 25 biomedical and drug discovery datasets.
The contribution of our work is threefold.

We define an edgetype transition matrix to represent network heterogeneity. The calculation of the matrix is mainly based on the path similarity of different edgetypes.

We develop an EM model to train a transition matrix via random walks on a heterogeneous graph as a unified framework and employ a stochastic gradient descent (SGD) method to learn node embedding in an efficient manner. The learned node vector can include not only the topological information of network structure, but also the edge type information, which indicates different relationships among nodes.

We evaluate our model in the drug discovery domain by predicting drugtarget associations using the highest available quality datasets as ground truth. Validation of the edge2vec model is addressed via three prediction tasks, all realistic biomedical discovery use cases. Validation results indicate that edge2vec adds value relative to existing methodology for drug discovery knowledge discovery.
In the following sections, first, we introduces edge2vec and its importance; second, we discusses related work about node embedding learning as well as heterogeneous network analysis; third, we explains edge2vec; fourth, we evaluates edge2vec based on later drug discovery; fifth, we illustrates two case studies to visualize edge2vec results, And in the end we concludes and points out future work.
Related Work
Network Representation: Network representation is useful in a variety of applications such as network classification [12, 13], content recommendation [14–16], community detection [17–19] and link prediction [20]. Networks are easily and naturally represented by adjacency matrix, but such matrices are generally sparse and high dimension, thus not well suited to statistical learning [4]. How to represent network information in low dimension is an important task. There are classical methods of network representation which is dimension reduction based on calculating eigenvector, such as LLE [21, 22], Laplacian Eigenmap [23, 24], MDS [25], IsoMap [26], and DGE [27]. However, these methods do not perform well in largescale networks.
Representation Learning based on Deep Neural Network: In deep learning, more and more encoderdecoder models have been proposed to solve network representation problems. By optimizing a deterministic distance measure, those models can learn a node embedding from its neighbor nodes so as to project nodes into a latent space with a predefined dimensionality.
Recently, deep neural network [28] based representation learning has been widely used in the natural language processing. Word2vec [29] is the deep learning model developed by Google to represent a word in a low dimension dense vector, which has proven to be successful in natural language processing [30]. By close analogy, topological paths neighboring a node may be handled like sequences of words, and word2vec can be adapted to network representation learning to reduce computing complexity and improve performance relative to conventional approaches. Accordingly, several recent publications have proposed word2vecbased network representation learning frameworks, such as DeepWalk [4], GraRep [31], TADW [31], CNRL [32], LINE [5], node2vec [6], and metapath2vec [7]. All of the above frameworks utilize the skipgram model [33, 34] to learn a representation of a node incorporating its topological context, so nodes with similar topological information will have similar numerical representations. Node representations are learned via skipgram model by optimizing the likelihood objective using SGD with negative sampling [35].
Sampling Strategy: Similar to word sequences from documents, node sequences may be sampled from the underlying network as an ordered sequence of nodes [7]. Accordingly, different network representation learning frameworks adopt different node sampling strategies. DeepWalk [4] deploys a truncated random walk to sample node sequences, and uses the skipgram model to learn the representation of node sequences. However, DeepWalk only considers the firstorder proximity between nodes. Moreover, it applies to unweighted networks. Practically, LINE is applicable for both weighted and unweighted networks and easily scales to largescale networks with millions of nodes. The problem is that embedding of some looselyconnected nodes, which have few connected edges, heavily depends on their connected neighbors and unconnected negative samples [36]. Most prior methods do not give full consideration to heterogeneity of nodes and edges. Thus Deepwalk, LINE, and Node2vec are not effective for representing these heterogeneous networks. Sun et al. [37] introduced a metapathbased similarity measurement to find similar objects of the heterogeneous information networks. Furthermore, Dong et al. proposed metapath2vec [7] to capture heterogeneous structure and semantic correlation exhibited from largescale networks by considering node types. However, one drawback of all previous methods is that they either only deal with homogeneous networks or do not consider edge semantics. When network contains nodes and edges with different types, the stateoftheart embedding results are no longer effective as all of them do not consider edge semantics. To represent heterogeneity, we have developed edge2vec to learn node representations with general, systematic consideration of edge semantics.
Representation learning in biomedical domains: In biomedical domains, there exist rich heterogeneous datasets about genes, proteins, genetic variations, chemical compounds, diseases, and drugs. Ongoing and expanding efforts to integrate and harness these datasets for datadriven discovery reflect widespread understandings of potential benefits to science and human health. For example, Chem2Bio2RDF [11] integrates over 25 different datasets related to drug discovery and comprises a large scale heterogeneous network. Such repositories hold complex relationships between many entity types. Representing this semantic complexity requires suitable embedding methods informed by these rich domains. Chen et al. [8] propose Semantic Link Association Prediction (SLAP) to predict ’missing links’ between drugs and targets in Chem2Bio2RDF. Subsequently, Fu et al. [38] applied an improved PathSim method with more than 50 metapaths on an extended Chem2Bio2RDF dataset to predict drug target interactions and rank metapaths based on Gini index. Although these previous studies focus on relationships between drugs and targets, none handle edge heterogeneity in graph embedding directly and generally. We regard this as an unmet need, both for the methodological value in exploring edge semantics, and also for the more practical value of applying and validating this novel methodology via biomedical data science use cases. Thus motivated, we propose edge2vec as an improved representation learning model, well suited for discovery on biomedical knowledge graphs.
Context: applied machine learning and data science: Machine learning is a big, diverse and rapidly advancing area of research, challenging to monitor and contextualize even to its scholars and practitioners. For biomedical data science, the challenge is compounded with applicability to complex real world datasets and tasks. This is applied machine learning, wherein the final evaluation depends on relevance and comprehensibility to the application domain. In this paper we strive to maintain this relevance and comprehensibility through concise, contextualizing notes and examples, with appropriate citations for further study. Terminology is a related challenge, for example the equivalence of “graph” and “network”, so to assist we provide a glossary as supplementary material.
Methods
In this section, we introduce edge2vec. The pipeline is shown in Fig. 1. We treat heterogeneous network embedding learning as an optimization problem and design an EM framework associated with a skip gram model to solve it. See Algorithm 1 pseudo code for details.
Edgetype transition matrix for network embedding
As word2vec [29] informed node2vec [6], we can represent a node and its network neighborhood analogous to a wordcontext relationship in a text corpus. Random walk paths of nodes are akin to word sequences. We thereby convert the node embedding learning problem into a node neighborhood optimization problem: given a node, we need to maximize the probability of neighbor nodes, which is Formula 1:
where V refers to the node collection of the network G(V,E); N(v) refers to the neighbor node collection of node v; θ is the node embedding parameterization to be learned.
However, this optimization only works well in homogeneous networks. As in heterogeneous networks, different types of nodes and edges occur with varying frequency. But low frequency node and edge types may be very important, depending on their semantics in the knowledge domain. For instance, in a scholarly citation network, venue nodes (i.e., conferences and journals) are fewer but more important than publication nodes. Since node2vec would treat all nodes equally, knowledge contained in the venue relationships would be lost. Likewise, throughout biomedical domains, node and edge semantics must be considered to avoid loss of critical knowledge. For one example, the edge relationship between an approved drug and its well validated protein target is highly and exceptionally informative, reflecting prodigious research efforts and expense. To address this need for edge semantics, we design an edgetype transition matrix which holds the transition weights between different edge types during the random walk process. Therefore, we consider not only the topological structure of the network but also edge semantics. Accordingly, the optimized version is shown in Formula 2:
M refers to the edgetype transition matrix. The matrix stores the random walk transition weights between different edge types. By employing the transition matrix as a prior distribution guiding the random walk process, we not only consider the distance between the nextstep node and the previousstep node but also the weight between the nextstep traversed edge type and the previousstep traversed edge type. Therefore, we can normalize by type so that the effect of low frequency node/edge types won’t be lost by dilution among high frequency node/edge types. As shown above, the optimization function maximizes the probability of generating the node neighborhood of a given node v, thus the transition probability from the current node v to its neighbor c can be seen in Formula 3:
where \(\vec {f_{i}}\) means the current step embedding for node i which will be updated in each batch. We calculate the inner product of two node embeddings, which are normalized by a Softmax function.
We designed an EM framework to combine the update of the transition matrix M and optimization of node context into a unified framework. An edgetype transition matrix is initialized with all values set to 1, meaning initially, all edge type transitions are regarded as equally probable. Then, we iteratively generate the random walk corpus of paths, optimizing the transition matrix based on the sampled frequencies of edge type transitions.
ExpectationMaximization framework
Expectation step
Assume we have a set of E= {e_{1},e_{2},e_{3}…e_{m}} different edge types in a network. From the previous iteration in the EM framework, we can get a collection of random walk paths for each node as P = {p_{1},p_{2},…p_{n}}. In each walk path p_{i}(i∈{1,2…n}), it is constructed like p_{i}={n_{1},n_{2},n_{3},...,n_{l}} where n_{i} is the ith node in p_{i} and l is a predefined walk length. Based on each path, we first extract all edges {T(n_{1},n_{2}),T(n_{2},n_{3}),…,T(n_{l−1},n_{l})} in the path by locating every start node n_{k} and end node n_{k+1} where k ∈{1,2,...,l−1},e_{k}=T(n_{i},n_{j}) refers to the edge type between n_{i} and n_{j}. After that, we calculate the number of times each type of edge e_{j}(e_{j}∈E) appears in the walk path p_{i}. The same calculation is applied to all walk paths. In the end, for each edge type e_{j}, we get a vector representation v_{j}, where the ith dimension in the v_{j} refers to the number of times e_{j} appears in walk path p_{i}. One assumption of our model is for a pair of edge type e_{1} and e_{2}, the distribution of each edge type sampled from the random walk paths is a valid estimator for the transition correlation for the graph. Hence, by calculating the correlation between their associated vector v_{i} and v_{j} in the walks, we can regard the correlation score as their updated transition weight. Therefore, we can define the formula for updating transition matrix as Formula 4:
where E[·] is the expectation value and σ is related standard derivation value. M (e_{i}, e_{j}) refers to the updated transition weight between edge type i and j. v_{i} and v_{j} are vector representation of e_{i} and e_{j} on all walk paths. By using Pearson correlation analysis, we can get a pairwise correlation score between two edge types to check the distribution difference. Larger weight value means larger correlation between the pair of edge types. However, as the range of the correlation score varies from 1 to +1, it makes no sense if we keep the original negative weights between a pair of edge types. Because we involve the optimized transition weights to the random walk probability, and the probability can’t be negative, thus we normalize by transformation to a Sigmoid function to solve this issue and restrict the transition probability in a range of between 0 and 1. Moreover this nonlinear transformation can better help to capture the patterns of transition probability than other linear or quadratic transformation functions [39]. The definition of Sigmoid (·) is shown as Formula 5:
In summary, the nonlinear transformed correlation ensures three characteristics of the biased random walk on a heterogeneous network: First, a random walk tends to pass on edges with same edgetype. Based on the correlation calculation in Formula 4, given an edgetype, the correlation with itself is always +1, which is the highest transition weight inside the transition matrix. Second, the Sigmoid function guarantees optimization convergence of transition matrix. Transition weights are adjusted according to the Sigmoid function by training based on the correlations calculated from the random walks until a stable final value is reached. Third, only edgetypes with closer relationships tend to have higher transition weights. Although some edgetypes are globally more common and likely to appear in random walks, we consider the specific cooccurrence rates between edgetypes in the same random walk. For example, if edgetype a appears (20,30,40) times in three random walks, while edgetype b appears (1,1,1) times in the same randoms walks. The transition weight from b to a is still considerably low.
Maximization step
In each iteration in the EM framework, based on the updated edgetype transition matrix M in the expectation step and the network topological structure, the biased random walk process generates a new paths with information of nodes and node neighbors. The transition matrix contributes to the calculation of random walk probabilities, thereby including the influence of edgetype information in sampling, which can reduce the negative effects caused by skewed type distribution issues. Even though some types of edges appear less frequently in the network, if the transition weights between those edgetypes and other edgetypes are high, the edge still has a high probability to get visited during the random walk process. Another important feature is that based on the expectation step, as well as Formula 4, for an edgetype e, M_{e,e} is always the largest among all possible edgetype pairs toward e, which means random walk prefers to keep the same kind of edgetype. So, during the random walk process, given the current node v and the previous node u, the probability for the next candidate node n is calculated as Formula 6 and demonstrated in Fig. 2:
where T(v,u) refers to the edgetype between node v and node u. α_{pq}(k,u) is defined based on the distance d_{ku} between next step node candidate k and previous traversed node u. The distance function is defined as Formula 7:
As seen in Algorithm 1, at the beginning, we initialize walk paths as empty, all values in the transition matrix as 1, we use function GenerateTransitionMatrix(·) to utilize an EM framework to get walk paths and the matrix M. In maximization steps, the function takes transition matrix in the last iteration as input, invokes the HeteroRandomWalk(·) function to get walk paths, the probability of random walk is mainly based on Formula 6. In expectation steps, the function utilizes the updated walk paths to optimize the transition matrix by Formula 4. We can retrieve an optimized edgetype transition matrix, which holds the correlation between edgetypes, via the EM framework. At the same time, we can also get the random walks as a node “corpus”, which holds the correlation between nodes. We therefore represent the whole heterogeneous network as a collection of random walk paths, which can be used as the input of the next step for embedding optimization.
Skip gram for embedding optimization
With the help of the EM framework, we can get the transition matrix M and random walks w as the input layer to train the node embedding via a one layer neural network. To optimize the Formula 2, we use the stochastic gradient descent (SGD) method to get optimized node embeddings. Considering all nodes to maximize Formula 2 would be slow and computationally inefficient. Hence, in addition to the known neighbor node t, we use the negative sampling method to generate k negative nodes towards a given node v. And the K negative nodes u_{i} where i∈{1,2,...,k} are randomly sampled from the uniformed distribution D(t) with probability P(t). Moreover, we take logarithm on Formula 2 to reduce calculation complexity. And the final objective function turns to be Formula 8 in the end:
The goal of the objective function is to maximize the similarity with the positive neighbour node and minimize the similarity with negative neighbor nodes.
Results
In this section, we describe the biomedical dataset used to test edge2vec and demonstrate the advantage of our model in three evaluation tasks. Moreover, we have a separate section for parameter tuning to retrieve the best model in both efficacy and efficiency points of view.
Biomedical dataset: Chem2Bio2RDF
Chem2Bio2RDF [11] is a richly heterogeneous dataset integrating data from multiple public sources spanning biomedical subdomains including bioinformatics, cheminformatics and chemical biology. The dataset includes 10 node types and 12 edge types. For details of each node/edgetype description, please refer to Table 1. In total, there are 295,911 nodes and 727,997 edges, a relatively sparsely connected network. There exist multiple edge types between two given node types, for example, two edge types between node types “gene” and “compound.” Node and edge type distributions are highly skewed. For instance, there are more than 20,000 compound nodes but a relative few are well studied in biological experiments, such as approved drugs, while most have few high confidence biological associations. Overall, the heterogeneity comprised by these network characteristics present significant challenges for embedding learning, and moreover, the particulars and specific semantics of this biomedical knowledge graph are essential considerations in optimizing learning power. Figure 3 shows the whole network structure of Chem2Bio2RDF.
Given the proposed edg2vec, we set up parameters with p = q = 0.25; embedding dimension d = 128; for other parameters we use the defaults from node2vec. After those parameters are assigned, we use Chem2BioRDF to train our edge2vec model. To evaluate the fitness of the generated node embeddings, we propose three evaluation tasks in the following three sections.
Evaluation Metrics
In this paper, we evaluate our model from both classification and information retrieval viewpoints.
Precision, recall, F1 score and Hamming loss are four metrics reported in classification tasks. Precision implies the ratio of correct positive results returned by the classifier; recall implies the ratio of correct positive results are returned; F1 score balances both precision and recall by taking their harmonic average. All above three metrics are in a range of 0 and 1, the higher the better. While the Hamming loss is the fraction of labels that are incorrectly predicted. The score is also in a range of 0 and 1, but the lower the better.
Precision@K, recall@K, MAP, NDCG and reciprocal rank are five metrics reported in information retrieval related tasks. Precision@K and recall@K imply the precision and recall score in the Top K ranked results. MAP refers to “mean average precision”, which implies the average precision score for all searching queries. NDCG refers to “normalized discounted cumulative gain”, which is a metric to measure not only the accuracy of searching results but also the ranked position of correct results. Like NDCG, reciprocal rank also considers the correct results ranking positions in the returned ranking list. It is the multiplicative inverse of the rank of the first correct result among all searching queries.
Entity multiclassification
We first propose a node multiclassification task. In this task, we take the types of nodes away so the network only has nodes, edges, and edgetypes. We run edge2vec and cluster nodes based on the result of edge2vec to see whether nodes with similar types will be clustered together. In the Chem2BioRDF dataset, there are 10 different node types with different scale number. In order to build up a suitable dataset for the classification model, for each node type, we randomly sample equal number of nodes from the dataset. In this way, we have a natural baseline as precision = 0.1 for a random classifier. Each node is represented as an instance; the 128 dimension vectors are regarded as 128 different features. Its related node type is the response variable. We use a linear support vector machine as the classification model to predict the node’s labels, and use a 10fold validation to evaluate the returned metrics. Three network embedding methods including DeepWalk, LINE and node2vec are our baseline algorithms. For node2vec, we take p = q = 0.25 which is the same setting as edge2vec. Other settings for all three algorithms are just default settings according to their related publications. For each node, after we learn its node embeddings for all baselines, we concatenate the embedding with the number of edges it has for each edgetype to integrate edgetype information into all baseline models as well. For example, if there are four edgetypes in a network and a node has one edge with type 1, two edges with type 2, three edges with type 3 and zero edge with type 4, we concatenate an additional four dimensional vector (1,2,3,0) to the original learned embedding. As metapath2vec requires metapath definitions (manually curated) and thereby only uses selected metapathpattern matched nodes for training node embeddings, metapath2vec is not comparable with other algorithms for a multiclassification task, which is also a drawback of metapath2vec.
We use precision, recall, F1 score macro, and Hamming loss as four evaluation metrics. These are all commonly used evaluation metrics particularly for classification problem. Precision is the fraction of relevant instances among the retrieved instances, while recall is the fraction of relevant instances that have been retrieved over the total amount of relevant instances. F1 measure is the harmonic average of the precision and recall, which balances the two metrics. Hamming loss is the fraction of labels that are incorrectly predicted. Details of the evaluation results can be seen in Table 2. To verify our model’s superiority, we run our model five times and calculate the performance differences between our model and the best baseline on each metric for all the runs, and apply a Ttest to check whether the performance difference is significantly above 0 or not.
From the evaluation results, we can find all four algorithms can predict node types far better than a random classifier. It means even we treat this heterogeneous network as a homogeneous one, there is still some meaningful information stored in these node embeddings. DeepWalk and node2vec have similar results which is no wonder because DeepWalk can be regarded as a particular node2vec model when p = q = 1. While LINE performs the best among all three baselines. It means for this medical network, local structure (one step neighbours and two step neighbours contains most information of a node). However, our proposed edge2vec model outperforms all baseline algorithms,. In all four evaluation metrics, our model has at least 20% improvement in each evaluation metric. It reflects that our model can better predict node labels via its node embedding. Moreover, in all steps of edge2vec, we only use edgetype information during the random walk to generate edgetype transition metrics, and no node type information. Therefore, we can rigorously validate model performance in node type prediction.
Compoundgene bioactivity prediction
One highvalue biomedical challenge is to predict bioactivity between a compound and protein target (often referenced by the encoding gene). Such predictions can accelerate early stage drug discovery by informing and/or replacing expensive screening campaigns via virtual screening. Therefore, we considered the real world bioactivity prediction use case for a validation task. Besides the three baselines in the first task, we add metapath2vec to our baseline as well. As metapath2vec needs to define metapaths a priori, three metapaths, which are compoundgenecompound; compound genegenecompound; compounddruggenedrugcompound were selected as the metapaths for our baseline metapath2vec. As metapaths need to be symmetric, we have to take the network as undirected when training metapath2vec node embeddings. Our ground truth is from another work [38] in which the authors generated 600,000 negative compound gene pairs and 145,6222 positive pairs. These ground truth pairs do not exist in Chem2BioRDF so it can be used as ground truth to evaluate of the result of edge2vec for bioactivity prediction. As the label for a compoundgene pair in ground truth is either ’positive’ or ’negative’, the prediction task is a binary classification task. Here, we randomly select 2,000 positive pairs and 2,000 negative pairs from the ground truth. And a random classifier will have an accuracy value as 0.5 naturally. Similar to the approach in the multiclass classification task, for each compoundgene pair, we use the difference of both embeddings together to form a new 128dimension embedding to represent the pair, and we apply a logistic regression classifier to train a prediction model. Each dimension of the pair is also regarded as a feature. The relationship between the compound and gene is a binary label for each pair. In the training and testing process, If the prediction score is above 0.5, we label the pair as “positive”, otherwise as “negative”. We deploy the same evaluation metrics as the multiclass classification task plus area under an ROC curve (AUROC). The detailed result is shown in Table 3. To verify our model’s superiority, we run our model five times and calculate the performance differences between our model and the best baseline on each metric for all the runs, and apply a Ttest to check whether the performance difference is significantly above 0 or not. In Fig. 4, we also report the ROC curve for edge2vec and baseline models based on their prediction scores, where we can find our model curve significantly performs better than the baselines.
Some interesting findings are observed from the experiments. First of all, among all three well known baseline algorithms (DeepWalk, LINE and node2vec), LINE still outperforms the other two baselines. And the result of DeepWalk is similar to that of node2vec. So, edge2vec is reliable and functionally stable for the two tasks. For metapath2vec, we leverage both metapath2vec and metapath2vec++ in our baseline models. As metapath2vec relies too much on selected metapaths, none of the three metapaths performs well.
Among these three metapaths, we find metapath compounddruggenedrugcompound works the best, implying that prediction accuracy is improved by the additional node types. Although the number of drug nodes is trivial compared with the number of compound nodes and gene nodes, drug nodes have larger effects than compounds and genes in terms of bioactivity prediction. So it is necessary to treat different types of nodes separately within an embedding model. Compared with metapath2vec, metapath2vec++ however achieves worse result in all three metapaths. edge2vec outperforms all baseline models. The F1 measure is around 0.9 which is far better than the random classifier with a score of 0.5. Also it has around 10% improvement compared to the LINE result which is the best of baseline results.
Compoundgene search ranking
Bioactivity prediction as a binary classification task, like single point high throughput screening in the wet lab, predicts active or inactive only. This is helpful, but more useful is the capability to predict ranking of hits by a measure that increases the probability of success and overall efficiency in costly follow up efforts. Hence, this comprises our final evaluation task: compoundgene search ranking. By analogy, the number of hits returned by a search algorithm is generally less important than the ranking and particularly the top ranked hits. Thus, our final task can be described as an information retrieval or search efficiency task. To limit runtime cost, from the ground truth, we select 70 compounds, which contain more than one positive pair with a gene. For each compound, we calculate the top 100 similar nodes in Chem2BioRDF. Compared with the positive pairs of 70 compounds in ground truth, we evaluate the searching result using metrics such as precision, recall, MAP, NDCG, and mean reciprocal rank (MRR). These metrics care not only whether the bioactive genes are returned but also the ranking of the returned genes. For a compound node, if its bioactive genes shows up with a higher rank in the top 100 returned ranking list, the evaluation metrics will have larger values. After the bioactivity prediction task, we choose the best metapath among all three metapaths, which is compounddruggenedrugcompound. We evaluate the embedding results from LINE, node2vec, and edge2vec, as well as metapath2vec with the best metapath. Evaluation details are shown in Table 4. To verify our model’s superiority, we run our model five times and calculate the performance differences between our model and the best baseline on each metric for all the runs, and apply a Ttest to check whether the performance difference is significantly above 0 or not. From the evaluation table, we find DeepWalk and node2vec still have similar results, and both outperform LINE. metapath2vec is almost meaningless because all returned evaluation metrics are approaching to 0, which means it can barely retrieve future positive genes to compounds. Metapaht2vec++ performs slightly better than its previous performance in other two tasks and is comparable to LINE. And overall, node2vec works the best in all baseline algorithms. Compared with all baseline algorithms, our edge2vec outperforms all the baseline methods. Although the retrieved scores are all relatively small, there is around 10% improvement in precision and a little better in the rest of evaluation metrics at least. This, edge2vec adds value in this critical task of compoundgene ranking, which can improve cost efficiency in virtual screening follow up efforts of early stage drug discovery.
Parameter tuning
In our EM framework for edge transition matrix training, in order to get the best fit to retrieve transition relationships between edge types, we have tried various correlation methods including cosine, Spearman, Pearson and Wilcoxon signedrank. To standardize the returned correlation scores into a reasonable (above 0) and comparable range, we have tried various activation function such as ReLU, Sigmoid and traditional standardization methods. Our experiments show that using the combination Sigmoid activation function and Pearson correlation similarity performs the best to represent the transition relationship between edges. Hence, we decided to use this combination for building up our framework. During the transition matrix training process, there are four important parameters to be tuned. We list them with default value below.

1
Number of walks on per node, r = 1

2
Walk length in each random walk path, w = 50

3
The ratio of nodes sampled for training edge transition matrix, p = 0.01

4
The number of iterations for training edge transition matrix, N = 10
The default parameter settings are used to train our edge2vec model and compare with baseline models in previous sections. In this section, we vary each of them and fix the rest to examine the parameter sensitivity of our model. We leverage all generated results on solving node multiclass classification task and use the Macro F1 score as the judgment to evaluate related models. The result of our tuning process is shown in Fig. 5.
To test how much that numbers of walk per node can affect our model, we test five cases r = {1,10,20,30,40} and the result is in Fig. 5a. We can see that more numbers of walks on per node leads to an increase in Macro F1 score. The reason might be that more walks on a node can better reflect the edge relationships around the node and avoid the negative influence of walk randomness. However, even though it shows a positive trend, the increase is small. So a short number of walks on per node should be able to capture enough edge relation information around the node.
In Fig. 5b, with the increase of walk length, the Macro F1 score increases in the beginning and decreases later on. In general, as the random walk length increasing, it will tend to contain all types of edges. As we don’t consider the sequence of the edge types in the same walk, increasing walk length can add noise obfuscating edge type transition relationships.
Figure 5c shows the different ratio of nodes involved in the EM training process to generate edge transition matrix. It might be no need to involve all nodes when training the edge transition matrix as long as all edge types can be covered in random walks in each iteration. Although with more nodes involved, the overall trend of Macro F1 score has a positive sign, the increase of Macro F1 score is not huge and the trend even fluctuates a little bit. We thereby conclude that using a sampling of nodes to train the edge type transition matrix is sufficient and computationally efficient.
Figure 5d refers to the influence of number of iterations in the EM process for edge type transition matrix training. We believe the reason why when N = 1 the Macro F1 score outperforms than N = 10 is by chance as when we increase the number of iterations, the overall trend of Macro F1 score also increases. From the Fig. 5d, the convergence is fast, and a few iterations can already generate a satisfactory Macro F1 score.
Discussion
To discuss how we can apply our node2vec model on a biomedical data set, we conduct two case studies to show the practical values of our model. The first one is to rank the positive links between nodes, which can be used for similarity search and recommendation; the second one is to cluster and visualize similar gene nodes that belong to the same gene family.
Ranking positive bindings for similarity search
To verify how well our model can be used for similarity search and recommendation use cases, we carried out a ranking experiments using the links identified in the existing network. We randomly selected three widely used drugs from the network, which are Hyaluronic acid (DB08818), Calcium Phosphate (DB11348), Copper (DB09130), and Cupric Chloride (DB09131). Each of them has multiple target genes to interact with. The selected pairs of drugs and target genes exist in the network, and we want to reproduce the links using the cosine similarity score based on the embedding vectors. As we can see, our proposed edge2vec embedding can represent node similarity significantly better than the node2vec embedding. The cosine similarity scores for the drug targets of calcium phosphate were all above 0.6, indicating strong similarity between the drug and the target genes. However, using node2vec embedding the cosine similarity scores between calcium phosphate and its targets were all below 0.4, and some of them demonstrated strong dissimilarity like CADPS2 as a target gene of calcium phosphate. The same findings for the other three drugs and their target genes. In addition, all of the target genes for those drugs can be identified as similar nodes with high rankings using edge2vec embeddings. Details are shown in Table 5. we further performed a pairwise ttest [40] study to see whether the similarity scores generated by two models are significantly different or not. If edge2vec has significantly higher similarity score than node2vec, it means our model can better predict those positive drugtarget pairs in the network. In our result, the difference between two embedding approaches is 0.0103 with a pvalue of 0.0001. It means our embedding approach can better rank and retrieve the existing links in the network than node2vec.
Gene clustering analysis
In order to further prove the usefulness of our node embedding results, we carried out a clustering analysis of gene nodes. We arbitrarily selected five gene families, which are ATP binding cassette transporters (ABC), Gap junction proteins (GJ), DEADbox helicases (DDX), Cytochrome C Oxidase Subunits (COX), and Pseudoautosomal region 2 (PAR2). Each gene family refers to a collection of genes originated from the same root gene and performing similar biochemical functions. From each given gene family, five genes are randomly selected to perform clustering analysis. Then, we use principal component analysis (PCA) to project the default 128 dimensions of all gene embeddings into a 2D space. Ideally, genes belonging to the same family should locate closer to each other than genes from different families. The resulting 2D plot has shown that the twenty five genes in five gene families have been well clustered based on their node embedding vectors. Results can be visualized in Fig. 6, where each pentagon refers to a gene and different colors indicate different gene families. It is easy to observe that genes are clustered by family in the 2D space. In the 2D projection, genes in family COX are all located in the bottom right corner of the plot, and genes in family GJ tend to stay in the bottom left corner. In a word, our edge2vec model can generate node embeddings highly reflecting their family information.
Conclusions
In this paper, we propose edge2vec, which incorporates edge semantics to add value over previous methods, as evaluated by knowledge discovery tasks in the domain of biomedical informatics. Edge2vec employs an EM framework associated with a onelayer neural network, to learn node embeddings which perform better than previous methods for heterogeneous networks. The novelty of our work is to generate an edgetype transition matrix so that during the process to generate the node random walk corpus, heterogeneity of the network is also considered. It can reduce the skewed type distribution issue via weighted sampling. Moreover, compared with other stateofart heterogeneous network embedding methods such as metapath2vec, our edge2vec has no restrictions and can deal with the situation where there are multiple relationships between two types of nodes. To illustrate efficiency and accuracy of our proposed model, we evaluate it on biomedical dataset Chem2BioRDF and propose three evaluation tasks including node multiclass classification, link prediction, and search rank efficiency. Edge2vec outperforms all baseline algorithms significantly. Furthermore, we illustrate the effect of edge2vec in biomedical domains using two case studies to explain the biological meanings of the prediction. Moreover, edge2vec can work well in both undirected and unweighted networks, and computational cost is only moderately increased relative to node2vec by choice of appropriate random walk strategy in the EM framework.
There are certainly promising future directions to be explored, which can be subdivided into (1) algorithmic modifications applicable to heterogeneous graphs generally,and (2) domain knowledge based enhancements applicable to characteristics of biomedical datasets and use cases. Informed by metapath2vec, we could change the objective function by using a nodetype based negative sampling. Instead of random sampling from all types of nodes, we could sample negative nodes based on the ratio of each node type. Another opportunity for improvement involves adding domain knowledge into our existing model. During the random walk generation process, we have already considered both node distance (p,q) and edgetype (transition matrix M). In addition to these two attributes, we could add some predefined rules guiding random walks based on our domain knowledge. Another problem for such complex networks is the scale free issue, i.e. skewed degree distributions, where a relative few hub nodes account for the bulk of connections. To reduce this hub node effect in the network, we can also come up with new methods such as novel restriction rules in the random walk, or mitigate those effects by adding node degree related information to optimize a new objective function. For the node attribute prediction task, we can actually apply a semisupervised approach: Given a sampling of nodes with known attributes, we can add this information into our embedding learning process and predict the attributes of remaining nodes. Or we can apply an unsupervised approach: given the relationship between the target attribute with other known attributes, we use this relationship information and other known attributes information in our embedding learning process, and in the end directly predict node target attributes. These enhancements could extend our edge2vec model to better address specific downstream tasks such as node attribute prediction with unsupervised or semisupervised training strategies.
Availability of data and materials
The dataset used in this work can be found at Github ^{Footnote 1}. There are three files within the zipped folder: chem2bio2rdf.txt is the heterogeneous graph we used to train our model. In this file, each line is a RDF triplet which contains two entities and their relations. Entity type and relation type can be obtained from their RDF representations directly. negative.txt and positive.txt stored the negative and positive bindings between genes and compounds, which are all directly generated from the original chem2bio2rdf paper.
The dataset is processed from the original dataset published at BMC Bioinformatics ^{Footnote 2}. We use the chem2bio2rdf.txt, internal_testset_label/positive.txt and internal_testset_label/negative.txt from semantic_network_dataset.zip directly.
The source code is stored at at Github ^{Footnote 3}. For details to run the code, please refer to the instructions in the Github link. In order to load the data into edge2vec, please convert all RDF format data points to hashed ids first, and save in csv format. Then run transition.py to generate and store the transition matrix. Then, transition.py is utilized to load the data and transition matrix for embedding optimization.
For evaluation, please use negative.txt and positive.txt as ground truth files.
References
 1
Wilcke X, Bloem P, de Boer V. The knowledge graph as the default data model for learning on heterogeneous knowledge. Data Sci. 2017:1–19.
 2
Goodfellow I, Bengio Y, Courville A. Deep Learning; 2016.
 3
Cai H, Zheng V, Chang K. A comprehensive survey of graph embedding: Problems, techniques, and applications. IEEE Trans Knowl Data Eng. 2018; 30:1616–37.
 4
Perozzi B, AlRfou R, Skiena S. Deepwalk: Online learning of social representations. In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining: 2014. p. 701–10.
 5
Tang J, Qu M, Wang M, Zhang M, Yan J, Mei Q. Line: Largescale information network embedding. In: Proceedings of the 24th International Conference on World Wide Web: 2015. p. 1067–77.
 6
Grover A, Leskovec J. node2vec: Scalable feature learning for networks. In: Proceedings of KDD: 2016. p. 855–64.
 7
Dong Y, Chawla NV, Swami A. metapath2vec: Scalable representation learning for heterogeneous networks. In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. ACM: 2017. p. 135–44.
 8
Chen B, Ding Y, Wild DD. Assessing drug target association using semantic linked data. PLoS Comput Biol. 2012; 8(7):1002574.
 9
Seal A, Ahn Y, Wild D. Optimizing drugtarget interaction prediction based on random walk on heterogeneous networks. J Cheminform. 2015;7(40).
 10
Himmelstein D, Baranzini S. Heterogeneous network edge prediction: A data integration approach to prioritize diseaseassociated genes. PLoS Comput Biol. 2015.
 11
Chen B, Dong X, Jiao D, Wang H, Zhu Q, Ding Y, Wild DJ. Chem2bio2rdf: a semantic framework for linking and data mining chemogenomic and systems chemical biology data. BMC Bioinformatics. 2010; 11(1):255.
 12
Bhagat S, Cormode G, Muthukrishnan S. Node classification in social networks. In: Social Network Data Analytics. Springer: 2011. p. 115–48.
 13
Sen P, Namata G, Bilgic M, Getoor L, Galligher B, EliassiRad T. Collective classification in network data. AI Mag. 2008; 29(3):93.
 14
Fouss F, Pirotte A, Renders JM, Saerens M. Randomwalk computation of similarities between nodes of a graph with application to collaborative recommendation. IEEE Trans Knowl Data Eng. 2007; 19(3):355–69.
 15
Yu X, Ren X, Sun Y, Gu Q, Sturt B, Khandelwal U, Norick B, Han J. Personalized entity recommendation: A heterogeneous information network approach. In: Proceedings of the 7th ACM International Conference on Web Search and Data Mining. ACM: 2014. p. 283–92.
 16
Gao Z, Gao Z, Huang H, Jiang Z, Yan Y. An endtoend model of predicting diverse ranking on heterogeneous feeds; 2018.
 17
Gao Z, Liu X. Personalized community detection in scholarly network. iConference 2017 Proc. 2017;Vol. 2.
 18
Liu X, Yu X, Gao Z, Xia T, Bollen J. Comparing communitybased information adoption and diffusion across different microblogging sites. In: Proceedings of the 27th ACM Conference on Hypertext and Social Media. ACM: 2016. p. 103–12.
 19
Zhang C, Gao Z, Liu X. How others affect your twitter# hashtag adoption? examination of communitybased and contextbased information diffusion in twitter. IConference 2016 Proc. 2016.
 20
LibenNowell D, Kleinberg J. The linkprediction problem for social networks. J Assoc Inf Sci Technol. 2007; 58(7):1019–31.
 21
Roweis ST, Saul LK. Nonlinear dimensionality reduction by locally linear embedding. Science. 2000; 290(5500):2323–6.
 22
Saul LK, Roweis ST. An introduction to locally linear embedding. 2000. unpublished. Available at: http://www.cs.toronto.edu/~roweis/lle/publications.html. Accessed 2000.
 23
Belkin M, Niyogi P. Laplacian eigenmaps and spectral techniques for embedding and clustering. In: Advances in Neural Information Processing Systems: 2002. p. 585–91.
 24
Tang L, Liu H. Leveraging social media networks for classification. Data Min Knowl Discov. 2011; 23(3):447–78.
 25
Cox TF, Cox MA. Multidimensional Scaling: CRC press; 2000.
 26
Tenenbaum JB, De Silva V, Langford JC. A global geometric framework for nonlinear dimensionality reduction. Science. 2000; 290(5500):2319–23.
 27
Chen M, Yang Q, Tang X. Directed graph embedding. In: IJCAI: 2007. p. 2707–12.
 28
Collobert R, Weston J. A unified architecture for natural language processing: Deep neural networks with multitask learning. In: Proceedings of the 25th International Conference on Machine Learning. ACM: 2008. p. 160–7.
 29
Mikolov T, Karafiát M, Burget L, Cernockỳ J, Khudanpur S. Recurrent neural network based language model. In: Interspeech: 2010. p. 3.
 30
Pennington J, Socher R, Manning C. Glove: Global vectors for word representation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP): 2014. p. 1532–43.
 31
Cao S, Lu W, Xu Q. Grarep: Learning graph representations with global structural information. In: Proceedings of the 24th ACM International on Conference on Information and Knowledge Management. ACM: 2015. p. 891–900.
 32
Tu C, Wang H, Zeng X, Liu Z, Sun M. Communityenhanced network representation learning for network analysis. 2016. arXiv preprint arXiv:1611.06645.
 33
Mikolov T, Chen K, Corrado G, Dean J. Efficient estimation of word representations in vector space. 2013. arXiv preprint arXiv:1301.3781.
 34
Mikolov T, Sutskever I, Chen K, Corrado GS, Dean J. Distributed representations of words and phrases and their compositionality. In: Advances in Neural Information Processing Systems: 2013. p. 3111–9.
 35
Levy O, Goldberg Y, Dagan I. Improving distributional similarity with lessons learned from word embeddings. Trans Assoc Comput Linguist. 2015; 3:211–25.
 36
Xu Y. An empirical study of locally updated largescale information network embedding (line). PhD thesis. Los Angeles: University of California; 2017.
 37
Sun Y, Han J, Yan X, Yu PS, Wu T. Pathsim: Meta pathbased topk similarity search in heterogeneous information networks. Proc VLDB Endowment. 2011; 4(11):992–1003.
 38
Fu G, Ding Y, Seal A, Chen B, Sun Y, Bolton E. Predicting drug target interactions using metapathbased semantic network analysis. BMC Bioinformatics. 2016; 17(1):160.
 39
Kedem D, Tyree S, Sha F, Lanckriet GR, Weinberger KQ. Nonlinear metric learning. In: Advances in Neural Information Processing Systems: 2012. p. 2573–81.
 40
Box JF. Guinness, gosset, fisher, and small samples. Stat Sci. 1987:45–52.
Acknowledgements
Not applicable.
Funding
The work was supported by National Natural Science Foundation of China (No. 71573162). The funding bodies had no role in the design of this study, the collection, analysis, and interpretation of data, or the writing of this manuscript.
Author information
Affiliations
Contributions
GZ prepared the dataset, programmed the algorithm, and initially drafted the manuscript. XL and YD conceived and guided the project. JJY revised the manuscript with emphasis on biomedical data science applications. GF helped to prepare the ground truth data set and comparison and helped with case study and editing. ST helped to implement the baselines. CO made contributions on the background writing. CG, BF, DW and QY are discussed the scientific ideas, reviewed and approved the final manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Gao, Z., Fu, G., Ouyang, C. et al. edge2vec: Representation learning using edge semantics for biomedical knowledge discovery. BMC Bioinformatics 20, 306 (2019). https://doi.org/10.1186/s1285901929142
Received:
Accepted:
Published:
Keywords
 Knowledge graph
 Heterogeneous network
 Biomedical knowledge discovery
 Representation learning
 Graph embedding
 Node embedding
 Edge semantics
 Applied machine learning
 Data science
 Linked data
 Semantic web
 Network science
 Systems biology