Skip to main content

CAT-DTI: cross-attention and Transformer network with domain adaptation for drug-target interaction prediction

Abstract

Accurate and efficient prediction of drug-target interaction (DTI) is critical to advance drug development and reduce the cost of drug discovery. Recently, the employment of deep learning methods has enhanced DTI prediction precision and efficacy, but it still encounters several challenges. The first challenge lies in the efficient learning of drug and protein feature representations alongside their interaction features to enhance DTI prediction. Another important challenge is to improve the generalization capability of the DTI model within real-world scenarios. To address these challenges, we propose CAT-DTI, a model based on cross-attention and Transformer, possessing domain adaptation capability. CAT-DTI effectively captures the drug-target interactions while adapting to out-of-distribution data. Specifically, we use a convolution neural network combined with a Transformer to encode the distance relationship between amino acids within protein sequences and employ a cross-attention module to capture the drug-target interaction features. Generalization to new DTI prediction scenarios is achieved by leveraging a conditional domain adversarial network, aligning DTI representations under diverse distributions. Experimental results within in-domain and cross-domain scenarios demonstrate that CAT-DTI model overall improves DTI prediction performance compared with previous methods.

Peer Review reports

Introduction

Drug discovery is highly valued in the current biomedical field [1]. In drug discovery, verifying whether a drug interacts with a certain target is a key step in proving drug effectiveness [2]. In vitro screening experiments are feasible but labor-intensive, expensive and time-consuming [3]. The utilization of computerized screening for potential DTI candidates has been substantiated as an effective strategy to aid biologists in the identification of genuine DTIs through wet-lab experiments [4]. Therefore, computer-aided DTI prediction has aroused great interest and received increasing attention.

Existing DTI prediction methods can generally be divided into three categories: ligand-based, structure-based and machine learning-based [5]. Traditional structure-based and ligand-based in silico virtual screening methods have gained increased attention owing to the demonstrated relative effectiveness [6,7,8]. However, these traditional methods have significant limitations. For example, the widely adopted molecular docking method is relatively inefficient, sometimes slow due to the huge amount of computation and the scoring function has relatively low accuracy [9]. In short, ligand-based methods face limited application scope and performance challenge due to the limited number of known ligands for some proteins. In addition, structural-based methods are limited by the lack of three-dimensional structures and ligand information for the majority of protein, thereby constraining the development.

Traditional machine learning models such as support vector machine (SVM) [10] and random forest (RF) [11] are also used for DTI prediction [12]. For example, Faulon et al. [13] used molecular features and reaction features as input to the SVM kernel function for DTI prediction. Wang et al. [14] used the features extracted by the Boruta algorithm as the input of the RF algorithm for DTI prediction. However, although these methods are simple and effective, the performance is far from satisfactory.

Recently, as an important branch of machine learning, deep learning has also made significant progress in DTI prediction. In the early days, researchers used hand-crafted descriptors of drugs and proteins to make predictions through a fully connected neural network [15]. Subsequently, Lee et al. proposed DeepConv-DTI [16], which used convolution neural network (CNN) to extract protein features, employed the extended connectivity fingerprints (ECFP) algorithm [17] to calculate drug features, and predicted DTI through fully connected network (FCN). However, the interaction characteristics of drug-protein pairs are ignored. Furthermore, the adoption of advanced feature extraction techniques, such as the DynCNN module in SAG-DTA [18] and DrugVQA [19], highlight the continuous efforts to optimize feature extraction methods.

Over the past few years, graph neural network (GNN) have demonstrated excellent predictive performance in addressing key prediction challenges in the field of bioinformatics by utilizing the powerful feature representation learning capability [20,21,22]. To extract the topological information of drugs, Nguyen et al. designed GraphDTA [23] based on GNN, treating drugs as molecular graphs, using GNN and CNN to extract drug and protein features respectively to predict the affinity of drugs and targets. Despite the use of stronger feature extraction modules, the important fact that the interactions between molecules are mainly focused on the relevant substructures of drugs and proteins is ignored [24]. Furthermore, a simply connection of drug and protein features fails to capture the complex interactions between them.

In order to more accurately model the interactions between drugs and proteins, some studies have introduced the attention mechanism into DTI prediction [25]. Nowadays, attention mechanisms have been widely used for revealing the contribution of different components of a drug or target on interaction [19] and describing interactions between targets and drugs [26]. HyperAttentionDTI [27] assigned attention vectors to each atom and amino acid on the basis of CNN to enhance feature expression. Although this method considered the interactions representation between drugs and proteins, the limited receptive field of CNN limits the ability to capture global dependencies.

Inspired by the powerful ability of Transformer [28] to capture features between two sequences, Chen et al. proposed TransformerCPI [29], using Transformer to predict DTI from the SMILES of drugs and protein amino acid sequences. Huang et al. proposed MolTrans [30], which applied Transformer to extract features from the substructure of drug and protein sequence and combined them into interaction map for compound-protein interaction prediction. However, this method mainly focuses on the interaction features between drugs and protein substructures, while ignoring the importance of utilizing original feature information.

Due to the wide scope and complexity of the chemical and genomic fields, DTI prediction often faces great challenges in real-world scenarios. Recently, Bai et al. proposed the DrugBAN [31] using bilinear attention to capture the local interaction representation of drug and target for DTI prediction. In cross-domain prediction tasks, the conditional adversarial domain adaptation method is introduced to transfer source domain knowledge to the target domain and demonstrate excellent cross-domain generalization capability.

In order to cope with the problem that many models [16, 23, 27, 31] cannot fully capture global context information while retaining local features when processing global and local information, resulting in the inability to extract sufficiently effective feature information. In this work, we propose a model named CAT-DTI, whose protein features are extracted by a protein feature encoder combining CNN and Transformer, which fully considers global context information while capturing local features of protein sequences. Besides, attention-based methods [23, 32,33,34] generally focus more on extracting internal features of drugs and targets, but rarely introduce attention to mine DTI representations. With the aim of better preserving the internal features of drugs and proteins while deeply exploring the interaction information between them, we input the extracted drug and protein features into the cross-attention module for feature fusion. We also notice that the training of the model in a specific domain is mainly aimed at the distribution of the domain and the drug-target pairs to be predicted in practical applications may have a different distribution from the training data, resulting in the inability to directly transfer the existing knowledge to new scenarios. Therefore, in cross-domain tasks, we employ conditional domain adversarial network (CDAN) to better understand and predict DTI in domains that are distributed differently from the training data, thus enhancing the cross-domain generalization ability.

Our contributions are summarized in the following points. First, we propose a deep learning model named CAT-DTI, which uses GCN and CNN combined with Transformer to extract feature maps of drugs and proteins, respectively. Second, we employ cross-attention module that fuse drug and protein features, effectively capture and process the interaction features between drugs and proteins while retaining the internal feature information. For the cross-domain prediction task, we adopt the CDAN to enable the model to adapt and cope with the characteristics of new domains, improve the performance of CAT-DTI in cross domain DTI prediction tasks and enhance the generalization performance and practical value of CAT-DTI. Compared with other baseline methods, CAT-DTI achieves generally better predictive performance on three public datasets.

Methods

An overview of CAT-DTI framework is illustrated in Fig. 1a. Given drug SMILES and protein amino acid sequences as input, the protein and drug embeddings are generated. Drug embedding is input into GCN to extract feature representations of drug molecules (i.e., drug feature map \(F_D\)). The protein embedding is passed to the protein feature encoder as shown in Fig. 1b, which combines the CNN and Transformer to extract the protein feature map \(F_P\), capturing local features and global context information in the protein sequence simultaneously. Next, the cross-attention module interacts protein and drug features for feature fusion to capture the interaction relationship between drugs and targets, as shown in Fig. 1c. Specifically, we swap the key and value of protein attention with those of drug attention. After obtaining the feature maps, the original features are integrated to construct the final features for both drugs and proteins. Through max-pooling and concatenation, the joint feature f for drug and protein target is produced and input into the decoder to predict DTI. To enhance the generalization performance of CAT-DTI in real-world scenarios for novel drug-target pairs, we integrate the domain adaptation module CDAN into the framework, which is employed to adapt the representations of drugs and proteins, thereby facilitating effective alignment between source and target domain distributions.

Fig. 1
figure 1

Framework of the proposed CAT-DTI. a Overview of CAT-DTI framework. b Details of the protein feature encoder. c Processes of cross-attention

GCN for drug molecular graph

Regarding the drug feature extraction process, we transform drug SMILES into a corresponding 2D molecular graph. To capture the node information within the graph, we first initialize each atom node. Each atom is denoted by a 74-dimensional integer vector that encapsulates eight distinct attributes, including the atom type, the atom degree, the number of implicit Hs, the formal charge, the number of radical electrons, the atom hybridization, the number of total Hs and whether the atom is aromatic.

The drug feature encoder transmits and aggregates information on the drug molecular structure through a three-layer GCN, thereby achieving extraction and representation of drug feature. In each layer of GCN operation, each row of the drug representation represents an aggregated representation of adjacent atomic nodes in the drug molecule. Each GCN layer uses the information of neighboring atomic nodes to update the feature representation of each atomic node, allowing the model to effectively capture the correlation information between neighboring atomic nodes. We retain node-level drug representations for subsequent explicit learning of interactions with protein fragments. We set the maximum number of nodes in the graph to be \(m_d\). Therefore, the node feature matrix of each graph is denoted as \(M_d\in \mathbb {R}^{m_d\times 74}\). Furthermore, we employ a simple linear transformation to establish \(F_{d}=M_{d}W_{o}^{\top }\), resulting in a real-valued dense matrix \(F_d\in \mathbb {R}^{m_d\times D_d}\) as input features, where \(D_d\) is the drug embedding dimension. Finally, we obtain the drug feature map \(F_D\in \mathbb {R}^{m_d\times D_d}\) through the drug feature encoder, which can be expressed as:

$$\begin{aligned} H_d^{i+1}=\sigma (\text {GCN}(\widetilde{A},W_{gcn}^i,b_{gcn}^i,H_d^i)), \end{aligned}$$
(1)

where \(W_{gcn}^i\) and \(b_{gcn}^i\) are the weight matrices and bias vector of the i-th layer of GCN. \(\tilde{A}\) is the adjacency matrix with added self-connection. \(H_{d}^i\) denotes the hidden node representation of layer i with \(H_{d}^{0}=F_{d}\).

Feature encoder for protein

To enhance protein sequence feature representation and capture long-distance relationship between sequence tags, we introduce a protein feature encoder that combines CNN and Transformer. Traditional CNN may struggle with long sequences due to the limited local receptive fields, so we combine the global attention mechanism of Transformer to capture long-distance dependence in protein sequences. By fusing the local perception capabilitiy of CNN and the global attention mechanism of Transformer, our model simultaneously considers local features and global context information in protein sequences, thereby extracting more effective protein features.

It is worth noting that before the feed forward layer of Transformer, we add 1D CNN to process local information. By sliding the convolution kernel on the protein sequence, we captured the local pattern and substructure of the protein. Combined with the advantages of Transformer in handling long-range dependencies, our model achieves the fusion of local and global information in the protein feature encoding process, which is beneficial to enhance the representation of protein sequence features. In our work, a three-layer protein feature encoder is used to capture protein features, as shown in Fig. 1b, where each layer includes a multi-head self-attention, CNN and a feed-forward neural network. Specifically, the protein sequence is input to the protein feature encoder with the feature matrix \(F_p\in \mathbb {R}^{l_p\times D_p}\), where \(l_p\) is the length of the protein sequence and \(D_p\) is the protein embedding dimension. The matrices \(Q\in \mathbb {R}^{l_p\times D_p}\), \(K\in \mathbb {R}^{l_p\times D_p}\) and \(V\in \mathbb {R}^{l_p\times D_p}\) in different feature spaces based on the feature matrix \(F_p\) are generated by the linear layer as follows:

$$\begin{aligned} {\left\{ \begin{array}{ll}Q=F_p\cdot W_Q+b_Q\\ K=F_p\cdot W_K+b_K,\\ V=F_p\cdot W_V+b_V\end{array}\right. } \end{aligned}$$
(2)

where \(W_Q\in \mathbb {R}^{D_p\times D_p}\), \(W_K\in \mathbb {R}^{D_p\times D_p}\), \(W_V\in \mathbb {R}^{D_p\times D_p}\) are learnable parameter weights. \(b_Q\), \(b_K\) and \(b_V\) are bias vectors. Given Q, K and V matrices, the self-attention layer computes the attention weights as follows:

$$\begin{aligned} \text {Attention}(Q,K,V)=\text {Softmax}(\frac{Q\cdot K^\top }{\sqrt{d_k}})V, \end{aligned}$$
(3)

where \(d_k\) is the dimension of K. The output \(X_M\) of the multi-head attention layer is generated as follows:

$$\begin{aligned} \begin{aligned} X_M&=\text {MutiHead}(Q,K,V)=\text {Concat}(\text {Attention}(Q,K,V))W_M+b_M, \end{aligned} \end{aligned}$$
(4)

where \(W_M\in \mathbb {R}^{D_p\times D_p}\) is the learnable weight matrix and \(b_M\) is the bias vector.

The multi-head attention layer extracts information from diverse representation subspace, enhancing model robustness. Therefore, long-range relationships between amino acids spanning the entire sequence can be learned with self-attention weights. Additionally, the first ADD & Norm layer implements a residual connection with original protein feature matrix \(F_p\) and then follow by normalization, expressed as follows:

$$\begin{aligned} X_{AN}=\text {LayerNorm}(F_{p}+X_{M}), \end{aligned}$$
(5)

Subsequently, a three-layer CNN is inserted after the first ADD & Norm layer to extract local feature in the protein sequence:

$$\begin{aligned} X_{CNN}=\textrm{CNN}\left( X_{AN}\right) , \end{aligned}$$
(6)

After the second ADD & Norm layer, we derive the protein feature map \(F_{P}\in \mathbb {R}^{l_p \times D_p}\) as follows:

$$\begin{aligned} F_P=\text {LayerNorm}(X_{CNN}+X_{AN}), \end{aligned}$$
(7)

Cross-attention module

After obtaining the feature maps for drugs and proteins through the feature encoder, we introduce a cross-attention module to effectively model the interaction between drugs and proteins, thereby capturing enhanced representations of their interaction and provides more reliable feature representation for DTI prediction. By performing two-way information interaction between the key and value of protein attention and the key and value of drug attention, the information exchange and association between drug and protein is realized, thus capturing the interaction features between drug and protein target. In this process, protein features can adjust their own expression by attention weights of drug features, and vice versa. Such an interaction and adjustment mechanism enable the cross-attention module to promote information flow across feature maps, effectively fuse drug and protein features, and extract more comprehensive DTI feature representation. The cross-attention module is depicted in Fig. 1c and primarily consists of drug and protein attention.

In this section, we set \(D_e=D_d=D_p\). For drugs, the drug feature map \(F_D\) is passed through the linear layer to calculate the drug query vector \(Q_D^i\in \mathbb {R}^{m_d\times d_{head}}\), and then the protein feature map \(F_P\) is obtained through a linear layer, which is further calculated as the drug key vector \(K_D^i\in \mathbb {R}^{l{_p\times d}_\textit{head}}\) and value vector \(V_D^i\in \mathbb {R}^{l{_p\times d}_\textit{head}}\). The query, key and value for the drug are obtained as follows:

$$\begin{aligned} {\left\{ \begin{array}{ll}Q_D^i=F_D\cdot W_q^i\\ K_D^i=F_P\cdot W_k^i ,\\ V_D^i=F_P\cdot W_\nu ^i\end{array}\right. } \end{aligned}$$
(8)

where \(\begin{aligned}W_q^i, W_k^i, W_\nu ^i\in \mathbb {R}^{D_e\times d_{head}}\end{aligned}\) are different weight matrices in the linear layer and \(d_{head}=D_{e}/heads\) is the channel dimension. \(i=1,2,\cdots ,heads\), where heads are the number of attention heads.

Protein attention follows a process similar to drug attention. The protein feature map \(F_P\) is input into the linear layer to calculate the protein query vector \(Q_P^i\in \mathbb {R}^{l{_p\times d}_\textit{head}}\), and then the drug feature map is taken to generate the protein key vector \(K_P^i\in \mathbb {R}^{m{_d\times d}_\textit{head}}\) and protein value vector \(V_P^i\in \mathbb {R}^{m{_d\times d}_\textit{head}}\). The queries, keys and values of proteins are calculated by the following formulas:

$$\begin{aligned} \left. \left\{ \begin{aligned}Q_P^i&=F_P\cdot W_q^i\\K_P^i&=F_D\cdot W_k^i\\V_P^i&=F_D\cdot W_v^i\end{aligned}\right. \right. , \end{aligned}$$
(9)

where the weight matrices \(W_q^i\in \mathbb {R}^{D{_e}\times d_\text {head}}\), \(W_k^i\in \mathbb {R}^{D{_e}\times d_\text {head}}\) and \(W_v^i\in \mathbb {R}^{D{_e}\times d_\text {head}}\) share the same weights as drug attention. Through the application of a softmax function, the drug and protein attention matrices are computed as:

$$\begin{aligned}{} & {} A_D^i=\textrm{Softmax}\left( \frac{Q_D^i\cdot K_D^i {^\top }}{\sqrt{d_{K_D^i}}}\right) , \end{aligned}$$
(10)
$$\begin{aligned}{} & {} A_P^i=\textrm{Softmax}\left( \frac{Q_P^i\cdot K_P^i {^\top }}{\sqrt{d_{K_P^i}}}\right) , \end{aligned}$$
(11)

where \(d_{K_D^i}=d_{K_P^i}=d_{head}\) is the dimension of K for drug and protein. The drug/protein feature map for each head is obtained by multiplying the drug/protein attention matrix of each attention head with the corresponding drug/protein value matrix. Subsequently, the drug/protein feature maps of all attention heads are concatenated in the channel dimension and fed into the linear layer to obtain the final drug feature representation \(Z_P\in \mathbb {R}^{l_p\times D_p}\) and protein feature map \(Z_D\in \mathbb {R}^{m_d\times D_d}\) received attention:

$$\begin{aligned}{} & {} Z_D=\text {Concat}(A_D^i\times V_D^i)\times W_Z, \end{aligned}$$
(12)
$$\begin{aligned}{} & {} Z_P=\text {Concat}(A_P^i\times V_P^i)\times W_Z, \end{aligned}$$
(13)

where \(i=1,2,\cdots ,heads\) and \(W_Z\in \mathbb {R}^{D_e \times D_e}\) is the shared weight matrix.

Next, the feature maps of interest are combined with the original feature maps to obtain the final drug feature map \(F_{ZD}\in \mathbb {R}^{m_d\times D_d}\) and protein feature map \(F_{ZP}\in \mathbb {R}^{l_p\times D_p}\):

$$\begin{aligned}{} & {} F_{ZD}=0.5Z_D+0.5F_D, \end{aligned}$$
(14)
$$\begin{aligned}{} & {} F_{ZP}=0.5Z_P+0.5F_P, \end{aligned}$$
(15)

The drug and protein feature maps are downsampled by using a global max-pooling operation to generate one-dimensional drug feature vector \(d_{mp}\in \mathbb {R}^{D_d}\) and protein feature vector \(p_{mp}\in \mathbb {R}^{D_p}\):

$$\begin{aligned}{} & {} d_{mp}=\text {Maxpooling}(F_{ZD}), \end{aligned}$$
(16)
$$\begin{aligned}{} & {} p_{mp}=\text {Maxpooling}(F_{ZP}), \end{aligned}$$
(17)

Finally, we concatenate \(d_{mp}\) and \(p_{mp}\) to obtain the joint feature representation \(f\in \mathbb {R}^{2D_e}\):

$$\begin{aligned} f=\textrm{Concat}(d_{mp},p_{mp}), \end{aligned}$$
(18)

Drug-target interaction prediction

In order to predict the DTI probability, we input the joint representation f into the decoder, which consists of a fully connected classification layer. Finally, the DTI probability p is generated as follows:

$$\begin{aligned} p=\sigma \left( Wf+b\right) , \end{aligned}$$
(19)

where W and b are learnable weight matrix and bias vector.

During model training, we employ backpropagation to concurrently optimize the learnable parameters. Our objective in training is to minimize the cross-entropy loss function:

$$\begin{aligned} \mathcal {L}=-\sum _i\left( y_i\log {(p_i)}+(1-y_i)\log {(1-p_i)}\right) +\frac{1}{2}\lambda ||\theta ||_2^2, \end{aligned}$$
(20)

where \(y_i\) denotes the ground-truth label of the i-th drug-target pair. \(p_i\) represents DTI prediction score predicted by the model. \(\theta\) is the set of learnable weight matrices and bias vectors and \(\lambda\) is a hyperparameter for L2 regularization to prevent overfitting.

Cross-domain adaptation enhances generalization

Deep learning models show excellent performance on similar data (i.e., in-domain) that is distributed with the training data. However, the performance on different data with different distributions (i.e., cross-domain) is not satisfactory. To this end, we employ  the CDAN module to improve the generalization ability of CAT-DTI model from a source domain rich in labeled data to a target domain containing only unlabeled data. Figure 2 shows the framework after integrating the CDAN module into CAT-DTI (i.e., \({\text {CAT-DTI}}_{\textrm{CDAN}}\)), which consists of three key components: Feature Extractor \(F\left( *\right)\), Decoder \(G\left( *\right)\) and Discriminator \(D\left( *\right)\).

Fig. 2
figure 2

Diagram of cross-domain adaptation process. CDAN is a domain adaptation technique designed to address domain shift challenges with different distributions. We utilize CDAN to integrate the joint representation f of the source and target domain, along with classifier prediction g into the joint conditional representation distinguished by the discriminator. The discriminator is structured as a three-layer fully connected network with the specific goal of distinguishing the target domain from source domain by minimizing domain classification error

On the cross-domain task, given \(N_S\) labeled drug-target pairs \(P_S=\{(x_s^i,y_s^i)\}_{N_S}^{i=1}\) in the source domain and \(N_T\) unlabeled drug-target pairs \(P_t=\{(x_t^i)\}_{N_T}^{i=1}\) in the target domain. We rely on CDAN to adjust the distribution of samples to optimize cross-domain prediction performance. The feature extractor \(F\left( *\right)\) is the drug and protein feature encoder together with the cross-attention module to generate a joint representation of the input domain data, namely \(f_s^i=F(x_s^i)\) and \(f_t^j=F(x_t^j)\). For the decoder \(G\left( *\right)\), we employ a fully connected classification layer and follow a softmax function as \(G\left( *\right)\) to obtain predicted classification results \(g_s^i=G(f_s^i)\in \mathbb {R}^2\) and \(g_t^j=G(f_t^j)\in \mathbb {R}^2\). Subsequently, the joint representation f and the classifier prediction g are embedded into a joint conditional representation \(c\in \mathbb {R}^{2D_e}\), which is defined as follows:

$$\begin{aligned} c=\text {FLATTEN}(f\otimes g), \end{aligned}$$
(21)

where \(\text {FLATTEN}\) performs a flattening operation on the outer product of the f and g vectors and \(\otimes\) is the outer product.

Adhering to CDAN principles, we employ a domain discriminator \(D\left( *\right)\) to align the joint representation f and predicted classification distribution g of the source and target domains. \(D\left( *\right)\) is a domain discriminator composed of a three-layer FCN that learns to distinguish whether a joint conditional representation c originates from the source or target domain. \(F\left( *\right)\) and \(G\left( *\right)\) are trained to minimize the cross-entropy loss \(\mathcal {L}\) of the source domain with source label information, generating a joint conditional representation c that confuses the discriminator \(D\left( *\right)\). In the cross-domain task, we utilize two losses: one for optimizing classification prediction and the other for optimizing the distribution alignment of the source and target domain:

$$\begin{aligned}{} & {} \mathcal {L}_S(F,G)=\mathbb {E}_{(x_s^i,y_s^i)\thicksim P_s}\mathcal {L}\left( G\left( F\left( x_s^i\right) \right) ,y_s^i\right) , \end{aligned}$$
(22)
$$\begin{aligned}{} & {} \begin{aligned} \mathcal {L}_{ad\nu }(F,G,D)&=\mathbb {E}_{x_t^i\thicksim P_t}\log\left( 1-D\left( f_t^i,g_t^i\right) \right) +\mathbb {E}_{x_s^j\thicksim P_s}\log(D(f_{s}^j,g_s^j)), \end{aligned} \end{aligned}$$
(23)

where \(\mathcal {L}_S\) is the cross-entropy loss on the labeled source domain and \(\mathcal {L}_{ad\nu }\) is the adversarial loss for the domain discriminator.

The optimization problem is written as a minimax paradigm:

$$\begin{aligned} \max _{\begin{array}{c}D\end{array}}\min _{\begin{array}{c}F,G\end{array}}\mathcal {L}_S(F,G)-\omega \mathcal {L}_{ad\nu }(F,G,D), \end{aligned}$$
(24)

where \(\omega\) is a hyper parameter for weighting \(\mathcal {L}_{ad\nu }\). By introducing adversarial training in \(\mathcal {L}_{ad\nu }\), the difference in data distribution between the source domain and target domain is reduced, thereby enhancing the generalization ability of cross-domain prediction.

Experiments and results

Datasets and data processing

We comprehensively evaluate CAT-DTI and six baseline models on three public datasets: BindingDB, BioSNAP and Human. The BindingDB database records the binding affinity information of small drug molecules and proteins that have been verified through experiments and mainly studies the interaction between drug-like molecules and proteins. In experiments, we use a low-bias version of the BindingDB dataset constructed by previous research [35]. The BioSNAP dataset is constructed according to previous research [30, 36] derived from the DrugBank database [37], including 4510 drugs and 2181 proteins. It is designed as a balanced dataset containing validated positive samples and an equal number of unseen negative samples. Drawing on the previous studies [19, 29], we also employ a balanced version of the Human dataset containing equal numbers of positive and negative samples. The statistics of the three datasets are shown in Table 1.

Table 1 Details of datasets used in this work

In experiments, we use different splitting strategies on the datasets for in-domain and cross-domain tasks. For in-domain evaluation, each experimental dataset is randomly split into training, validation and test sets in a ratio of 7:1:2. For cross- domain tasks, the decision to exclude the Human dataset stems from its comparatively limited sample size. In order to ensure the model has robust performance in cross-domain scenarios, the datasets should have sufficient data volume and sample diversity, so our cross-domain performance evaluation focuses on the large-scale BindingDB and BioSNAP datasets. We utilize the datasets from the previous study [31], which adopts a clustering-based pair split strategy to build cross-domain scenario and cluster drugs and target proteins from BindingDB and BioSNAP datasets respectively for cross-domain performance evaluation. Specifically, a single-linkage clustering method is used to cluster from the bottom and hierarchically to ensure that the distance between samples in different clusters always exceeds a predefined minimum distance threshold, which helps prevent the formation of clusters that are too close. For each dataset, the single-linkage algorithm is used for the clustering of drugs and proteins based on the ECFP4 [17] fingerprint and pseudo-amino acid composition (PSC) [38], respectively. Since the clustering-based pair split enables the quantitative construction of cross-domain task by taking into account the similarity between drugs and proteins, we use Jaccard distance and cosine distance on ECFP4 and PSC respectively to accurately measure pairwise distances. During the clustering of drugs and proteins, the distance threshold is set to 0.5 to ensure that the clusters do not become too large while maximizing the separation of different samples. So far, 2,780 drug clusters and 1,693 protein clusters have been obtained in the BindingDB dataset and 2,387 drug clusters and 1978 protein clusters have been obtained in the BioSNAP dataset. Through the clustering-based pair split strategy, the source and target domain are characterized by non-overlapping sets with different distributions. Following the general setup of domain adaptation, we use all labeled source domain data and 80% unlabeled target domain data as the training set and the remaining 20% labeled target domain data as the test set. While cross-domain evaluation presents greater challenges compared to in-domain random splitting, it emerges as a more efficacious methodology for assessing the generalization capacity of model in the practical realm of drug discovery.

Baselines

We compare CAT-DTI with the following baselines.

  • SVM [10] and RF [11] are used as a classifier to classify encoded drug and protein features.

  • GraphDTA [23] uses GNN to encode drug molecule graphs and CNN to encode protein sequences. The learned drug and protein representation vectors are combined with a simple concatenation.

  • TransformerCPI [29] treats drugs and proteins as two sequences, generating representations of protein sequences and drug atoms. The interaction feature is captured by Transformer decoder and the interaction probability is output by a fully connected layer.

  • MolTrans [30] is a deep learning model that utilizes Transformer to encode drug and protein information and learns the interactions between substructures through a CNN-based interaction module.

  • DrugBAN [31] encodes drug SMILES through GCN to obtain drug features and uses CNN to capture protein features. Drug-target pairwise interactions are captured by a bilinear attention network, followed by a fully-connected classifier for DTI prediction. Besides, conditional domain adversarial network is used to adjust the distribution of source and target domains in cross-domain prediction tasks.

Evaluation metrics

Since DTI prediction is a classification task, we choose the widely recognized evaluation metrics AUROC (i.e., the area under the receiver operating characteristic curve) and AUPRC (i.e., the area under the precision-recall curve) as our primary assessment criteria. Additionally, we provide reports on Accuracy, Sensitivity, Specificity, and F1 score. In principally, the higher the AUROC value means the better the performance. The definitions of these evaluation metrics are given as:

$$\begin{aligned}{} & {} \text {Accuracy}=\frac{TP+TN}{TP+FP+TN+FN}, \end{aligned}$$
(25)
$$\begin{aligned}{} & {} \text {Sensitivity}=\frac{TP}{TP+FN}, \end{aligned}$$
(26)
$$\begin{aligned}{} & {} \text {Specificity}=\frac{TN}{TN+FP}, \end{aligned}$$
(27)
$$\begin{aligned}{} & {} F1=\frac{2TP}{2TP+FP+FN}, \end{aligned}$$
(28)

where the true positive (TP) and true negative (TN) are the number of drug-target with interaction and drug-target without interaction that are successfully identified, respectively. The false positive (FP) and false negative (FN) represent the number of drug-target with interaction and drug-target without interaction examples that are incorrectly identified.

Experimental setting

During the experiments, we allow our model to run for up to 100 epochs on all datasets, setting the experimental batch size to 32. We use the Adam optimizer with a learning rate of \(2.5\times 10^{-5}\) for the in-domain tasks and \(5\times 10^{-5}\) for the cross-domain tasks. The maximum sequence length allowed for proteins is set to 1000 and the maximum number of atoms allowed for drug molecules is 290. The number of hidden neurons in the fully connected decoder is 512. Five independent experiments are run for each dataset split. The best performing model is the one that exhibits the highest AUROC on the validation set, which is used on the test set to get the performance metrics.

Performance comparison

In-domain performance evaluation

In the in-domain scenario, we use regular CAT-DTI in our experiments, so we do not embed the CDAN into the model. We compare CAT-DTI with six baselines in the random split setting: SVM [10], RF [11], GraphDTA [23], TransformerCPI [29], MolTrans [30] and DrugBAN [31]. Table 2 shows the comparison on BindingDB, BioSNAP and Human datasets. The experimental results indicate that CAT-DTI demonstrates competitive performance across all evaluation metrics in the in-domain scenario. Specifically, CAT-DTI outperforms other comparative methods in terms of evaluation metrics such as AUROC, AUPRC, F1, and specificity on the BioSNAP and Human datasets, which is attributed to its strong capability in extracting key features of drugs and proteins, effectively capturing their interacting characteristics. It is noteworthy that CAT-DTI do not achieve the best performance on evaluation metrics such as AUROC, AUPRC, and F1 on the BindingDB dataset. One possible reason is that the key protein information in the BindingDB is mainly reflected in local features, which makes it challenging for the Transformer part of the protein feature encoder in CAT-DTI to fully utilize its potential, and indirectly affects the model’s ability to capture local features, leading to a decline in model performance. Nevertheless, CAT-DTI still outperforms other comparative methods in most cases.

Table 2 Comparison results of CAT-DTI and baselines on three datasets

Overall, the experimental results on the three datasets all confirm the effectiveness of the CAT-DTI for DTI prediction. The improvement of experimental results is attributed to our method not only capturing protein local features but also analyzing global context information for protein features. Moreover, the incorporation of the cross-attention module enables the model to concurrently consider the impact of drug features on proteins and the influence of protein features on drugs. This bidirectional interaction empowers CAT-DTI to comprehensively comprehend and capture the intricate interactions between drugs and proteins, thereby achieving the effective fusion of drug and protein target features.

Cross-domain performance evaluation

In-domain classification tasks under random split are relatively simple and of limited practical value. In order to better simulate real-world situations, we focus on the more challenging cross-domain DTI prediction, where the training data and test data have different distribution characteristics. In order to deeply explore the knowledge transferability in cross-domain prediction, we embed the CDAN module into the CAT-DTI model, which means using CAT-DTI\(_{\text {CDAN}}\) for cross-domain prediction.

We present the cross-domain performance evaluation results on BindingDB and BioSNAP datasets in Table 3. All methods show a significant drop compared to the previous in-domain prediction results due to the reduced information overlap between training and test datasets. However, our newly proposed CAT-DTI model clearly outperforms other state-of-the-art models on both datasets. Specifically, the AUROC and AUPRC of CAT-DTI on the BioSNAP dataset are 10.4% and 7.1% higher than the second-ranked DrugBAN. At the same time, the AUROC and AUPRC of CAT-DTI on the BindingDB dataset are 12.4% and 15.4% higher than those of DrugBAN. What’s more worth mentioning is that even when the CDAN module is incorporated into DrugBAN (i.e., DrugBAN\(_{\text {CDAN}}\)), the performance of CAT-DTI outperforms DrugBAN\(_{\text {CDAN}}\) model. Experimental results demonstrate that the proposed CAT-DTI can effectively handle cross-domain DTI prediction problems. Compared with previous methods, CAT-DTI not only exhibits improved accuracy but also demonstrates strong cross-domain generalization capability. We attribute the superiority of CAT-DTI in cross-domain generalization to its unique framework design and the collaborative interaction of key components. The introduced cross-attention module enables CAT-DTI to simultaneously consider drug and protein features, establishing bidirectional information correlation and aiding in a more comprehensive and accurate capture of interactions across different domains. This feature fusion method enhances the adaptability of CAT-DTI to diverse data distributions, thereby improving cross-domain generalization performance. Additionally, the protein feature encoder combines CNN and Transformer to provide CAT-DTI with comprehensive modeling capability for local and global information within protein sequences.

Table 3 Cross-domain performance comparison of CAT-DTI and other baselines on BindingDB and BioSNAP Datasets

In recent years, domain adaptation techniques have attracted extensive attention in academia due to their excellent cross-domain knowledge transfer capabilities. In our work, we explore and improve cross-domain DTI prediction by combining CAT-DTI model with CDAN. As shown in Table 3, the proposed CAT-DTI\(_{\text {CDAN}}\) model has a significant performance improvement after integrating the domain adaptation module. Specifically, compared with the CAT-DTI model, CAT-DTI\(_{\text {CDAN}}\) improves AUROC and AUPRC on the BindingDB dataset by 6.6% and 9.2% and also has a gratifying improvement on the BioSNAP dataset. In order to more clearly observe the improvement effect of adding the domain adaptation module CDAN to CAT-DTI, we draw the prediction results of CAT-DTI and CAT-DTI\(_{\text {CDAN}}\) on BindingDB and BioSNAP datasets as radar charts, as shown in Fig. 3. In this way, we can more intuitively observe the positive impact of the CDAN module on cross-domain tasks. By reducing cross-domain distribution bias, CAT-DTI demonstrates a substantial enhancement in its generalization performance with the incorporation of the CDAN module. These experimental results further confirm the significant superiority of CAT-DTI in cross-domain generalization ability.

Fig. 3
figure 3

Cross-domain performance comparison of CAT-DTI with and without CDAN module on BindingDB and BioSNAP datasets

We analyze that the cross-domain generalization performance of CAT-DTI\(_{\text {CDAN}}\) has been improved is attributed to CAT-DTI\(_{\text {CDAN}}\) can better optimize feature alignment and effectively reduce the distribution differences between different domains after introducing the CDAN module. Through the adversarial learning mechanism, CAT-DTI\(_{\text {CDAN}}\) is able to adjust the feature representations of the source domain and the target domain to make them statistically more similar, thereby improving the model’s performance on the target domain. This domain adaptation mechanism enables CAT-DTI to better adapt to new and unseen domain data and enhances the generalization ability of the model. Specifically, the introduction of the CDAN module helps reduce domain differences and makes CAT-DTI more robust when processing new drug and protein interaction data, thus improving its prediction accuracy and adaptability.

Ablation experiments

To explore the effect of CNN combined with Transformer for protein feature extraction, DTI feature captured by cross-attention module and domain adaptation module CDAN on model prediction performance, we perform ablation experiments. In the in-domain task, we compare the experimental results of CAT-DTI with three variant models, including CAT-DTI using only CNN for protein feature encoder after removing the Transformer (i.e., Without Transformer), the model using only Transformer for protein feature encoder after removing the CNN (i.e., Without CNN) and the model removing the cross-attention module (i.e., Without cross-attention). In the cross-domain task, in order to evaluate the efficacy of CDAN, we compare the evaluation results of CAT-DTI\(_{\text {CDAN}}\) with CDAN module removed (i.e., CAT-DTI) and DrugBAN fused with CDAN (i.e., DrugBAN\(_{\text {CDAN}}\)). The experimental results are illustrated in Fig. 4.

Fig. 4
figure 4

AUROC and AUPRC for random split and clustering-based split strategies on BindingDB and BioSNAP datasets

By analyzing the experimental results, we observed an enhancement in the predictive performance of CAT-DTI when the model integrated the complete module, which confirms the effectiveness of the mechanism that combines CNN and Transformer for protein feature extraction, highlights the efficient interactive capabilities of the cross-attention module, and indicates the improvement in cross-domain performance of CAT-DTI with the introduction of the CDAN module.

Case study

In order to further verify the effectiveness of CAT-DTI, we selected two representative targets for case study, namely P00519 (Tyrosine-protein kinase ABL1) and P35228 (Nitric oxide synthase, inducible). Based on the testing result from the test set of BindingDB, we select drug-target pairs containing P00519 and P35228, and then rank the candidate drugs for each protein in descending order according to their predicted interaction scores, choosing the top 6 for each target as depicted in Table 4. After detailed review, the drug and target pairs listed in the table have been confirmed to have interactions in the Drugbank database. Taking the example of the drug-target pair involving the target protein P00519 and its first candidate drug DB08901 (Ponatinib) from Table 4, the tyrosine-protein kinase ABL1 plays a pivotal role in various processes related to cell growth and survival. It coordinates actin cytoskeleton dynamics by regulating protein tyrosine phosphorylation. On the other hand, DB08901 (Ponatinib) is a novel Bcr-Abl tyrosine kinase inhibitor that has been proven to inhibit the tyrosine kinase activity of Abl and is used to treat chronic myelogenous leukemia. Therefore, the drug numbered DB08901 has been verified as a ligand for the target protein numbered P00519.

Table 4 Top-ranked list of predicted drugs for two proteins
Table 5 DTI prediction results for five drugs

Furthermore, we have selected additional 5 drugs to expand our case study, including DB00786 (Marimastat), DB01254 (Dasatinib), DB06155 (Rimonabant), DB00482 (Celecoxib), and DB00481 (Raloxifene). Based on the testing results, we once again selected drug-target pairs from the test set of BindingDB dataset that contain the aforementioned drugs and have higher predicted interaction scores, as shown in Table 5. Similarly, all drug-target pairs in Table 5 can be found in the DrugBank database, indicating that evidence of interactions for these drug-target pairs can be found in the DrugBank database. For example, DB01254 (Dasatinib) is a tyrosine kinase inhibitor that can inhibit the activity of P00519 (Tyrosine-protein kinase ABL1) and P12931 (Proto-oncogene tyrosine-protein kinase Src).

The above cases demonstrate that our proposed CAT-DTI can effectively predict potential drug-target pairs, possessing the capability to identify potential candidate drugs and thus improving the virtual screening stage of drug discovery.

Conclusion

In this work, we propose a deep learning model named CAT-DTI, which is based on cross-attention and Transformer to enhance the accuracy of predicting drug-target interactions. We employ GCN for extracting drug features, while the acquisition of protein target features uses CNN combined with Transformer, which can not only capture local features of proteins, but also take into account global context information. The introduction of the cross-attention module effectively facilitated bidirectional feature interactions between drugs and proteins, leading to the extraction of more critical DTI features. Furthermore, with the help of CDAN, our model exhibits good adaptability and predictive performance in cross-domain task, which enhances the generalization performance of CAT-DTI. Compared with other state-of-the-art models and traditional machine learning models, experimental results show that CAT-DTI improves DTI prediction performance in both in-domain and cross-domain setting, especially making promising progress in cross-domain prediction tasks.

Availibility of data and materials

The data and code can be found online at: https://github.com/ZXT0212/CAT-DTI.

References

  1. Agamah FE, Mazandu GK, Hassan R, Bope CD, Thomford NE, Ghansah A, et al. Computational/in silico methods in drug target and lead prediction. Brief Bioinf. 2020;21(5):1663–75.

    Article  Google Scholar 

  2. Zhao Q, Yang M, Cheng Z, Li Y, Wang J. Biomedical data and deep learning computational models for predicting compound-protein relations. IEEE/ACM Trans Comput Biol Bioinform. 2021;19(4):2092–110.

    Article  Google Scholar 

  3. Pan X, Lin X, Cao D, Zeng X, Yu PS, He L, et al. Deep learning for drug repurposing: methods, databases, and applications. Wires Comput Mol Sci. 2022;12(4): e1597.

    Article  CAS  Google Scholar 

  4. Hu S, Zhang C, Chen P, Gu P, Zhang J, Wang B. Predicting drug-target interactions from drug structure and protein sequence using novel convolutional neural networks. BMC Bioinf. 2019;20:1–12.

    Article  Google Scholar 

  5. Chen R, Liu X, Jin S, Lin J, Liu J. Machine learning for drug-target interaction prediction. Molecules. 2018;23(9):2208.

    Article  PubMed  PubMed Central  Google Scholar 

  6. Himmat M, Salim N, Al-Dabbagh MM, Saeed F, Ahmed A. Adapting document similarity measures for ligand-based virtual screening. Molecules. 2016;21(4):476.

    Article  PubMed  PubMed Central  Google Scholar 

  7. Sieg J, Flachsenberg F, Rarey M. In need of bias control: evaluating chemical data for machine learning in structure-based virtual screening. J Chem Inf Model. 2019;59(3):947–61.

    Article  CAS  PubMed  Google Scholar 

  8. Maia EHB, Assis LC, De Oliveira TA, Da Silva AM, Taranto AG. Structure-based virtual screening: from classical to artificial intelligence. Front Chem. 2020;8:343.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  9. Su M, Yang Q, Du Y, Feng G, Liu Z, Li Y, et al. Comparative assessment of scoring functions: the CASF-2016 update. J Chem Inf Model. 2018;59(2):895–913.

    Article  PubMed  Google Scholar 

  10. Cortes C, Vapnik V. Support-vector networks. Mach Learn. 1995;20:273–97.

    Article  Google Scholar 

  11. Breiman L. Random forests. Mach Learn. 2001;45:5–32.

    Article  Google Scholar 

  12. Ballester PJ, Mitchell JB. A machine learning approach to predicting protein-ligand binding affinity with applications to molecular docking. Bioinformatics. 2010;26(9):1169–75.

    Article  CAS  PubMed  Google Scholar 

  13. Faulon JL, Misra M, Martin S, Sale K, Sapra R. Genome scale enzyme-metabolite and drug-target interaction predictions using the signature molecular descriptor. Bioinformatics. 2008;24(2):225–33.

    Article  CAS  PubMed  Google Scholar 

  14. Wang X, Cao T, Jia CM, Tian X, Wang Y. Quantitative prediction model for affinity of drug-target interactions based on molecular vibrations and overall system of ligand-receptor. BMC Bioinf. 2021;22(1):1–18.

    Article  Google Scholar 

  15. Tian K, Shao M, Wang Y, Guan J, Zhou S. Boosting compound-protein interaction prediction by deep learning. Methods. 2016;110:64–72.

    Article  CAS  PubMed  Google Scholar 

  16. Lee I, Keum J, Nam H. DeepConv-DTI: Prediction of drug-target interactions via deep learning with convolution on protein sequences. PLoS Comput Biol. 2019;15(6): e1007129.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  17. Rogers D, Hahn M. Extended-connectivity fingerprints. J Chem Inf Model. 2010;50(5):742–54.

    Article  CAS  PubMed  Google Scholar 

  18. Zhang S, Jiang M, Wang S, Wang X, Wei Z, Li Z. SAG-DTA: prediction of drug-target affinity using self-attention graph network. Int J Mol Sci. 2021;22(16):8993.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  19. Zheng S, Li Y, Chen S, Xu J, Yang Y. Predicting drug-protein interaction using quasi-visual question answering system. Nat Mach Intell. 2020;2(2):134–40.

    Article  Google Scholar 

  20. Wei L, Zou Q, Liao M, Lu H, Zhao Y. A novel machine learning method for cytokine-receptor interaction prediction. Comb Chem High T Scr. 2016;19(2):144–52.

    CAS  Google Scholar 

  21. Wei L, Bowen Z, Zhiyong C, Gao X, Liao M. Exploring local discriminative information from evolutionary profiles for cytokine-receptor interaction prediction. Neurocomputing. 2016;217:37–45.

    Article  Google Scholar 

  22. Wei L, Long W, Wei L. Mdl-cpi: multi-view deep learning model for compound-protein interaction prediction. Methods. 2022;204:418–27.

    Article  CAS  PubMed  Google Scholar 

  23. Nguyen T, Le H, Quinn TP, Nguyen T, Le TD, Venkatesh S. GraphDTA: predicting drug-target binding affinity with graph neural networks. Bioinformatics. 2021;37(8):1140–7.

    Article  CAS  PubMed  Google Scholar 

  24. Jia J, Zhu F, Ma X, Cao ZW, Li YX, Chen YZ. Mechanisms of drug combinations: interaction and network perspectives. Nat Rev Drug Discov. 2009;8(2):111–28.

    Article  CAS  PubMed  Google Scholar 

  25. Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473. 2014;.

  26. Chen W, Chen G, Zhao L, Chen CYC. Predicting drug-target interactions with deep-embedding learning of graphs and sequences. J Phys Chem A. 2021;125(25):5633–42.

    Article  CAS  PubMed  Google Scholar 

  27. Zhao Q, Zhao H, Zheng K, Wang J. HyperAttentionDTI: improving drug-protein interaction prediction by sequence-based deep learning with attention mechanism. Bioinformatics. 2022;38(3):655–62.

    Article  CAS  PubMed  Google Scholar 

  28. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Adv Neural Inf Process Syst. 2017;30.

  29. Chen L, Tan X, Wang D, Zhong F, Liu X, Yang T, et al. TransformerCPI: improving compound-protein interaction prediction by sequence-based deep learning with self-attention mechanism and label reversal experiments. Bioinformatics. 2020;36(16):4406–14.

    Article  CAS  PubMed  Google Scholar 

  30. Huang K, Xiao C, Glass LM, Sun J. MolTrans: molecular interaction transformer for drug-target interaction prediction. Bioinformatics. 2021;37(6):830–6.

    Article  CAS  PubMed  Google Scholar 

  31. Bai P, Miljković F, John B, Lu H. Interpretable bilinear attention network with domain adaptation improves drug-target prediction. Nat Mach Intell. 2023;5(2):126–36.

    Article  Google Scholar 

  32. Shin B, Park S, Kang K, Ho JC. Self-attention based molecule representation for predicting drug-target interaction. In: Machine Learning for Healthcare Conference. PMLR; 2019. p. 230–248.

  33. Zeng Y, Chen X, Peng D, Zhang L, Huang H. Multi-scaled self-attention for drug-target interaction prediction based on multi-granularity representation. BMC bioinf. 2022;23(1):1–15.

    Article  Google Scholar 

  34. Cheng Z, Yan C, Wu FX, Wang J. Drug-target interaction prediction using multi-head self-attention and graph attention network. IEEE/ACM Trans Comput Biol Bioinf. 2021;19(4):2208–18.

    Article  Google Scholar 

  35. Bai P, Miljković F, Ge Y, Greene N, John B, Lu H. Hierarchical clustering split for low-bias evaluation of drug-target interaction prediction. In: 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE; 2021. p. 641–644.

  36. Zitnik M, Sosic R, Leskovec J. BioSNAP Datasets: Stanford biomedical network dataset collection. http://snap.stanford.edu/biodata. 2018;.

  37. Wishart DS, Knox C, Guo AC, Cheng D, Shrivastava S, Tzur D, et al. DrugBank: a knowledgebase for drugs, drug actions and drug targets. Nucleic Acids Res. 2008;36(suppl-1):D901–6.

    Article  CAS  PubMed  Google Scholar 

  38. Cao DS, Xu QS, Liang YZ. propy: a tool to generate various modes of Chou’s PseAAC. Bioinformatics. 2013;29(7):960–2.

    Article  CAS  PubMed  Google Scholar 

Download references

Acknowledgements

We are grateful to the anonymous reviewers for their constructive comments on the original manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant Nos. U22A2024, and 62271328), in part by the Shenzhen Science and Technology Program (Grant Nos. JCYJ20220818095809021, and KQTD20210811090219022).

Author information

Authors and Affiliations

Authors

Contributions

BYL and WLC supervised the study. XTZ designed the model and conducted the experiments. The manuscript was drafted by XTZ and revised by BYL with the support from WLC. All authors read and approved the final manuscript.

Corresponding authors

Correspondence to Weilin Chen or Baiying Lei.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zeng, X., Chen, W. & Lei, B. CAT-DTI: cross-attention and Transformer network with domain adaptation for drug-target interaction prediction. BMC Bioinformatics 25, 141 (2024). https://doi.org/10.1186/s12859-024-05753-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12859-024-05753-2

Keywords