Skip to main content

Semi-supervised learning improves regulatory sequence prediction with unlabeled sequences

Abstract

Motivation

Genome-wide association studies have systematically identified thousands of single nucleotide polymorphisms (SNPs) associated with complex genetic diseases. However, the majority of those SNPs were found in non-coding genomic regions, preventing the understanding of the underlying causal mechanism. Predicting molecular processes based on the DNA sequence represents a promising approach to understand the role of those non-coding SNPs. Over the past years, deep learning was successfully applied to regulatory sequence prediction using supervised learning. Supervised learning required DNA sequences associated with functional data for training, whose amount is strongly limited by the finite size of the human genome. Conversely, the amount of mammalian DNA sequences is exponentially increasing due to ongoing large sequencing projects, but without functional data in most cases.

Results

To alleviate the limitations of supervised learning, we propose a paradigm shift with semi-supervised learning, which does not only exploit labeled sequences (e.g. human genome with ChIP-seq experiment), but also unlabeled sequences available in much larger amounts (e.g. from other species without ChIP-seq experiment, such as chimpanzee). Our approach is flexible and can be plugged into any neural architecture including shallow and deep networks, and shows strong predictive performance improvements compared to supervised learning in most cases (up to \(70\%\)).

Availability and implementation

https://forgemia.inra.fr/raphael.mourad/deepgnn.

Peer Review reports

Introduction

Complex diseases are caused by the combination of multiple mutations with lifestyle and environmental factors [1]. Complex diseases are frequent in the population and include Alzheimer’s disease, autoimmune diseases, asthma and some cancers [2,3,4]. Over the past decade, genome-wide association studies (GWASs) have systematically identified thousands of single nucleotide polymorphisms (SNPs) associated with these diseases [1]. However, an important limitation of GWASs comes from the difficulty of gaining insight into the underlying biological mechanism, since over 95\(\%\) of associated SNPs are located outside coding sequences [5]. Interestingly, the majority of the non-coding SNPs (75\(\%\)) are located within regulatory elements [5], suggesting an important role for these SNPs in the deregulation of gene expression.

Deep neural networks have achieved state-of-the-art performance for the prediction of regulatory elements from DNA sequences. The first models were based on the stacking of one or few convolutional layer(s) followed by a dense network, thereby capturing combinations of DNA motifs specific to DNA binding proteins [6, 7]. However, these initial models could not model long-range relations among DNA motifs within the sequence, and thus missed the complex grammar of the regulatory genome. To tackle this issue, several approaches were proposed such as recurrent neural networks with LSTM for few hundred bases [8], dilated convolutions [9] and transformers [10] for hundreds of kilobases. These models were trained using labeled data (supervised learning), but recently new models were trained without labeled data (unsupervised learning) [11, 12].

Semi-supervised learning trains a model using not only labeled data generally available in small amount, but also using unlabeled data often available in large amount [13]. Graph neural networks (GNNs) were recently proposed for semi-supervised learning graph-structured data that is based on an efficient variant of convolutional neural networks which operate directly on graphs [14]. Instead of training individual embeddings for each node of the graph, a GNN aggregates features from a node’s local neighborhood, thereby borrowing information from unlabeled nodes. In the simplest GNN, the graph convolutional network (GCN), the aggregation simply consists in a linear combination from a node’s local neighborhood embedding [14]. Many improvements of the GCN were later proposed, including approximate personalized propagation of neural predictions (APPNP) with an improved propagation scheme based on personalized PageRank [15], GraphSAGE which concatenates the aggregation and the original node’s embeddings [16] and Graph Attention layer (GAT) [17] where edges have weights computed by an attention mechanism.

Here, we propose a novel framework for regulatory sequence prediction using semi-supervised learning with graph neural network. For this purpose, neural networks are trained not only from labeled sequences (e.g. human genome with ChIP-seq experiment), but also from unlabeled sequences (from other species without ChIP-seq experiment, e.g. chimpanzee, rabbit or dog). In order to incorporate unlabeled sequences in the training, a graph neural network connecting homologous sequences is used. Compared to supervised learning, the proposed semi-supervised learning allows to train models from a much larger number of sequences, without needing additional experiments since many unlabeled (unannotated) genomes are already available, while the vast majority of functional experiments such as ChIP-seq are only available in human, or to some extent in mouse.

Materials and methods

Human experimental data

We used publicly available CTCF, ESR1, POL2, H3K4me3 ChIP-seq data and input data of human lymphoblastoid GM12878 from Gene Expression Omnibus (GEO) accession GSE31477 and GSE170139 from ENCODE [18]. We used publicly available ATAC-seq data of lymphoblastoid GM12878 from Gene Expression Omnibus (GEO) accession GSE170918 from ENCODE [18]. All the data was mapped on hg38 and peak calling was done using macs3 [19].

Mouse experimental data

We used publicly available CTCF, POL2 and H3K4me3 ChIP-seq data and input data of mouse lymphoblastoid CH12 (GM12878 analog) from Gene Expression Omnibus (GEO) accession GSE49847 from Mouse ENCODE [20]. All the data was mapped on mm10 and peak calling was done using macs3 [19].

Labeled DNA sequences

We binned the labeled genome (e.g. human genome assembly hg38) into non-overlapping genomic intervals of 200 b, where a given bin was considered as either bound by a TF (if overlapping \(> 50\%\)), or not bound otherwise (classification setting). For each 200 b bin, we extracted the surrounding DNA sequence of 1 kb (containing the bin and the context).

Homologous DNA sequences

Given a labeled bin sequence (e.g. from hg38), the homologous sequences in other species’ genomes (e.g. from panTro6, oryCun2 and canFam3) were mapped using the following approach. We first mapped the labeled bin to another species genome using liftover. If a labeled bin mapped to different loci in another species genome but the loci were close to each other (separated by less than 50 b), the different loci were merged into one loci. Then the DNA sequence of 1 kb surrounding the loci was extracted and considered as a homologous sequence. If the bin mapped to different and distant loci (\(>50\) b) or if the bin did not map to any loci, then no homologous sequence was extracted (the homologous sequence was filled with N’s).

To search for homologous sequences, we used the following mammalian genomes: Rhesus monkey (rheMac10), marmoset (calJac3), chimpanzee (panTro6), pygmy chimpanzee (panPan2), Sumatran orangutan (ponAbe3), gorilla (gorGor6), olive baboon (papAnu4), crab-eating macaque (macFas5), Bolivian squirrel monkey (saiBol1), northern white-cheeked gibbon (nomLeu3), gray mouse lemur (micMur2), small-eared galago (otoGar3), mouse (mm10), rat (rn7), ferret (musFur1), rabbit (oryCun2), pork (susScr11), cat (felCat9), dog (canFam3), horse (equCab3), cow (bosTau9) and opossum (monDom5).

Model architecture

Fig. 1
figure 1

Sketch of the proposed semi-supervised model. A A convolutional network within a graph neural network (so called CNN-GNN). B Comparison with the classical convolutional network (CNN)

The proposed model is illustrated in Fig. 1. The model takes two inputs: (1) a 1 kb labeled DNA sequence from one species (for instance, human) and the corresponding homologous sequences from other species (for instance, chimpanzee, mouse and dog), and (2) a graph matrix connecting the homologous sequences from the different species. As output, in a classification setting, the model predicts the probability of the labeled DNA sequence to bind a given TF within a 200 b DNA sequence centered on the 1 kb DNA sequence. Alternatively, in a regression setting the model predicts the ChIP-seq signal.

As done in state-of-the-art [21], we first model the DNA sequences using a convolutional network (left side of the network, Fig. 1A). The convolutional neural network (CNN) can be composed of only one convolutional layer (shallow convolutional network). Alternatively, the convolutional network can be composed of a convolutional layer followed by dilated convolutions with residual connections (deep convolutional network), providing better performances in most situations by modeling long-range relations. Each convolutional layer comprises 64 filters (also called kernels). In the first layer, filters of size = 24 were used. In the next layers (present in the deep convolutional network), filters of size = 3 with an increasing dilation rate of 2 for each additional layer were used instead. The convolutional network is followed by a global maximum pooling and a dense layer of 10 neurons. If graph convolution was not used (classical CNN), then a last layer of 1 neuron was used to predict the outcome. Alternatively, graph convolution was used as described in the paragraph below.

In parallel, the model takes as a secondary input the graph matrix connecting homologous sequences between species (right side of the network, Fig. 1A). Depending on the graph convolution used, the graph matrix can be preprocessed. For instance, for the GCN, APPNP or GSC convolutions, the graph matrix is preprocessed as follows:

$$\begin{aligned} A^* = \hat{D}^{-1/2} \hat{A} \hat{D}^{-1/2}, \end{aligned}$$
(1)

where \(\hat{A} = A + I\) is the adjacency matrix with added self-loops (I) and \(\hat{D}\) is its degree matrix. For the GraphSAGE or GAT convolutions, no graph matrix preprocessing is needed. Lastly, a graph convolution is used to convolve the activation values from the dense layer over the sequence homology matrix producing sequence output with 1 neuron. We chose the GraphSAGE convolution for our model as providing the best results, although we compared different graph convolutions in the Section Results and Discussion, Subsection Comparison with the baseline model. For the GraphSAGE, graph convolution was done over the dense layer of 10 neurons as follows:

$$\begin{aligned} {\textbf {H}} = \big [ \text {AGGREGATE}({\textbf {X}}) || {\textbf {X}} \big ] {\textbf {W}} + {\textbf {b}}, \end{aligned}$$
(2)

where \(\text {AGGREGATE}\) aggregates a node’s neighborhood by summing, i.e. it aggregates the features over the different homologous sequences. A last layer of 1 neuron then combines the aggregated features and predicts the outcome. For classification, a sigmoid activation is used, while for regression, a softplus activation is used. Because the proposed model combined a GNN on the top of a CNN, we called it “CNN-GNN”.

An important aspect of the CNN-GNN is that parameters are shared before the graph convolution layer, i.e. parameters are shared for the convolution layers (and for the dense layer) when training on the multiple input sequences (labeled and homologous sequences). Parameter sharing allows to keep the number of parameters to train almost the same between the CNN-GNN and the standard CNN without graph convolution. In fact, only very few parameters are dedicated to graph convolution (for GraphSAGE: only 10 additional parameters).

We compared the CNN-GNN with the baseline CNN where graph convolution was not implemented (Fig. 1B). The baseline CNN takes as an input only a labeled sequence and no homologous sequences.

Model training

Model training was done using the following hyper-parameters: ADAM optimizer, a minimum of 5 epochs, a maximum of 30 epochs, a patience of 3 epochs, a learning rate = 0.0005, a clipping norm = 0.0001, a batch size = 100. For classification, binary cross-entropy was used as loss, whereas for regression, the Poisson loss was used instead. The models were trained with 100 thousand sequences in balanced manner: 50 thousand bin sequences corresponding to observed ChIP-seq peaks (or other experimental peaks) and 50 thousand randomly drawn bin sequences without ChIP-seq peaks. When the number of peaks was not enough to obtain 50 thousand bin sequences, upsampling was used to reach 50 thousands.

Prediction of SNP effect

We used the deep learning models to predict the effect of SNPs on functional data such as protein binding or chromatin accessibility. To compute the effect of a SNP, we used the following approach for classification as in [6]. First, model predictions were computed both for the DNA sequence comprising the reference SNP allele (\(p_{ref}\)) and the same DNA sequence but with the alternative SNP allele (\(p_{alt}\)). Then, the SNP effect was computed as \((p_{alt}-p_{ref}) \cdot max(0,p_{alt},p_{ref})\), where \(p_{alt}\) is the prediction from the alternative allele and \(p_{ref}\) the prediction from the reference allele.

To predict SNP effect with the CNN-GNN, graph convolution was not used for prediction (only used for training), since it is not possible to accurately know the alternative and reference alleles in homologous sequences from other species. Hence, the sequence homology matrix values were set up to zeroes, except for the labeled sequence and itself (first row and first column of the matrix). Then the SNP effect was computed as detailed in the previous paragraph.

Implementation and availability

The model was developed using Tensorflow and Keras. It is available at the Github repository: https://forgemia.inra.fr/raphael.mourad/deepgnn

Results and discussion

Conservation of protein binding sites in related species

Fig. 2
figure 2

Conservation of CTCF DNA binding sites as predicted from the CTCF motif

The working hypothesis of the proposed semi-supervised model is the following: transcription factor binding sites as encoded by a DNA motif (or multiple motifs) are evolutionary conserved, meaning that the position of a DNA motif in a genome is conserved in evolutionary close species. This hypothesis implies that DNA sequences in unlabeled evolutionary close genomes could be used to improve predictions.

To verify this working hypothesis, we assessed whether the CTCF binding sites (as encoded by the CTCF DNA motif MA00138 from Jaspar database) in the human genome are conserved in evolutionary close genomes. For this purpose, we binned the human genome (hg38 assembly) into 200 b bins. Then, we liftovered the human genome bins to other mammalian genomes to obtain homologous bins as detailed in Subsection Materials and Methods, Homologous DNA sequences. Then, for each bin of the human and the corresponding homologous bins in the other mammalian genomes, we scanned the DNA sequences for the presence of a CTCF motif. We illustrated the results with a genomic loci of the ERCC4 gene in Fig. 2. In this loci, three CTCF binding peaks were found in the human genome (hg38). For each peak, a CTCF binding motif was found in the human genome. By looking at other mammalian genomes, we found that, for each human peak, the CTCF binding motif was also present in apes (chimpanzee: panTro6 and gorilla: gorGor6). By looking at more distantly related species, we still could find the DNA motif for some peaks. For instance, for the peak at the left side, the DNA motif was also conserved in the macaque (rheMac10), the mouse (mm10), the pig (susScr11) and the cow (bosTau9), but absent in the dog (canFam3), the cat (felCat9) and the rabbit (oryCun2).

We concluded that using unlabeled genomes could be used to improve the prediction of regulatory sequences by account for the homology of sequences between related genomes since DNA motif position is often conserved.

Prediction of regulatory elements

The proposed model takes as input sets of homologous 1 kb DNA sequences between species and a graph of homologies between the DNA sequences encoded as a binary matrix. The model then predicts the binding of a given TF within a 200 b DNA sequence centered on the 1 kb DNA sequence. The model can predict the binding as a classification problem (whether bin overlaps a binding peak) or as a regression problem (the ChIP-seq coverage signal).

We binned the human genome into bins of 200 b, considered as either bound by a TF (if overlapping \(> 50\%\)), or not bound otherwise. For each 200 b bin, we extracted the surrounding DNA sequence of 1 kb (containing the bin and the context). We also mapped homologous DNA sequences in other mammal genomes using the liftover program (see more details in Subsection Materials and Methods, Homologous DNA sequences). As training and validation data, we used bins from chromosomes 1–15, and as testing data, we used bins from chromosomes 16–22 and X.

Comparison with the baseline model

Fig. 3
figure 3

Comparison of prediction performances between the semi-supervised model (here called CNN–GNN) with the baseline model (CNN) where graph convolution was not used. Ten trainings were done for each model to make boxplots. A Comparison of shallow CNN and shallow CNN–GNN in a classification setting, in term of area under the roc curve (AUROC). B Comparison of shallow CNN and shallow CNN-GNN in a classification setting, in term of area under the precision recall curve (AUPR). C Comparison of shallow CNN and shallow CNN–GNN in a regression setting, in term of Pearson correlation. D Comparison of shallow CNN and shallow CNN–GNN in a classification setting for ATAC peaks, H3K4me3 peaks and POL2 peaks. E Comparison of AUROC between deep CNN and deep CNN–GNN in a classification setting. F Comparison of AUPR between deep CNN and deep CNN–GNN in a classification setting. G Comparison of Pearson correlation between deep CNN and deep CNN–GNN in a regression setting

We then comprehensively compared the proposed semi-supervised model (CNN–GNN) with the baseline model (CNN) where graph convolution was not used. For a fair comparison between the two models, no fine-tuning was used to favor any of the two models, and the same hyperparameters were used for both models. We first built a semi-supervised model with only one convolutional layer (shallow CNN), with the equivalent baseline model. We trained the models to predict the binding of CTCF from ChIP-seq data as a binary classification setting (peak versus no peak). In term of area under the curve (AUROC), we found a moderate increase with the CNN-GNN when trained with sequences from 23 mammalian genomes (1 labeled genome plus 22 unlabeled genomes) compared to the CNN (from 0.955 to 0.958, \(+0.3\%\), Fig. 3A). With AUROC \(> 0.95\), one would think that predicting CTCF peaks is very accurate. However, genomic data are highly imbalanced with few positive labels (peak presence) compared to negative labels (peak absence), which makes the AUROC a less appropriate metric to evaluate prediction performance. When we instead compared the area under the precision-recall curve (AUPR), a better metric to account for imbalanced data, we instead observed a strong increase for the CNN–GNN compared to the baseline CNN (from 0.224 to 0.272, +\(20.8\%\), corresponding to a raw increase of 0.047, between CNN–GNN with 23 mammals and baseline CNN, Fig. 3B). Interestingly, we observed that the AUPR monotonically increased with the number of mammalian genomes used in the CNN-GNN, demonstrating that the higher number of mammalian genomes included, the better prediction. In a regression setting, where the models were trained to predict the CTCF ChIP-seq signal within a bin, we also found a strong increase of Pearson correlation for the CNN-GNN with 23 mammals compared to the baseline CNN (from 0.296 to 0.358, \(+20.9\%\), corresponding to a raw increase of 0.062; Fig. 3C). We didn’t find that the correlation was driven by outliers (Additional file 1: Fig S1).

We then assessed prediction performance with other types of data, including ATAC-seq, H3K4me3 and POL2 ChIP-seq data (Fig. 3D). For ATAC-seq, AUROC was moderately increased (from 0.878 to 0.898, \(+2.3\%\)), while AUPR was strongly increased (from 0.183 to 0.239, \(+30.5\%\)). For H3K4me3, AUROC was increased from 0.941 to 0.961 (\(+2.2\%\)), while a very important increase of AUPR was observed (from 0.213 to 0.301, \(+41.8\%\)). For POL2 ChIP-seq, AUROC was moderately increased (from 0.899 to 0.912, \(+1.4\%\)), while AUPR was strongly increased (from 0.124 to 0.203, \(+63.5\%\)). We also assessed transcription factors NANOG, OCT4 and SOX2 that maintain pluripotency of embryonic stem cells and whose binding sites might not be well conserved across species. For those TFs, there were slight improvements of AUROC for NANOG (from 0.832 to 0.841, \(+1.2\%\)), for OCT4 (from 0.880 to 0.897, \(+1.9\%\)), and for SOX2 (from 0.900 to 0.908, \(+1\%\)). There was no improvement of AUPR for NANOG (from 0.060 to 0.055, \(-7.5\%\)), but improvements for OCT4 (from 0.109 to 0.118, \(+8.2\%\)) and for SOX2 (0.031 to 0.032, \(+2.9\%\)) (Additional file 2: Fig S2).

We also sought to assess if our semi-supervised approach could improve the performance for other CNN architectures, for instance when plugged on the top of a deeper CNN architecture. For this purpose, we used a deeper CNN architecture where one convolutional layer was followed by 10 dilated convolutional layers with residual connections similar to [9]. Similarly to the shallow CNN, the CNN–GNN did not improve significantly the classification performance in terms of AUROC (from 0.983 to 0.984, \(+0.1\%\); Fig. 3E). However, in term of AUPR, the CNN-GNN improved the CNN (from 0.544 to 0.597, \(+9.9\%\); Fig. 3F). In the regression setting, the CNN-GNN greatly improved the CNN (from 0.541 to 0.631, \(+16.6\%\); Fig. 3G).

Different graph neural network convolutions were compared with our model implementing a GraphSAGE convolution (Additional file 3: Fig S3). In term of AUROC, we found that the best performances were obtained on average with GraphSAGE (0.958), compared to those obtained with GCN (0.953), GAT (0.955), APPNP (0.957) and GSC (0.957). In term of AUPR, we also found the best performances on average with GraphSAGE (0.272), compared to GCN (0.259), GAT (0.257), APPNP (0.261) and GSC (0.261).

We thus concluded that our semi-supervised approach borrowing information from unlabeled genomes could greatly improve the performance of predictions in many different situations, including classification and regression, shallow and deep CNNs, and transcription factor binding (CTCF), chromatin accessibility (ATAC), histone mark (H3K4me3) and polymerase II binding (POL2).

Cross-species prediction

Fig. 4
figure 4

Comparison of prediction performance of our semi-supervised model trained from the human labeled genome and 9 unlabeled genomes to predict mouse experiments, as compared to the baseline model. A CTCF classification with shallow CNN. Area under the ROC curve (AUROC) and area under the PR curve (AUPR) are plotted. B CTCF classification with deep CNN. C H3K4me3 classification with shallow CNN. D POL2 classification with shallow CNN

Previous studies have shown that the grammar of regulatory sequences can be transferred across species, for instance between human and mouse, to make accurate tissue-specific predictions [21]. In fact, cross-species prediction could be used to annotate the genome of many mammalian species for which no functional data are available, but for which a large body of human data is available. Hence, we explored if our CNN–GNN could improve predictions across species, compared to the baseline CNN. For this purpose, we first trained our model on human CTCF ChIP-seq data and predicted on the mouse genome. In a classification setting, we found that a shallow CNN–GNN trained from human labeled genome and 9 unlabeled genomes slightly improved AUROC (from 0.965 to 0.967, \(+0.3\%\)), but strongly increased AUPR (from 0.169 to 0.211, \(+24.7\%\)) (Fig. 4A). Similarly, a deep CNN-GNN slightly improved AUROC (from 0.991 to 0.992, \(+0.1\%\)), but significantly increased AUPR (from 0.508 to 0.588, \(+15.8\%\)) (Fig. 4B). We also explored other functional data. For H3K4me3 classification, we found that a shallow CNN–GNN improved AUROC (from 0.893 to 0.918, \(+2.8\%\)), but significantly increased AUPR (from 0.274 to 0.321, \(+17.1\%\)) (Fig. 4C). For POL2 classification, we found that a shallow CNN-GNN slightly improved both AUROC (from 0.866 to 0.878, \(+1.3\%\)) and AUPR (from 0.218 to 0.242, \(+10.1\%\)) (Fig. 4D).

Results thus showed the superiority of our semi-supervised approach as compared to supervised learning for cross-species predictions in mammals for both shallow and deep CNNs, and for transcription factor binding (CTCF), histone mark (H3K4me3) and polymerase II binding (POL2).

Multigenome training compared to semi-supervised learning

Fig. 5
figure 5

Comparison of prediction performance of our semi-supervised model trained from labeled human genome and 22 unlabeled mammalian genomes with the baseline model trained from human and mouse labeled genomes (also called multi-genome training [21]), and the baseline model trained only from human (one-genome training). A Area under the ROC curve (AUROC) from shallow CNNs. The baseline CNN trained on labeled human data (hg38), the baseline CNN trained on both labeled human and mouse data (hg38 + mm10), and the CNN-GNN trained on 10 or 23 mammalian genomes were compared. B Area under the PR curve (AUPR) from shallow CNNs. C Area under the ROC curve (AUROC) from deep CNNs. D Area under the PR curve (AUPR) from deep CNNs

Multi-genome training was previously proposed to improve the training of deep learning models [21]. Multi-genome training consists in training a model not only from one labeled genome (for instance from human genome), but from multiple labeled genomes (for instance from the human and the mouse genomes at the same time) followed by fine-tuning on the target labeled genome. Multi-genome training is a simple approach to improve predictive performance, but it is limited by the availability of functional data from multiple genomes. For mammals, most of the functional data are available in human, and to a certain extent in mouse. In comparison, our semi-supervised learning instead allows to exploit many more genomes, e.g. all available mammalian genomes sequenced.

We compared multi-genome training (MGT) with our semi-supervised learning (SSL) and one-genome training (OGT) for both shallow and deep CNNs (Fig. 5). For shallow CNNs, we observed a slight improvement of MGT compared to the OGT in term of AUROC (from 0.955 to 0.957, \(+0.2\%\); Fig. 5A), also for SSL with 23 mammals compared to the OGT (from 0.955 to 0.958, \(+0.3\%\)). However, in term of AUPR, we found a slight improvement for MGT (from 0.224 to 0.230, \(+2.5\%\); Fig. 5B), but a strong increase for SSL with 23 mammalian genomes as compared to OGT (from 0.224 to 0.272, \(+20.8\%\)). For deep CNNs, we observed an improvement of MGT compared to the OGT in term of AUROC (0.983 to 0.985, \(+2.4\%\); Fig. 5C), and almost no improvement of SSLwith 23 mammalian genomes compared to OGT (from 0.983 to 0.984, \(+0.1\%\)). In term of AUPR, we found an increase for MGT compared to OGT (from 0.543 to 0.572, \(+5.3\%\); Fig. 5D), and a stronger increase for SSL compared to OGT (from 0.544 to 0.597, \(+9.9\%\)).

These results thus revealed that semi-supervised learning with more than 20 unlabeled genomes provides similar or better performances than multi-genome training with two labeled genomes, while the former does not necessitate any additional functional data as compared to the latter.

Prediction of SNP effects

Fig. 6
figure 6

SNP effect prediction with our semi-supervised model trained from human labeled genome and 22 unlabeled genomes, compared to the baseline model. Allelic imbalance ChIP-seq SNPs were downloaded from ADASTRA database [22]. A Scatter plot of predicted CTCF binding SNP effect versus observed effect with a deep CNN. B Scatter plot of predicted CTCF binding SNP effect versus observed effect with a deep CNN–GNN. C Comparison of Pearson correlation between predicted and observed CTCF binding SNP effects for CNN and CNN–GNN, for 10 different model trainings. D Scatter plot of predicted ESR1 binding SNP effect versus observed effect with a deep CNN. E Scatter plot of predicted ESR1 binding SNP effect versus observed effect with a deep CNN–GNN. F Comparison of Pearson correlation between predicted and observed CTCF binding SNP effects for CNN and CNN–GNN, for 10 different model trainings

An important application of deep learning models in genomics is to predict the impact of a SNP on functional data. For instance, studies have shown that deep learning can predict the impact of a SNP on gene expression, histone modification, or protein binding [6, 7, 10]. Hence, we explored the ability of our CNN-GNN to improve the prediction of SNP effect on functional data, and compared it to the baseline CNN. For this purpose, we first trained a deep CNN for CTCF peak classification, and then predicted the impact of a SNP on CTCF binding, as done in [7]. We found a good Pearson correlation between the observed effect as estimated by ChIP-seq allelic imbalance (from ADASTRA database [22]), and the predicted effect (R \(=0.312\); Fig. 6A). We then trained a deep CNN-GNN and observed an increase of Pearson correlation (R \(=0.458\), \(+46.8\%\); Fig. 6B). Repeating the same experiment with different model training runs led to the same conclusion (p = \(9 \times 10^{-11}\); Fig. 6C). We also trained a deep CNN for ESR1 peak classification, and then predicted the impact of a SNP on ESR1 binding. We found a moderate Pearson correlation between the observed effect and the predicted effect (R \(= 0.150\); Fig. 6D). With the CNN-GNN, we observed an increase of Pearson correlation (R \(= 0.256\), \(+70.7\%\); Fig. 6E). Repeating the same experiment also showed a significant difference (p = 0.03; Fig. 6F).

We thus found that our semi-supervised learning could improve the prediction of SNP effect on functional data, and could thus be used to prioritize SNPs in clinical studies.

Conclusion

In this article, we propose a novel framework for regulatory sequence prediction using semi-supervised learning with graph neural network. For this purpose, neural network training does not only exploit labeled sequences (e.g. human genome with ChIP-seq experiment), but also unlabeled sequences available in much larger amounts (e.g. from other species without ChIP-seq experiment, such as chimpanzee). For this purpose, a graph neural network is built on the top of a classical convolutional neural network in order to propagate feature information from homologous unlabeled sequences. The framework is flexible and can be plugged into any neural architecture including shallow and deep networks, for which the framework shows predictive performance improvements in both classification and regression tasks. Moreover, improvements were found not only for CTCF ChIP-seq whose DNA motif is known to be conserved among mammals, but surprisingly also for ATAC-seq, H3K4me3 ChIP-seq and POL2 ChIP-seq, which clearly demonstrates the versatility of our approach. Globally, we found strong predictive performance improvements compared to supervised learning in most cases (up to \(70\%\)), when using 22 unlabeled mammalian genomes. Including more unlabelled genomes (hundreds or even thousands) will likely bring even stronger improvements, and will thus represent a very appealing avenue for future model improvements.

A strong drawback of supervised learning for DNA sequence prediction in human and most species is the limit of the genome size, e.g. 3.3 Gb in human. Of course, accounting for variants can increase the amount of DNA sequences, but still represent only a small amount of DNA sequence variation. Multi-genome training was previously proposed to exploit labeled sequences from multiple genomes during model training, allowing in a simple manner to train from a larger DNA sequence space [21]. While such approach shows good performance improvements, it requires the availability of many labeled genomes for a given functional experiment, which are most of the time not available. Semi-supervised learning further extend this philosophy by exploiting during training DNA sequences from genomes that are not even labeled, which represents the majority of genomes. Semi-supervised learning was successfully applied for cross-species prediction, allowing the annotation of genomes in species without functional data. Moreover, we found that such approach could improve the prediction of the impact of SNPs and thus represent an interesting avenue for human genetic research.

Another approach to leverage unlabeled data is the use of self-supervised learning as a pre-training with large language models (LLMs). DNABERT was proposed as an adaptation of the BERT (Bidirectional Encoder Representations from Transformers) model for pre-training before DNA sequence classification [12]. More recently, a self-supervised model was shown to learn gene structure and DNA motifs without any supervision [23]. Compared to self-supervised learning, semi-supervised learning has the advantage of borrowing information of labeled sequenced during training allowing to guide the training of features, and does not necessitate to build models with a very large number of parameters as LLMs.

There are several limitations of the proposed approach. First, the use of our approach is likely limited to species which are not too evolutionary distant. For instance, it is very unlikely that including plant genomes would help to predict regulatory sequences from a mammalian genome. Second, the interpretability of the low-level features from the CNN (e.g. DNA motifs) is likely to be affected by the GNN, since the features are learned not only from one species (labeled sequence) but also from other species (from the homologous sequences), and therefore reflect “consensus” DNA motifs between different genomes. Third, we did not try to plug our GNN on the top of other networks such as the transformer network, and thus do not know whether our approach could bring improvement. Fourth, our approach provides improvement without any additional functional data in homologous species, but as a drawback it requires higher computational resources. Fifth, an improvement of our GNN would be to combine semi-supervised learning from unlabeled genomes with multi-genome training on multiple labeled genomes (for instance, both human and mouse genomes). Sixth, the homology matrix was computed before model training by liftover between sequences. An improvement would be to infer the homology matrix from the sequence embedding itself during training. However, our initial result with the Graph Attention layer which computes an attention weight from the embedding did not yield improvements. Further analyses must be required to show the superiority of the graph attention mechanism.

Availability of data and materials

All the code and data are available at: https://forgemia.inra.fr/raphael.mourad/deepgnn

References

  1. Visscher PM, Wray NR, Zhang Q, Sklar P, McCarthy MI, Brown MA, Yang J. 10 years of GWAS discovery: biology, function, and translation. Am J Human Genet. 2017;101(1):5–22.

    Article  CAS  Google Scholar 

  2. Dorn GW, Cresci S. Genome-wide association studies of coronary artery disease and heart failure: where are we going? Pharmacogenomics. 2009;10(2):213–23 (PMID: 19207022).

    Article  CAS  PubMed  Google Scholar 

  3. Billings LK, Florez JC. The genetics of type 2 diabetes: What have we learned from GWAS? Ann N Y Acad Sci. 2010;1212(1):59–77.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  4. Collins AL, Sullivan PF. Genome-wide association studies in psychiatry: What have we learned? Br J Psychiatry. 2013;202(1):1–4.

    Article  PubMed  PubMed Central  Google Scholar 

  5. Maurano MT, Humbert R, Rynes E, Thurman RE, Haugen E, Wang H, Reynolds AP, Sandstrom R, Qu H, Brody J, Shafer A, Neri F, Lee K, Kutyavin T, Stehling-Sun S, Johnson AK, Canfield TK, Giste E, Diegel M, Bates D, Hansen RS, Neph S, Sabo PJ, Heimfeld S, Raubitschek A, Ziegler S, Cotsapas C, Sotoodehnia N, Glass I, Sunyaev SR, Kaul R, Stamatoyannopoulos JA. Systematic localization of common disease-associated variation in regulatory DNA. Science. 2012;337(6099):1190–5.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  6. Alipanahi B, Delong A, Weirauch MT, Frey BJ. Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning. Nat Biotechnol. 2015;33:831–8.

    Article  CAS  PubMed  Google Scholar 

  7. Zhou J, Troyanskaya OG. Predicting effects of noncoding variants with deep learning-based sequence model. Nat Methods. 2015;12(10):931–4.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  8. Quang D, Xie X. DanQ: a hybrid convolutional and recurrent deep neural network for quantifying the function of DNA sequences. Nucleic Acids Res. 2016;44(11):e107–e107.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Kelley DR, Reshef YA, Bileschi M, Belanger D, McLean CY, Snoek J. Sequential regulatory activity prediction across chromosomes with convolutional neural networks. Genome Res. 2018;28(5):739–50.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  10. Avsec Ž, Agarwal V, Visentin D, Ledsam JR, Grabska-Barwinska A, Taylor KR, Assael Y, Jumper J, Kohli P, Kelley DR. Effective gene expression prediction from sequence by integrating long-range interactions. Nat Methods. 2021;18(10):1196–203.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  11. Ng P. dna2vec: Consistent vector representations of variable-length k-mers. 2017.

  12. Ji Y, Zhou Z, Liu H, Davuluri RV. DNABERT: pre-trained bidirectional encoder representations from transformers model for DNA-language in genome. Bioinformatics. 2021;37(15):2112–20.

    Article  CAS  PubMed  Google Scholar 

  13. Zhu X, Goldberg AB. Introduction to semi-supervised learning, Synthesis Lectures on Artificial Intelligence and Machine Learning. Morgan & Claypool Publishers, 2009.

  14. Kipf TN, Welling M. Semi-supervised classification with graph convolutional networks. 2016. CoRR, arXiv:abs/1609.02907.

  15. Klicpera J, Bojchevski A, Günnemann S. Personalized embedding propagation: combining neural networks on graphs with personalized PageRank. 2018. CoRR arXiv:abs/1810.05997.

  16. Hamilton WL, Ying R, Leskovec J. Inductive representation learning on large graphs. 2018.

  17. Velickovic P, Cucurull G, Casanova A, Romero A, Lio P, Bengio Y. Graph attention networks. 2018.

  18. The ENCODE Consortium. An integrated encyclopedia of DNA elements in the human genome. Nature. 2012;489(7414):57–74.

    Article  Google Scholar 

  19. Zhang Y, Liu T, Meyer CA, Eeckhoute J, Johnson DS, Bernstein BE, Nusbaum C, Myers RM, Brown M, Li W, Liu XS. Model-based Analysis of ChIP-Seq (MACS). Genome Biol. 2008;9(9):R137.

    Article  PubMed  PubMed Central  Google Scholar 

  20. Yue F, Cheng Y, Breschi A, Vierstra J, Wu W, Ryba T, Sandstrom R, Ma Z, Davis C, Pope BD, Shen Y, Pervouchine DD, Djebali S, Thurman RE, Kaul R, Rynes E, Kirilusha A, Marinov GK, Williams BA, Trout D, Amrhein H, Fisher-Aylor K, Antoshechkin I, DeSalvo G, See L-H, Fastuca M, Drenkow J, Zaleski C, Dobin A, Prieto P, Lagarde J, Bussotti G, Tanzer A, Denas O, Li K, Bender MA, Zhang M, Byron R, Groudine MT, McCleary D, Pham L, Ye Z, Kuan S, Edsall L, Wu Y-C, Rasmussen MD, Bansal MS, Kellis M, Keller CA, Morrissey CS, Mishra T, Jain D, Dogan N, Harris RS, Cayting P, Kawli T, Boyle AP, Euskirchen G, Kundaje A, Lin S, Lin Y, Jansen C, Malladi VS, Cline MS, Erickson DT, Kirkup VM, Learned K, Sloan CA, Rosenbloom KR, Lacerda de Sousa B, Beal K, Pignatelli M, Flicek P, Lian J, Kahveci T, Lee D, James Kent W, Ramalho Santos M, Herrero J, Notredame C, Johnson A, Vong S, Lee K, Bates D, Neri F, Diegel M, Canfield T, Sabo PJ, Wilken MS, Reh TA, Giste E, Shafer A, Kutyavin T, Haugen E, Dunn D, Reynolds AP, Neph S, Humbert R, Scott Hansen R, De Bruijn M, Selleri L, Rudensky A, Josefowicz S, Samstein R, Eichler EE, Orkin SH, Levasseur D, Papayannopoulou T, Chang K-H, Skoultchi A, Gosh S, Disteche C, Treuting P, Wang Y, Weiss MJ, Blobel GA, Cao X, Zhong S, Wang T, Good PJ, Lowdon RF, Adams LB, Zhou X-Q, Pazin MJ, Feingold EA, Wold B, Taylor J, Mortazavi A, Weissman SM, Stamatoyannopoulos JA, Snyder MP, Guigo R, Gingeras TR, Gilbert DM, Hardison RC, Beer MA, Ren B, and Consortium TME. A comparative encyclopedia of DNA elements in the mouse genome. Nature. 2014;515(7527):355–64.

  21. KD R. Cross-species regulatory sequence activity prediction. PLOS Comput Biol. 2020;16(7): e1008050.

    Article  Google Scholar 

  22. Abramov S, Boytsov A, Bykova D, Penzar DD, Yevshin I, Kolmykov SK, Fridman MV, Favorov AV, Vorontsov IE, Baulin E, Kolpakov F, Makeev VJ, Kulakovskiy IV. Landscape of allele-specific transcription factor binding in the human genome. Nat Commun. 2021;12(1):2751.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  23. Benegas G, Batra SS, Song YS. DNA language models are powerful zero-shot predictors of non-coding variant effects. 2022.

Download references

Acknowledgements

The author is grateful to Nathalie Vialaneix and Céline Brouard, and the SaAB team from INRAE MIAT lab.

Funding

This work was supported by INRAE MIAT visiting professor fellowship and University of Toulouse.

Author information

Authors and Affiliations

Authors

Contributions

RM conceived and designed the project. RM implemented the model and analyzed the data. RM wrote the manuscript. The authors read and approved the final manuscript.

Corresponding author

Correspondence to Raphaël Mourad.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The author declares that he has no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1: Fig S1

. Scatterplots of observed CTCF ChIP-seq signal compared to predicted signal by semi-supervised model (here called CNN-GNN) and by the baseline model (CNN) where graph convolution was not used.

Additional file 2: Fig S2

. Comparison of prediction performances between the semi-supervised (here called CNN-GNN) with the baseline model (CNN) where graph convolution was not used for NANOG, OCT4 and SOX2 peak classifications. Ten trainings were done for each model to make boxplots. Area under roc curve (AUROC). Area under the precision recall curve (AUPR).

Additional file 3: Fig S3

. Comparison of prediction performances between different graph neural network convolution layers. A) Comparison in term of area under the roc curve (AUROC). B) Comparison in term of area under the precision recall curve (AUPR).

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Mourad, R. Semi-supervised learning improves regulatory sequence prediction with unlabeled sequences. BMC Bioinformatics 24, 186 (2023). https://doi.org/10.1186/s12859-023-05303-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12859-023-05303-2

Keywords