Abstract

Motivation

Nucleotide-binding leucine-rich repeat (NLR) family is a class of immune receptors capable of detecting and defending against pathogen invasion. They have been widely used in crop breeding. Notably, the correspondence between NLRs and effectors (CNE) determines the applicability and effectiveness of NLRs. Unfortunately, CNE data is very scarce. In fact, we’ve found a substantial 91 291 NLRs confirmed via wet experiments and bioinformatics methods but only 387 CNEs are recognized, which greatly restricts the potential application of NLRs.

Results

We propose a deep learning algorithm called ProNEP to identify NLR-effector pairs in a high-throughput manner. Specifically, we conceptualized the CNE prediction task as a protein–protein interaction (PPI) prediction task. Then, ProNEP predicts the interaction between NLRs and effectors by combining the transfer learning with a bilinear attention network. ProNEP achieves superior performance against state-of-the-art models designed for PPI predictions. Based on ProNEP, we conduct extensive identification of potential CNEs for 91 291 NLRs. With the rapid accumulation of genomic data, we expect that this tool will be widely used to predict CNEs in new species, advancing biology, immunology, and breeding.

Availability and implementation

The ProNEP is available at http://nerrd.cn/#/prediction. The project code is available at https://github.com/QiaoYJYJ/ProNEP.

1 Introduction

Pathogens are the main biological stresses faced by plants (Singh et al. 2023). They cause great economic losses (Trumbore et al. 2015) and pose a serious threat to global food security. Some plants have evolved sophisticated immune systems to sense pathogens and activate the immune responses to fight them. One of the key components of the immune system is the nucleotide-binding leucine-rich repeat (NLR) receptors family (Liu et al. 2023). It is activated by virulence proteins (i.e. effectors; Contreras et al. 2023) secreted by pathogens, triggering hypersensitivity and cell death (Wan et al. 2019) to fight pathogen invasion. NLR has attracted increasing attention recently. A large number of NLRs have been found (Jones et al. 2016, Liu et al. 2021, Calle García et al. 2022), and some NLRs have been used for breeding disease-resistant crops. For example, Rpi-blb2 was originally identified in Solanum bulbocastanum as an NLR to resist the late blight disease caused by Phytophthora infestans (Orbegozo et al. 2016). The gene encoding Rpi-blb2 is transferred to Solanum demissum to confer a broad-spectrum resistance to isolates of P.infestans (Haverkort et al. 2009, Wu et al. 2017), which can greatly reduce the costs of disease control and increase the yield of potatoes (Haverkort et al. 2008). To successfully apply NLRs in resistance breeding, identifying the targets (i.e. pathogen effectors) of NLRs is crucial (Białas et al. 2018). Wet experiment is the main and reliable method to identify the presence of NLR and effector pairs (i.e. the correspondence between NLRs and effectors; CNE), but it requires high experimental cost and a long experimental period (Dao et al. 2019). Therefore, CNE data is very scarce. In fact, some databases have collected NLR data but there is no CNE information (Liu et al. 2021, Li et al. 2023). Through a systematic review of the literature, we identified a total of 91 291 NLRs (Liu et al. 2021) that have been verified through wet experiments and bioinformatics methods. In contrast, only 387 CNEs of 259 NLRs have been experimentally confirmed. The limited number of CNEs restricts the application of NLRs in breeding practice. Therefore, accurately identifying CNEs on large-scale data becomes urgent.

Deep learning is one of the fastest-growing topics in biology (Tang et al. 2021, Ao et al. 2022, Choudhary et al. 2022, Li et al. 2023, Wang et al. 2023, Yan et al. 2023) and achieved great success in the last few years (Zeng et al. 2022). Especially in protein fields, deep learning has been used extensively (Chen et al. 2023), such as protein structure prediction models (e.g. AlphaFold2; Jumper et al. 2019), drug-target interaction (DTI) prediction models (e.g. GraphDTA; Nguyen et al. 2021), and protein–protein interaction (PPI) prediction models (e.g. Topsy-Turvy; Singh et al. 2022). One big advantage of deep learning in solving protein problems is that it can learn from large volumes of protein data to embed them into a complex latent feature space (Lin et al. 2023). In addition, large protein language models together with attention mechanisms (Bahdanau et al. 2014) can learn interactions among sequence elements (Kaplan et al. 2020, Li et al. 2021, Hoffmann et al. 2022, Nijkamp et al. 2022, Song et al. 2022, Zhang et al. 2022). Deep learning is also capable of processing and integrating protein information from various data sources (Kulmanov et al. 2018), such as sequences (Chowdhury et al. 2022, Jin et al. 2022, Wang et al. 2023), structures (Berman et al. 2000), interaction networks (Szklarczyk et al. 2019), and even domain information (Mistry et al. 2021). Through this integration, deep learning models can capture the complex relationships and interdependencies between different modalities, thereby enhancing the accuracy and robustness of automatic predictions (Sapoval et al. 2022, Ao et al. 2023). However, to our best knowledge, there are currently no automatic tools to identify CNEs.

In this study, we conceptualized CNE identification as the problem of PPI prediction based on the fact that NLR and effector interact directly or indirectly through guard proteins or bait proteins in the process of immune response. Although we may benefit from the latest deep learning methods in PPI predictions, two challenges remain in our work. The first challenge is how to build a well-performing deep learning model based on a limited dataset (387 CNEs, verified by traditional wet experiments; see above). The sparsity of the training data remains a fundamental challenge for deep learning approaches (Lake et al. 2017). The second challenge is how to learn the pairing of NLR and effector. CNE is essentially determined by the interactions between residues in NLR and effector. The ability to simulate complex interactions between two inputs is critical to CNE identifications. In previous PPI studies, the focus has been on capturing position-wise interactions between residues by applying element-wise operations (e.g. sum and product) over protein features (Sledzieski et al. 2021). Learning all pairwise interactions between residues still requires further exploration.

To address these challenges, we presented a computational method for CNE identification called ProNEP based on transfer learning and the bilinear attention network (Kim et al. 2018, Yu et al. 2018). Transfer learning based on pre-trained models has proven successful in both natural language processing (Peters et al. 2018, Devlin et al. 2019) and pattern recognition domains (Shin et al. 2016). They can make use of the large data from related tasks to improve the performance of the target task with limited data. We first transferred protein knowledge from a pre-trained model from Bepler and Berger [a semi-supervised bidirectional long short-term memory (Bi-LSTMs) neural network] to embed NLRs and effectors to generate sequence features (XN and XE). XN and XE capture the local and global information of the structure, function, and evolution of NLRs and effectors sequences, respectively. Then, the dimensions of XN and XE are reduced using the Convolutional Neural Network (CNN) module to obtain SN and SE which capture a series of local residue patterns by expanding the receptive field through a chain of convolutional layers. Next, SN and SE are fed into a pairwise interaction module that consists of a bilinear attention network to output a joint representation that learns detailed interactions between all residues of NLRs and effectors in a multiplicative way. Finally, we used an Multilayer Perceptron (MLP) to decode the joint representation into the CNE prediction. The evaluation on the independent test dataset shows that ProNEP demonstrates strong performance, with an Area Under the Receiver Operating Characteristic Curve (AUROC) of 0.966 and an Area Under the Precision-Recall Curve (AUPRC) of 0.747. The evaluation on two unseen datasets where 10% NLRs and effectors were deliberately excluded from the training and validation sets, respectively, shows the robust generalization of ProNEP, with both AUROC exceeding 0.90 and both AUPRC over 0.69. Moreover, ProNEP accurately predicted all CNEs in nine newly published datasets. These results suggest that ProNEP can rapidly and accurately identify potential CNEs on a large scale, thereby advancing species conservation and agricultural production. In summary, the main contributions of ProNEP are: (i) it is the first, to our knowledge, CNE prediction model that can work in a high-throughput way based solely on sequence data; (ii) it introduces a bilinear attention mechanism to capture all the residue interactions between NLRs and effectors.

2 Methods

2.1 Pre-trained module

The ProNEP model takes as inputs an NLR sequence of length m and an effector sequence of length n. We generated feature embeddings XNRm× 6165 and XERn× 6165 by embedding the NLRs and the effectors, respectively, using the pre-trained model from Bepler and Berger (2021). This pre-trained model is a supervised Bi-LSTM neural network. The bidirectional architecture enables the pre-trained model to effectively consider contextual dependencies in both the forward and backward directions, extracting more comprehensive protein features. By leveraging transfer learning based on the pre-trained model, we transferred the generic features and knowledge learned from large-scale protein sequence data to the CNE prediction problem. Overall, XN and XE captured the local and global features of structure, function, and evolution of NLRs and effectors, respectively.

2.2 CNN module

The CNN module consists of four consecutive 1D-convolutional layers. The initial convolutional layer is employed to capture local residue patterns using a kernel size of 3. Subsequently, the next three layers continue to expand the receptive field, allowing the CNN module to extract longer or larger local residue patterns from the protein feature matrix XN and XE while achieving dimensionality reduction. In each convolutional layer, a convolution operation is performed on the input matrix, followed by a non-linear transformation through the Rectified Linear Unit (ReLU) activation function. The CNN module is written as:

(1)
(2)

where SNRm×d0 and SERn×d0 are the final feature representations for NLRs and the effectors, WN(l) and WE(l) denote the learnable weight matrices, bN(l) and bE(l) denote the bias vectors in the lth CNN layer. For any given layer l, SN(l) and SE(l) denote the lth hidden protein representation. Initially, SN(0) is XN and SE(0) is XE.

2.3 Interaction learning module

Here, we introduced the multi-head bilinear attention network (Kim et al. 2018) to obtain all pairwise interactions between NLR and effector proteins. This bilinear approach effectively reduces the input dimensionality, thereby decreasing computational demands, while simultaneously enhancing the granularity of residue interaction information.

Upon applying a CNN model for dimensionality reduction, we obtained the feature vectors SNRm×d0 and SERn×d0 for the respective input sequences. A single-head pairwise interaction MRd0 × d0 can be obtained by the Hadamard calculation of the bilinear interaction map:

(3)

where URm×k and VRm×k are learnable weight matrix of NLR and effector features, respectively. P Rk is a learnable weight vector, 1 Rd0 represents a fixed all-ones vector, and ○ denotes Hadamard calculation. Specifically, each element in Equation (3) can be represented by:

(4)

where SNi is the ith column of SN and SEj is the jth column of SE, respectively, denoting the ith and jth residue representations of the NLRs and effectors.

We obtained the final joint representation fh through bilinear pooling layers over the interaction maps M:

(5)
(6)

where h is for heads and fhk denotes the kth element of the intermediate representation in the hth head.

Last, we added a sum pooling to the joint representation vectors to obtain a compact feature map:

(7)

where the SumPool(·) function is a one-dimensional and non-overlapped sum pooling operation with stride s. It reduces the dimensionality of fh Rk to f Rk/s. To get the prediction probability, the joint representation f is fed into the decoder, which is an MLP followed by a sigmoid function:

(8)

Finally, we jointly optimized all learnable parameters by backpropagation. The training objective is to minimize the cross-entropy loss as follows:

(9)

where yi is the ground-truth label of the ith NLR-effector pair and pi is the output probability of the model.

2.4 Experimental settings

We implemented ProNEP in PyTorch 2.1.0. LN and LE were set to the maximum sequence length of NLRs and effectors, respectively, with padding. The pre-trained module generated a 6165-dimensional feature vector representation. CNN reduced the feature dimension with the number of filters [128, 128, 128, 128] and corresponding kernel sizes [3, 5, 7, 4]. In the bilinear attention module, we employed two attention heads. The MLP was configured with an input layer of size 256, a single hidden layer of size 512, and an output vector of size 2. During training, a batch size of 16 is utilized, and the Adam optimizer with a learning rate of 5 × 10−5 is applied. The models are trained for 100 epochs to ensure convergence. We selected the model with the best AUROC performance on the validation dataset and then evaluated the final performance on the test dataset.

We conducted a comparative analysis involving six models: Support Vector Machine (SVM), Random Forest (RF), D-SCRIPT, Topsy-Turvy, PIPR, and Alphafold3. To construct SVM and RF models, we utilized pretrained models to obtain feature representations of the sequence data. However, the model performance was poor, with AUROC and AUPRC scores reaching only 0.6771 and 0.382 at best, respectively. Alternatively, we employed Profeat to extract 2001-dimensional routine sequence features, capturing protein structural and physicochemical properties (Li et al. 2006), to build up SVM and RF models. D-SCRIPT leveraged sequence dissimilarity calculations to generate contact maps and employed a CNN-based interaction module to learn the interactions among sequence residues. Topsy-Turvy represents an enhancement to the D-SCRIPT model integrating both bottom-up (inferring properties from the characteristics of the individual protein sequences) and top-down (inferring properties from the pattern of already known PPIs in the species of interest) approaches. PIPR incorporates a deep residual recurrent convolutional neural network in the Siamese architecture. AlphaFold3 is the latest tool capable of efficiently and accurately predicting protein complex structures. When training these models, we used a threshold that maximized the F1 score. SVM with Radial Basis Function (RBF) kernel and RF are trained with the default parameters in scikit-learn. For the other four models, we followed the hyperparameter values used in the original literature. To address the issue of imbalanced training data, we incorporated a weighted-random-sampler in all the mentioned models.

3 Results

3.1 Datasets

We have collected data on NLR-effector pairs that have been validated through wet experiments from the literature and created the database NERRD. The database contains information including plant species, NLR name, NLR type, NLR protein sequences, effector name, pathogen, and effector protein sequences. To develop ProNEP, we used this database to create a dataset. In the positive samples, there are 387 CNEs validated by wet experiments, including 259 NLRs and 111 effectors. We randomly paired the NLR and effector sequences from the positive samples to generate negative samples (Hashemifar et al. 2018) based on a positive-negative ratio of 1:10. The dataset was divided into three parts: 80% for the training dataset, 10% for the validation dataset, and 10% for the independent test dataset, with no overlap among them (Supplementary Table S1). We conducted 10 repeated experiments, each was trained with different negative samples and tested with the same independent test set (Supplementary Table S2). To assess ProNEP’s generalization capabilities on new data, we created two new datasets: unseen-NLR and unseen-effector. For the unseen-NLR dataset, we selected 10% of NLRs from the full set and chose interaction data including these proteins for the test set, ensuring that no NLRs from this set were present in the training and validation sets (Supplementary Table S3). This dataset mimics real-world applications where users predict potential CNEs by inputting NLR data not included in our dataset. The same approach was employed to generate the unseen-effector dataset (Supplementary Table S4), which simulates a scenario where researchers predict potential CNEs by inputting their own effector data. Additionally, we tested two other pre-trained models: ProtTrans (Elnaggar et al. 2022) and ESM-1b (Rao et al. 2020). We found that ProtTrans and the Bepler and Berger model performed comparably, while ESM-1b showed weaker performance (Supplementary Table S5). We speculate that this may be due to the model’s limitations regarding sequence length. We chose the Bepler and Berger model to construct ProNEP due to its small size (Supplementary Table S5) and the requirement of online server implementation.

3.2 ProNEP framework

In the CNE prediction, we aim to develop a function, denoted as F, which can predict the likelihood of the interaction between a given NLR-effector pair. ProNEP takes a pair of protein sequences, denoted as LN and LE, with respective lengths of m and n as input and produces an interaction probability, p ∈ [0, 1], as its output. It consists of four main components: a pre-trained module for extracting protein sequence features, a CNN dimension reduction module, a bilinear attention module capturing all protein residue interactions, and an MLP to output prediction probabilities, as shown in Fig. 1. We employed the pre-trained module to generate feature embeddings XNRm×6165 and XERn×6165 for LN and LE, capturing both local and global aspects of the protein structure, function, and evolution across various dimensions (Bepler and Berger 2021). Subsequently, the CNN module is utilized to reduce the dimensionality of the features produced by the pre-trained module, resulting in the final feature representations SNRm×d0 and SERn×d0 for NLRs and the effectors, respectively. Thereafter, we deployed a bilinear attention network module to generate the joint feature representations of SN and SE to learn the multiplicative interactions between all residues of both proteins. Finally, the interaction probability p ∈ [0, 1] is predicted through an MLP. It represents the likelihood of an interaction between a given pair of NLR and effector.

Overview of the fundamental framework of ProNEP. We first used the model proposed by Bepler and Berger as a pre-trained module to generate XN ∈ Rm×6165 and XE ∈ Rn×6165, which contain local and global information on the structure, function, and evolution of proteins. Then, XN and XE were transformed to SN ∈ Rm×d0 and SE ∈ Rn×d0 through the CNN module. The bilinear attention module was used to capture local interactions between SN and SE. In this module, two attention maps were first computed with SN and SE. Then, through the bilinear attention network, the joint representations f1 and f2 were calculated and summed to obtain a compact feature map f. Finally, an MLP was used as a classifier to output the prediction probability p ∈ [0,1] based on the final joint representation f
Figure 1.

Overview of the fundamental framework of ProNEP. We first used the model proposed by Bepler and Berger as a pre-trained module to generate XNRm×6165 and XERn×6165, which contain local and global information on the structure, function, and evolution of proteins. Then, XN and XE were transformed to SNRm×d0 and SERn×d0 through the CNN module. The bilinear attention module was used to capture local interactions between SN and SE. In this module, two attention maps were first computed with SN and SE. Then, through the bilinear attention network, the joint representations f1 and f2 were calculated and summed to obtain a compact feature map f. Finally, an MLP was used as a classifier to output the prediction probability p ∈ [0,1] based on the final joint representation f

3.3 Evaluation of ProNEP and other algorithms

In this study, we compared ProNEP with other models, including SVM (Manganaro et al. 2023, Zhu et al. 2023, Zou et al. 2023), RF, D-SCRIPT (Sledzieski et al. 2021), Topsy-Turvy (Singh et al. 2022), PIPR (Chen et al. 2019), and Alphafold3 (Abramson et al. 2024). Since our training data is significantly imbalanced, with negative samples outnumbering positive samples, we used metrics, such as AUROC and AUPRC to evaluate the model performance. AUROC and AUPRC do not depend on a specific cutoff for their calculation. We chose the model with the highest AUROC performance on the validation dataset as the best-performing model, and then tested it on the independent test dataset. ProNEP achieved the highest AUROC and AUPRC, reaching 0.9685 and 0.947, respectively (Fig. 2). The second best is PIPR, with AUROC and AUPRC reaching 0.879 and 0.678 (Fig. 2). On the contrast, the performance of SVM is poor (Table 1 and Fig. 2). Although D-SCRIPT and Topsy-Turvy adopt the same feature engineering strategy as ProNEP, they do not perform as well as ProNEP (Table 1 and Fig. 2). Additionally, we tested the performance of SVM, RF, D-SCRIPT, Topsy-Turvy, Alphafold3, PIPR, and ProNEP on the unseen datasets (unseen-NLR and unseen-effector). ProNEP performs best on these two datasets, with AUROC and AUPRC scores reaching 0.9292 and 0.7134 on the unseen-NLR dataset and 0.9076 and 0.6973 on the unseen-effector dataset (Fig. 3).

AUPRC and AUROC of SVM, RF, D-SCRIPT, Topsy-Turvy, Alphafold3, PIPR, and ProNEP on the independent test dataset
Figure 2.

AUPRC and AUROC of SVM, RF, D-SCRIPT, Topsy-Turvy, Alphafold3, PIPR, and ProNEP on the independent test dataset

AUPRC and AUROC of SVM, RF, D-SCRIPT, Topsy-Turvy, Alphafold3, PIPR, and ProNEP on two unseen datasets
Figure 3.

AUPRC and AUROC of SVM, RF, D-SCRIPT, Topsy-Turvy, Alphafold3, PIPR, and ProNEP on two unseen datasets

Table 1.

Performance comparison of SVM, RF, D-SCRIPT, Topsy-Turvy, Alphafold3, PIPR, and ProNEP on the independent test dataset.a

ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
SVM0.781 ± 0.0240.157 ± 0.0280.323 ± 0.0370.277 ± 0.1680.617 ± 0.0650.176 ± 0.057
RF0.806 ± 0.0750.749 ± 0.0860.610 ± 0.1220.716 ± 0.1210.783 ± 0.1270.483 ± 0.117
D-SCRIPT0.609 ± 0.0730.324 ± 0.2330.719 ± 0.2520.335 ± 0.0560.686 ± 0.0650.306 ± 0.093
Topsy-Turvy0.672 ± 0.1170.225 ± 0.1280.546 ± 0.3290.221 ± 0.0460.64 ± 0.0460.359 ± 0.117
Alphafold30.1850.09520.91890.17260.56880.1185
PIPR0.912 ± 0.0080.569 ± 0.0320.594 ± 0.0260.581 ± 0.0260.879 ± 0.0160.678 ± 0.004
ProNEP0.914 ± 0.0280.945 ± 0.0350.915 ± 0.0360.928 ± 0.0220.966 ± 0.0120.747 ± 0.054
ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
SVM0.781 ± 0.0240.157 ± 0.0280.323 ± 0.0370.277 ± 0.1680.617 ± 0.0650.176 ± 0.057
RF0.806 ± 0.0750.749 ± 0.0860.610 ± 0.1220.716 ± 0.1210.783 ± 0.1270.483 ± 0.117
D-SCRIPT0.609 ± 0.0730.324 ± 0.2330.719 ± 0.2520.335 ± 0.0560.686 ± 0.0650.306 ± 0.093
Topsy-Turvy0.672 ± 0.1170.225 ± 0.1280.546 ± 0.3290.221 ± 0.0460.64 ± 0.0460.359 ± 0.117
Alphafold30.1850.09520.91890.17260.56880.1185
PIPR0.912 ± 0.0080.569 ± 0.0320.594 ± 0.0260.581 ± 0.0260.879 ± 0.0160.678 ± 0.004
ProNEP0.914 ± 0.0280.945 ± 0.0350.915 ± 0.0360.928 ± 0.0220.966 ± 0.0120.747 ± 0.054
a

ProNEP was trained 10 times with 10 different negative samples, while the other models, excluding Alphafold3, were trained 6 times. All models, except Alphafold3, used a threshold that maximized the F1 score.

Table 1.

Performance comparison of SVM, RF, D-SCRIPT, Topsy-Turvy, Alphafold3, PIPR, and ProNEP on the independent test dataset.a

ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
SVM0.781 ± 0.0240.157 ± 0.0280.323 ± 0.0370.277 ± 0.1680.617 ± 0.0650.176 ± 0.057
RF0.806 ± 0.0750.749 ± 0.0860.610 ± 0.1220.716 ± 0.1210.783 ± 0.1270.483 ± 0.117
D-SCRIPT0.609 ± 0.0730.324 ± 0.2330.719 ± 0.2520.335 ± 0.0560.686 ± 0.0650.306 ± 0.093
Topsy-Turvy0.672 ± 0.1170.225 ± 0.1280.546 ± 0.3290.221 ± 0.0460.64 ± 0.0460.359 ± 0.117
Alphafold30.1850.09520.91890.17260.56880.1185
PIPR0.912 ± 0.0080.569 ± 0.0320.594 ± 0.0260.581 ± 0.0260.879 ± 0.0160.678 ± 0.004
ProNEP0.914 ± 0.0280.945 ± 0.0350.915 ± 0.0360.928 ± 0.0220.966 ± 0.0120.747 ± 0.054
ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
SVM0.781 ± 0.0240.157 ± 0.0280.323 ± 0.0370.277 ± 0.1680.617 ± 0.0650.176 ± 0.057
RF0.806 ± 0.0750.749 ± 0.0860.610 ± 0.1220.716 ± 0.1210.783 ± 0.1270.483 ± 0.117
D-SCRIPT0.609 ± 0.0730.324 ± 0.2330.719 ± 0.2520.335 ± 0.0560.686 ± 0.0650.306 ± 0.093
Topsy-Turvy0.672 ± 0.1170.225 ± 0.1280.546 ± 0.3290.221 ± 0.0460.64 ± 0.0460.359 ± 0.117
Alphafold30.1850.09520.91890.17260.56880.1185
PIPR0.912 ± 0.0080.569 ± 0.0320.594 ± 0.0260.581 ± 0.0260.879 ± 0.0160.678 ± 0.004
ProNEP0.914 ± 0.0280.945 ± 0.0350.915 ± 0.0360.928 ± 0.0220.966 ± 0.0120.747 ± 0.054
a

ProNEP was trained 10 times with 10 different negative samples, while the other models, excluding Alphafold3, were trained 6 times. All models, except Alphafold3, used a threshold that maximized the F1 score.

3.4 Ablation study

To validate the effectiveness of the pre-trained module and the bilinear attention module in predicting CNE interactions, we conducted ablation experiments. We compared the complete model with three different variants of ProNEP: No pre-trained module, unilateral attention, and linear concatenation. These three variants differ in the computation of the joint representation of NLR and effector. In the first variant, we removed the pre-trained model and used word embedding to learn protein representations. The results show that, compared with the model without the pre-training module, ProNEP’s AUROC and AUPRC increased by 39.22% and 41.33% (Table 2), respectively. This indicated that the pre-trained model can capture richer feature representations and significantly improve the model performance. In the second variant, we replaced the bilinear attention with a unilateral attention, forming two models. The unilateral attention employs the neural attention mechanism (Tsubaki et al. 2019), which captured the interaction strengths of an NLR as a whole with subsequences of an effector (unilateral EFF attention; Table 2) and the interaction strengths of an effector as a whole with subsequences of an NLR (unilateral NLR attention; Table 2). The AUROC of the two models were 81.95% and 82.89%, and the AUPRC were 56.21% and 50.39%, respectively. Finally, we replaced the bilinear attention with a linear concatenation. That is, after the CNN module, we simply concatenated the vector representations of NLR and effector. This variant achieves an AUROC of 55.1% and an AUPRC of 65.2% (Table 2). The results of unilateral attention and linear connection variants show that the bilinear attention mechanism captures the CNE interaction information more effectively than the two variants. Overall, the ablation study highlights the contributions of the transfer learning and the bilinear attention network to identifying CNEs.

Table 2.

The ablation study of ProNEP on the independent test dataset.

ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
No-pre-trained0.8350.80670.83750.82180.56830.3787
Unilateral NLR attention0.83810.90270.83200.86590.82890.5621
Unilateral EFF attention0.84120.74450.85120.79430.81950.5039
Linear concatenation0.83750.890.83190.860.89880.6621
ProNEP0.91440.94540.91450.92830.96580.747
ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
No-pre-trained0.8350.80670.83750.82180.56830.3787
Unilateral NLR attention0.83810.90270.83200.86590.82890.5621
Unilateral EFF attention0.84120.74450.85120.79430.81950.5039
Linear concatenation0.83750.890.83190.860.89880.6621
ProNEP0.91440.94540.91450.92830.96580.747
Table 2.

The ablation study of ProNEP on the independent test dataset.

ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
No-pre-trained0.8350.80670.83750.82180.56830.3787
Unilateral NLR attention0.83810.90270.83200.86590.82890.5621
Unilateral EFF attention0.84120.74450.85120.79430.81950.5039
Linear concatenation0.83750.890.83190.860.89880.6621
ProNEP0.91440.94540.91450.92830.96580.747
ModelAccuracyPrecisionSensitivityF1-scoreAUROCAUPRC
No-pre-trained0.8350.80670.83750.82180.56830.3787
Unilateral NLR attention0.83810.90270.83200.86590.82890.5621
Unilateral EFF attention0.84120.74450.85120.79430.81950.5039
Linear concatenation0.83750.890.83190.860.89880.6621
ProNEP0.91440.94540.91450.92830.96580.747

3.5 Robustness assessment of ProNEP

In order to further evaluate the robustness of ProNEP, we selected the Arabidopsis thaliana and Hyaloperonospora arabidopsidis system with five known CNEs and Triticum aestivum and Blumeria graminis system with 20 known CNEs for testing. To construct the A.thaliana and H.arabidopsidis system dataset, we first excluded data of this system from our training and validation sets, then retrained the model. We then randomly selected 100 proteins from A.thaliana, mimicking the length distribution of NLR sequences, and another 100 proteins from H.arabidopsidis, mimicking the distribution of effectors, for testing. This process was repeated 10 times. Among these 10 datasets, 5 did not contain any positive samples (Table 3). ProNEP correctly predicted all five known CNEs and achieved a high prediction accuracy, ranging from 0.97 to 1 (Table 3). Additionally, we collected nine newly published CNE data for testing (Table 4), and ProNEP accurately predicted all of them. Overall, the results indicate that ProNEP has good generalization ability, which can be a valuable tool for researchers in identifying potential CNEs.

Table 3.

Results of two plant-pathogen systems.

ExperimentTrue positivesFalse positivesFalse negativesTrue negativesAccuracy
Arabidopsis thaliana—Hyaloperonospora arabidopsidis system
10001001
20001001
30001001
40001001
50001001
6200981
7320950.98
8230950.97
9110980.99
10310960.99
Triticum aestivum—Blumeria graminis system
1041950.95
2132940.95
3031960.96
4022960.96
5531910.96
6522910.96
7412930.97
8721900.97
9002980.98
10320950.98
ExperimentTrue positivesFalse positivesFalse negativesTrue negativesAccuracy
Arabidopsis thaliana—Hyaloperonospora arabidopsidis system
10001001
20001001
30001001
40001001
50001001
6200981
7320950.98
8230950.97
9110980.99
10310960.99
Triticum aestivum—Blumeria graminis system
1041950.95
2132940.95
3031960.96
4022960.96
5531910.96
6522910.96
7412930.97
8721900.97
9002980.98
10320950.98
Table 3.

Results of two plant-pathogen systems.

ExperimentTrue positivesFalse positivesFalse negativesTrue negativesAccuracy
Arabidopsis thaliana—Hyaloperonospora arabidopsidis system
10001001
20001001
30001001
40001001
50001001
6200981
7320950.98
8230950.97
9110980.99
10310960.99
Triticum aestivum—Blumeria graminis system
1041950.95
2132940.95
3031960.96
4022960.96
5531910.96
6522910.96
7412930.97
8721900.97
9002980.98
10320950.98
ExperimentTrue positivesFalse positivesFalse negativesTrue negativesAccuracy
Arabidopsis thaliana—Hyaloperonospora arabidopsidis system
10001001
20001001
30001001
40001001
50001001
6200981
7320950.98
8230950.97
9110980.99
10310960.99
Triticum aestivum—Blumeria graminis system
1041950.95
2132940.95
3031960.96
4022960.96
5531910.96
6522910.96
7412930.97
8721900.97
9002980.98
10320950.98
Table 4.

Detailed information on nine newly collected CNEs.

NLRPlantEffectorPathogenReferences
CSA1Arabidopsis thalianaHopBPseudomonas syringaeSchulze et al. (2022)
Mla1 (Crean et al., 2023)Hordeum vulgareAvra1Blumeria graminisCrean et al. (2023)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2Magnaporthe oryzaeBrabham et al. (2024)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2-2Magnaporthe oryzaeBrabham et al. (2024)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-1Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-2Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-3Blumeria graminisCrean et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaripbRalstonia solanacearumPrautsch et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaXopQRalstonia solanacearumPrautsch et al. (2023)
NLRPlantEffectorPathogenReferences
CSA1Arabidopsis thalianaHopBPseudomonas syringaeSchulze et al. (2022)
Mla1 (Crean et al., 2023)Hordeum vulgareAvra1Blumeria graminisCrean et al. (2023)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2Magnaporthe oryzaeBrabham et al. (2024)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2-2Magnaporthe oryzaeBrabham et al. (2024)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-1Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-2Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-3Blumeria graminisCrean et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaripbRalstonia solanacearumPrautsch et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaXopQRalstonia solanacearumPrautsch et al. (2023)
Table 4.

Detailed information on nine newly collected CNEs.

NLRPlantEffectorPathogenReferences
CSA1Arabidopsis thalianaHopBPseudomonas syringaeSchulze et al. (2022)
Mla1 (Crean et al., 2023)Hordeum vulgareAvra1Blumeria graminisCrean et al. (2023)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2Magnaporthe oryzaeBrabham et al. (2024)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2-2Magnaporthe oryzaeBrabham et al. (2024)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-1Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-2Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-3Blumeria graminisCrean et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaripbRalstonia solanacearumPrautsch et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaXopQRalstonia solanacearumPrautsch et al. (2023)
NLRPlantEffectorPathogenReferences
CSA1Arabidopsis thalianaHopBPseudomonas syringaeSchulze et al. (2022)
Mla1 (Crean et al., 2023)Hordeum vulgareAvra1Blumeria graminisCrean et al. (2023)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2Magnaporthe oryzaeBrabham et al. (2024)
Mla3 (Brabham et al., 2024)Hordeum vulgarePwl2-2Magnaporthe oryzaeBrabham et al. (2024)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-1Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-2Blumeria graminisCrean et al. (2023)
Mla13 (Crean et al., 2023)Hordeum vulgareAvra13-3Blumeria graminisCrean et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaripbRalstonia solanacearumPrautsch et al. (2023)
Roq1 (Prautsch et al., 2023)Nicotiana benthamianaXopQRalstonia solanacearumPrautsch et al. (2023)

4 Discussion

NLRs play a crucial role in plant resistance against a wide range of pathogens, including fungi, bacteria, oomycetes, and viruses (Belkhadir et al. 2004, Collier and Moffett 2009). They can recognize pathogen effectors and activate the immune response to defend against diseases. Identifying NLRs and their corresponding effectors is crucial, as it facilitates the intraspecific (Vuong et al. 2023) and interspecific (Orbegozo et al. 2016) transfer of NLRs through genetic engineering, contributing significantly to the enhanced disease defense. In this study, we introduce ProNEP, the first deep learning tool for the precise identification of NLR-effector pairs. ProNEP offers a powerful and efficient means to identify the interactions between NLRs and effectors in a high-throughput way, thereby advancing our understanding of plant immunity and enabling targeted strategies for enhancing crop resistance.

ProNEP demonstrates superior performance by leveraging both the transfer learning and the bilinear attention mechanisms. Transfer learning based on the pre-trained models has demonstrated significant potential in addressing protein-related challenges. For example, the CLEAN (Yu et al. 2023), built upon a 650-million-parameter protein language model (ESM1b; Rives et al. 2021), stands out for its superior performance in predicting enzyme functions. In the prediction of PPI, the D-SCRIPT (Sledzieski et al. 2021), utilizing a structure-based pre-trained protein language, have also outperformed other methods. Simultaneously, transfer learning exhibits robust adaptability and generalization capabilities when confronted with small-sample data. In our study, transfer learning augments protein information for CNE identification, which greatly enhances model performance. On the other hand, the acquisition of interaction representation is pivotal for the identification of CNEs, which can be effectively achieved through the bilinear attention network. The bilinear attention network was originally designed to address visual question answering (VQA) problems, focusing on pairs of image regions and question words to learn their interaction representation (Kim et al. 2018). The bilinear attention network has achieved notable success in several protein-related issues, particularly excelling in the domain of drug target prediction (Drug-BAN; Bai et al. 2023). However, its application in PPI research remains limited. D-SCRIPT, for instance, only uses a residue contact module to predict interactions between proteins. Topsy-Turvy builds on D-SCRIPT, combining it with global and local integrated diffusion embedding (GLIDE). GLIDE predicts PPI by quantifying the likelihood of interactions between each pair of proteins in a network through a combination of local (proximity-based) and global (diffusion state embedding-based) graph theory techniques. Despite using the same feature engineering strategies, D-SCRIPT and Topsy-Turvy do not perform as well as ProNEP, which employs a bilinear attention mechanism that precisely focuses on critical residue pair information in protein sequences. We expect the introduction of the bilinear attention network to improve the performance of PPI models by leveraging its powerful interactive representation learning capabilities.

ProNEP has three functions: Identify effectors for given NLRs; identify NLRs for given effectors; and calculate the interaction probability for given pairs of NLRs and effectors. We used ProNEP to identify 23 310 potential CNE interactions from 91 291 NLRs in public databases (Liu et al. 2021), involving 111 effectors. We have uploaded the results to the NERRD database. However, we also recognize that the diversity of effectors in real-world scenarios far exceeds the range represented in our current dataset. To mitigate the potential limitation brought by the diversity of effectors in our database, we recommend using two functions of ProNEP, “identify NLRs for given effectors” and “calculate the interaction probability for given pairs of NLRs and effectors.” These functions allow researchers to use their own effectors and NLRs that are not included in the database to predict potential interactions. In further study, we plan to enrich the diversity within our database by incorporating a larger and more diverse set of effector data to enhance the study of plant immune mechanisms. In summary, we anticipate widespread applications of ProNEP in predicting CNEs for more species, thereby advancing research in plant biology (Tamborski and Krasileva 2020), plant immunology (Yuan et al. 2021), and crop breeding (Zhang et al. 2023).

Acknowledgements

We thank the Ecosystem Big Data Research and Development Center of Northeast Forestry University (NEFU) providing the additional computational resources for this study. We sincerely thank two anonymous reviewers for their insightful comments and constructive suggestions, which have greatly enhanced the quality of this paper. The data underlying this article are available in the GitHub repository at https://github.com/QiaoYJYJ/ProNEP/tree/main/datasets/nlr.

Author contributions

Baixue Qiao (Data curation, Writing—Original draft, Visualization, Platform development). Shuda Wang (Platform development, Website design). Mingjun Hou (Data curation). Haodi Chen (Data curation). Zhengwenyang Zhou (Data curation). Xueying Xie (Data curation). Shaozi Pang (Data curation). Quan Zou (Writing—review and editing, Funding acquisition). Shanwen Sun (Conception and design, Project administration, Writing—review and editing, Funding acquisition).

Supplementary data

Supplementary data are available at Bioinformatics online.

Conflict of interest

None declared.

Funding

This work has been supported by the National Natural Science Foundation of China (62273086, 62471123, 62131004); and Fundamental Research Funds for the Central Universities (2572022BD04).

References

Abramson
J
,
Adler
J
,
Dunger
J
 et al.  
Accurate structure prediction of biomolecular interactions with AlphaFold 3
.
Nature
 
2024
;
630
:
493
500
.

Ao
C
,
Jiao
S
,
Wang
Y
 et al.  
Biological sequence classification: a review on data and general methods
.
Research (Wash D C)
 
2022
;
2022
:
0011
.

Ao
C
,
Ye
X
,
Sakurai
T
 et al.  
m5U-SVM: identification of RNA 5-methyluridine modification sites based on multi-view features of physicochemical features and distributed representation
.
BMC Biol
 
2023
;
21
:
93
.

Bahdanau
D
,
Cho
K
,
Bengio
YJC.
Neural machine translation by jointly learning to align and translate. arXiv, arXiv:1409.0473, 2014, preprint: not peer reviewed.

Bai
P
,
Miljković
F
,
John
B
 et al.  
Interpretable bilinear attention network with domain adaptation improves drug-target prediction
.
Nat Mach Intell
 
2023
;
5
:
126
36
.

Belkhadir
Y
,
Subramaniam
R
,
Dangl
JL.
 
Plant disease resistance protein signaling: NBS-LRR proteins and their partners
.
Curr Opin Plant Biol
 
2004
;
7
:
391
9
.

Bepler
T
,
Berger
B.
 
Learning the protein language: evolution, structure, and function
.
Cell Syst
 
2021
;
12
:
654
69.e3
.

Berman
HM
,
Westbrook
J
,
Feng
Z
 et al.  
The protein data bank
.
Nucleic Acids Res
 
2000
;
28
:
235
42
.

Białas
A
,
Zess
EK
,
De la Concepcion
JC
 et al.  
Lessons in effector and NLR biology of plant-microbe systems
.
Mol Plant Microbe Interact
 
2018
;
31
:
34
45
.

Brabham
HJ
, ,
De La Cruz
DG
,
,
Were
V
 et al.  
Barley MLA3 recognizes the host-specificity effector Pwl2 from
magnaporthe oryzae.
Plant Cell
 
2024
;
36
:
447
70
.

Calle García J, Guadagno A, Paytuvi-Gallart
A
 et al.  
PRGdb 4.0: an updated database dedicated to genes involved in plant disease resistance process
.
Nucleic Acids Res
 
2022
;
50
:
D1483
90
.

Chen
L
,
Yu
L
,
Gao
L.
 
Potent antibiotic design via guided search from antibacterial activity evaluations
.
Bioinformatics
 
2023
;
39
:
btad059
.

Chen
M
,
Ju
CJ-T
,
Zhou
G
 et al.  
Multifaceted protein-protein interaction prediction based on Siamese residual RCNN
.
Bioinformatics
 
2019
;
35
:
i305
i314
.

Choudhary
K
,
DeCost
B
,
Chen
C
 et al.  
Recent advances and applications of deep learning methods in materials science
.
NPJ Comput Mater
 
2022
;
8
:
59
.

Chowdhury
R
,
Bouatta
N
,
Biswas
S
 et al.  
Publisher correction: single-sequence protein structure prediction using a language model and deep learning
.
Nat Biotechnol
 
2022
;
40
:
1692
.

Collier
SM
,
Moffett
P.
 
NB-LRRs work a “bait and switch” on pathogens
.
Trends Plant Sci
 
2009
;
14
:
521
9
.

Contreras
MP
,
Pai
H
,
Selvaraj
M
 et al.  
Resurrection of plant disease resistance proteins via helper NLR bioengineering
.
Sci Adv
 
2023
;
9
:
eadg3861
.

Crean
EE
, ,
Bilstein-Schloemer
M
,
,
Maekawa
T.
 
A dominant-negative avirulence effector of the barley powdery mildew fungus provides mechanistic insight into barley MLA immune receptor activation
.
J Exp Bot
 
2023
;
74
:
erad285
.

Dao
F-Y
,
Lv
H
,
Wang
F
 et al.  
Identify origin of replication in Saccharomyces cerevisiae using two-step feature selection technique
.
Bioinformatics
 
2019
;
35
:
2075
83
.

Devlin
J
,
Chang
MW
,
Lee
K
 et al. BERT: pre-training of deep bidirectional transformers for language understanding. arXiv, arXiv: 1810.04805, 2019, preprint: not peer reviewed.

Elnaggar
A
,
Heinzinger
M
,
Dallago
C
 et al.  
ProtTrans: toward understanding the language of life through self-supervised learning
.
IEEE Trans Pattern Anal Mach Intell
 
2022
;
44
:
7112
27
.

Hashemifar
S
,
Neyshabur
B
,
Khan
AA
 et al.  
Predicting protein-protein interactions through sequence-based deep learning
.
Bioinformatics
 
2018
;
34
:
i802
i810
.

Haverkort
AJ
,
Boonekamp
PM
,
Hutten
R
 et al.  
Societal costs of late blight in potato and prospects of durable resistance through cisgenic modification
.
Potato Res
 
2008
;
51
:
47
57
.

Haverkort
AJ
,
Struik
PC
,
Visser
RGF
 et al.  
Applied biotechnology to combat late blight in potato caused by Phytophthora infestans
.
Potato Res
 
2009
;
52
:
249
64
.

Hoffmann
J, Borgeaud S,
Mensch A et al. Training compute-optimal large language models. arXiv, arXiv: 2203.15556, 2022, preprint: not peer reviewed.

Jin
J
,
Yu
Y
,
Wang
R
 et al.  
iDNA-ABF: multi-scale deep biological language learning model for the interpretable prediction of DNA methylations
.
Genome Biol
 
2022
;
23
:
219
.

Jones
JD
,
Vance
RE
,
Dangl
JL.
 
Intracellular innate immune surveillance devices in plants and animals
.
Science
 
2016
;
354
:
aaf6395
.

Jumper
J
,
Evans
R
,
Pritzel
A
 et al.  
Highly accurate protein structure prediction with AlphaFold
.
Nature
 
2019
;
596
:
583
9
.

Kaplan
J, McCandlish S, Henighan T
 et al. Scaling laws for neural language models. arXiv, arXiv: 2001.08361, 2020, preprint: not peer reviewed.

Kim
H
,
Jun
J
,
Zhang
BT.
Bilinear attention networks. arXiv, arXiv: 1805:07932, 2018, preprint: not peer reviewed.

Kulmanov
M
,
Khan
MA
,
Hoehndorf
R
 et al.  
DeepGO: predicting protein functions from sequence and interactions using a deep ontology-aware classifier
.
Bioinformatics
 
2018
;
34
:
660
8
.

Lake
BM
,
Ullman
TD
,
Tenenbaum
JB
 et al.  
Building machines that learn and think like people
.
Behav Brain Sci
 
2017
;
40
:
e253
.

Li
F
,
Guo
X
,
Bi
Y
 et al.  
Digerati—a multipath parallel hybrid deep learning framework for the identification of mycobacterial PE/PPE proteins
.
Comput Biol Med
 
2023
;
163
:
107155
.

Li
HL
,
Pang
YH
,
Liu
B.
 
BioSeq-BLM: a platform for analyzing DNA, RNA and protein sequences based on biological language models
.
Nucleic Acids Res
 
2021
;
49
:
e129
.

Li
X
,
Ma
L
,
Wang
Y
 et al.  
PlantNLRatlas: a comprehensive dataset of full- and partial-length NLR resistance genes across 100 chromosome-level plant genomes
.
Front Plant Sci
 
2023
;
14
:
1178069
.

Li
ZR
,
Lin
HH
,
Han
LY
 et al.  
PROFEAT: a web server for computing structural and physicochemical features of proteins and peptides from amino acid sequence
.
Nucleic Acids Res
 
2006
;
34
:
W32
W37
.

Lin
Z
,
Akin
H
,
Rao
R
 et al.  
Evolutionary-scale prediction of atomic-level protein structure with a language model
.
Science
 
2023
;
379
:
1123
30
.

Liu
Y
,
Zeng
Z
,
Zhang
Y-M
 et al.  
An angiosperm NLR atlas reveals that NLR gene reduction is associated with ecological specialization and signal transduction component deletion
.
Mol Plant
 
2021
;
14
:
2015
31
.

Liu
Y
,
Zhang
Y-M
,
Tang
Y
 et al.  
The evolution of plant NLR immune receptors and downstream signal components
.
Curr Opin Plant Biol
 
2023
;
73
:
102363
.

Manganaro
L
,
Sabbatini
G
,
Bianco
S
 et al.  
Non-small cell lung cancer survival estimation through multi-omic two-layer SVM: a multi-omics and multi-mources integrative model
.
CBIO
 
2023
;
18
:
658
69
.

Mistry
J
,
Chuguransky
S
,
Williams
L
 et al.  
Pfam: the protein families database in 2021
.
Nucleic Acids Res
 
2021
;
49
:
D412
9
.

Nguyen
T
,
Le
H
,
Quinn
TP
 et al.  
GraphDTA: predicting drug-target binding affinity with graph neural networks
.
Bioinformatics
 
2021
;
37
:
1140
7
.

Nijkamp
E
,
Ruffolo
JA
,
Weinstein
EN
 et al.  
ProGen2: exploring the boundaries of protein language models
.
Cell Syst
 
2022
;
14
:
968
78.e3
.

Orbegozo
J
,
Roman
ML
,
Rivera
C
 et al.  
Rpi-blb2 gene from Solanum bulbocastanum confers extreme resistance to late blight disease in potato
.
Plant Cell Tiss Organ Cult
2016
;
125
:
269
81
.

Peters
ME, Neumann M, Iyyer M
 et al. Deep contextualized word representations. arXiv, arXiv: 1802.05365, 2018, preprint: not peer reviewed.

Prautsch
J
, ,
Erickson
JL
,
,
Özyürek
S
 et al.  
Effector xopq-induced stromule formation in nicotiana benthamiana depends on ETI signaling components ADR1 AND NRG1
.
Plant Physiology
 
2023
;
191
:
161
76
.

Rao
R, Meier J, Sercu T
 et al. Transformer protein language models are unsupervised structure learners. bioRxiv, bioRxiv: 422761, 2020, preprint: not peer reviewed.

Rives
A
,
Meier
J
,
Sercu
T
 et al.  
Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences
.
Proc Natl Acad Sci USA
 
2021
;
118
:
e2016239118
.

Sapoval
N
,
Aghazadeh
A
,
Nute
MG
 et al.  
Current progress and open challenges for applying deep learning across the biosciences
.
Nat Commun
 
2022
;
13
:
1728
.

Schulze
S
,
Yu
L
,
Hua
C
 et al.  
The arabidopsis TIR-NBS-LRR protein CSA1 guards BAK1-BIR3 homeostasis and mediates convergence of pattern- and effector-induced immune responses
.
Cell Host Microbe
 
2022
;
30
:
1717
31.e6
.

Shin
H-C
,
Roth
HR
,
Gao
M
 et al.  
Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning
.
IEEE Trans Med Imaging
 
2016
;
35
:
1285
98
.

Singh
BK
,
Delgado-Baquerizo
M
,
Egidi
E
 et al.  
Climate change impacts on plant pathogens, food security and paths forward
.
Nat Rev Microbiol
 
2023
;
21
:
640
56
.

Singh
R
,
Devkota
K
,
Sledzieski
S
 et al.  
Topsy-Turvy: integrating a global view into sequence-based PPI prediction
.
Bioinformatics
 
2022
;
38
:
i264
i272
.

Sledzieski
S
,
Singh
R
,
Cowen
L
 et al.  
D-SCRIPT translates genome to phenome with sequence-based, structure-aware, genome-scale predictions of protein-protein interactions
.
Cell Syst
 
2021
;
12
:
969
82.e966
.

Song
B
,
Luo
X
,
Luo
X
 et al.  
Learning spatial structures of proteins improves protein-protein interaction prediction
.
Brief Bioinform
 
2022
;
23
:
bbab558
.

Szklarczyk
D
,
Gable
AL
,
Lyon
D
 et al.  
STRING v11: protein-protein association networks with increased coverage, supporting functional discovery in genome-wide experimental datasets
.
Nucleic Acids Res
 
2019
;
47
:
D607
D613
.

Tamborski
J
,
Krasileva
KV.
 
Evolution of plant NLRs: from natural history to precise modifications
.
Annu Rev Plant Biol
 
2020
;
71
:
355
78
.

Tang
YJ
,
Pang
YH
,
Liu
B.
 
IDP-Seq2Seq: identification of intrinsically disordered regions based on sequence to sequence learning
.
Bioinformatics
 
2021
;
36
:
5177
86
.

Trumbore
S
,
Brando
P
,
Hartmann
H.
 
Forest health and global change
.
Science
 
2015
;
349
:
814
8
.

Tsubaki
M
,
Tomii
K
,
Sese
J.
 
Compound-protein interaction prediction with end-to-end learning of neural networks for graphs and sequences
.
Bioinformatics
 
2019
;
35
:
309
18
.

Vuong
UT
,
Iswanto
ABB
,
Nguyen
Q-M
 et al.  
Engineering plant immune circuit: walking to the bright future with a novel toolbox
.
Plant Biotechnol J
 
2023
;
21
:
17
45
.

Wan
L
,
Essuman
K
,
Anderson
RG
 et al.  
TIR domains of plant immune receptors are NAD+-cleaving enzymes that promote cell death
.
Science
 
2019
;
365
:
799
803
.

Wang
R
,
Jiang
Y
,
Jin
J
 et al.  
DeepBIO: an automated and interpretable deep-learning platform for high-throughput biological sequence prediction, functional annotation and visualization analysis
.
Nucleic Acids Res
 
2023
;
51
:
3017
29
.

Wang
Y, Zhai Y, Ding Y
 et al. SBSM-Pro: support bio-sequence machine for proteins.arXiv, arXiv: 2308:10275, 2023, preprint: not peer reviewed.

Wu
C-H
,
Abd-El-Haliem
A
,
Bozkurt
TO
 et al.  
NLR network mediates immunity to diverse plant pathogens
.
Proc Natl Acad Sci USA
 
2017
;
114
:
8113
8
.

Yan
K
,
Lv
H
,
Guo
Y
 et al.  
sAMPpred-GAT: prediction of antimicrobial peptide by graph attention network and predicted peptide structure
.
Bioinformatics
 
2023
;
39
:
btac715
.

Yu
T
,
Cui
H
,
Li
JC
 et al.  
Enzyme function prediction using contrastive learning
.
Science
 
2023
;
379
:
1358
63
.

Yu
Z
,
Yu
J
,
Xiang
C
 et al.  
Beyond bilinear: generalized multimodal factorized high-order pooling for visual question answering
.
IEEE Trans Neural Netw Learn Syst
 
2018
;
29
:
5947
59
.

Yuan
M
,
Jiang
Z
,
Bi
G
 et al.  
Pattern-recognition receptors are required for NLR-mediated plant immunity
.
Nature
 
2021
;
592
:
105
9
.

Zeng
X
,
Wang
F
,
Luo
Y
 et al.  
Deep generative molecular design reshapes drug discovery
.
Cell Rep Med
 
2022
;
3
:
100794
.

Zhang
J
,
Nirmala
J
,
Chen
S
 et al.  
Single amino acid change alters specificity of the multi-allelic wheat stem rust resistance locus SR9
.
Nat Commun
 
2023
;
14
:
7354
.

Zhang
W
,
Meng
Q
,
Wang
J
 et al.  
HDIContact: a novel predictor of residue-residue contacts on hetero-dimer interfaces via sequential information and transfer learning strategy
.
Brief Bioinform
 
2022
;
23
:
bbac169
.

Zhu
W
,
Yuan
S-S
,
Li
J
 et al.  
A first computational frame for recognizing heparin-binding protein
.
Diagn
 
2023
;
13
:
2465
.

Zou
X
,
Ren
L
,
Cai
P
 et al.  
Accurately identifying hemagglutinin using sequence information and machine learning methods
.
Front Med
 
2023
;
10
:
128188
.

Author notes

= Baixue Qiao and Shuda Wang equal contribution.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted reuse, distribution, and reproduction in any medium, provided the original work is properly cited.
Associate Editor: Macha Nikolski
Macha Nikolski
Associate Editor
Search for other works by this author on:

Supplementary data