Next Article in Journal
Pharmacogenetic-Based Interactions between Nutraceuticals and Angiogenesis Inhibitors
Previous Article in Journal
In Silico Genome-Wide Analysis of Respiratory Burst Oxidase Homolog (RBOH) Family Genes in Five Fruit-Producing Trees, and Potential Functional Analysis on Lignification of Stone Cells in Chinese White Pear
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Prediction Method for Plant lncRNA-Protein Interaction

1
School of Computer Science and Technology, Dalian University of Technology, Dalian 116023, Liaoning, China
2
Department of Information Technology, Jomo Kenyatta University of Agriculture and Technology, Nairobi 62000-00200, Kenya
3
School of Bioengineering, Dalian University of Technology, Dalian 116023, Liaoning, China
4
College of Life Sciences, Zhejiang University, Hangzhou 310058, Zhejiang, China
*
Author to whom correspondence should be addressed.
Submission received: 26 April 2019 / Revised: 22 May 2019 / Accepted: 29 May 2019 / Published: 30 May 2019

Abstract

:
Long non-protein-coding RNAs (lncRNAs) identification and analysis are pervasive in transcriptome studies due to their roles in biological processes. In particular, lncRNA-protein interaction has plausible relevance to gene expression regulation and in cellular processes such as pathogen resistance in plants. While lncRNA-protein interaction has been studied in animals, there has yet to be extensive research in plants. In this paper, we propose a novel plant lncRNA-protein interaction prediction method, namely PLRPIM, which combines deep learning and shallow machine learning methods. The selection of an optimal feature subset and subsequent efficient compression are significant challenges for deep learning models. The proposed method adopts k-mer and extracts high-level abstraction sequence-based features using stacked sparse autoencoder. Based on the extracted features, the fusion of random forest (RF) and light gradient boosting machine (LGBM) is used to build the prediction model. The performances are evaluated on Arabidopsis thaliana and Zea mays datasets. Results from experiments demonstrate PLRPIM’s superiority compared with other prediction tools on the two datasets. Based on 5-fold cross-validation, we obtain 89.98% and 93.44% accuracy, 0.954 and 0.982 AUC for Arabidopsis thaliana and Zea mays, respectively. PLRPIM predicts potential lncRNA-protein interaction pairs effectively, which can facilitate lncRNA related research including function prediction.

1. Introduction

Ribonucleic acid (RNA) is a polymeric single-stranded molecule that constitutes living cells alongside deoxyribonucleic acid (DNA) and protein. Long non-coding RNAs (lncRNAs) are broad and myriad group of endogenous single-stranded polynucleotides non-protein coding transcripts with greater than 200 nucleotides sequence length [1]. Previously, lncRNAs were thought to be transcriptional noise because they are less efficiently spliced compared with messenger RNAs (mRNAs). Nevertheless, accumulated experimental evidence has suggested that lncRNAs have indispensable roles in development, hormone-dependent signaling, and stress responses in plants [2,3]. For example, lncRNA COLDAIR in Arabidopsis thaliana regulates flowering time and floral organ formation during vernalization [4]. Functional roles of lncRNAs in plants are lagging behind compared with other organisms such as animals, bacteria, and viruses [5]. The mechanism underlying the functions of lncRNAs is a captivating area of research. More and more studies have indicated that regulation of transcriptional, post-transcription processes and new insights into potential functions can be elucidated by lncRNAs binding with proteins [6]. Additionally, information about these interactions can hint at molecular causes of diseases in plants and animals [7]. lncRNA-protein interactions have been revealed by various experimental methods. For instance, using a fluorescence resonance energy transfer (FRET) approach coupled with fluorescence lifetime imaging microscopy (FLIM) to detect RNA-protein interactions in plant leaves [8]. Other experimental methods such as RNA-protein pull-down assays and RNA immunoprecipitation protocol have been adapted for detecting RNAs and their protein interaction partners in plants [9,10]. However, these wet-lab experimental methods are time-consuming, labor-intensive, and relatively few lncRNA-protein associations have experimental proof. Therefore, computational prediction of RNA and protein interaction partners to complement experimental approaches has become increasingly crucial.
To hasten research, many genome databases have been set up. The databases and web-based platforms available for human and mammals blend lncRNA information including functional annotation, phylogenetic conservation, associations with diseases, and relations between lncRNAs with other RNAs and proteins [5]. However, databases for plant lncRNAs are incomprehensive [5]. Over the last decade, many plant lncRNAs have been discovered by both experimental and computational screening [11]. Computational tools such as LncFinder [12], PlncPRO [13], PlantRNA Sniffer [14], and LncADeep [15] have produced a huge influx of plant lncRNA data. Two main approaches prevalent across lncRNA-protein interaction methods are feature-based and similarity-based [16]. Similarity-based methods assume that lncRNAs with similar functions interact with similar proteins, a concept known as ‘guilt-by-association’. Feature-based methods extract features from input data and employ methods such as support vector machine (SVM) to predict interactions between lncRNAs and proteins. In previous research, lncRNA-protein association prediction is based on biological characteristics such as sequence, structure, localization, and genomic context. For example, catRAPID [17] used physicochemical properties including secondary structure, hydrogen bonds, and van der Waals forces between proteins and lncRNAs to assess the tendency of interaction. Lu et al. developed lncPro [18] for lncRNA-protein interaction based on Van der Waals propensities, hydrogen bonding, and secondary structures. RPI-Pred [19] predicted ncRNA-protein interactions by integrating 3D structural and sequence features. LPI-ETSLP [20], a semi-supervised method based on sequence data, revealed lncRNA-protein association using eigenvalue transformation without the need for negative samples. Zhao et al. developed IRWNRLPI [21], a method that integrates random walk and neighborhood regularized logistic matrix factorization for forecasting lncRNA-protein association. Moreover, many network-based methods have been proposed to predict lncRNA-protein interaction based on the integration of heterogeneous networks including LPIHN, RWR, and LPI-NRLMF [22,23,24].
Traditional methods use hand crafted features that consume more time, require strong domain knowledge, and are problem-dependent. Therefore, it is immensely beneficial to develop a feature extraction method that is multi-layered (deep) to enhance representation power and provide insight into complex features [25]. Deep learning (DL) methods compute high-level representations from raw input. These models have been broadly utilized by researchers to understand the molecular mechanism in human and plant diseases [26,27,28]. As the hottest sub-field of machine learning, DL methods extract intrinsic features through multilayer architectures such as autoencoders (AE), convolutional neural networks (CNN), and recurrent neural networks (RNN) [29]. In computer vision, speech recognition, and text processing, DL has been observed to perform better than other popular machine learning methods [30]. It has also been successfully implemented in bioinformatics [31,32,33]. For example, DeepBind [31] is a DL-based model developed to identify sequence specificities of DNA and RNA binding protein. Similarly, DeepSEA [32] predicted noncoding-variant effect from chromatin-profiling sequences using DL. Recently, Pan et al. [34] predicted RNA-protein interaction by local-global deep CNN. Despite their significantly outstanding performance, selection of an optimal feature subset and subsequent efficient compression are still significant challenges in the existing methods. Therefore, deriving the most salient features is essential for improving performance. This can be achieved by imposing sparsity constraints including non-negativity on weights, weight decay regularization, and adding l1/l2-regularization terms to loss function [35,36]. l1/l2-regularization terms, referred to as mixed norm, minimizes reconstruction error in the AE model during training. Other constraints associated with AE include randomly corrupted data inputs, referred to as denoising AEs [37].
The most appealing characteristic of DL is the ability to extract high-level features to leverage the large number of unlabeled instances. Based on the extracted features, classifiers such as support vector machine (SVM) and random forest (RF) are used to build the prediction model. However, while a single classification model performs well, fusing multiple models through ensemble learning improves performance. An ensemble learning method is a meta-algorithm that trains several baseline models and combines them into a single predictive model. Apart from improved classification, ensemble methods perform well in problems involving noisy and imbalanced datasets [38]. Classification strengths of individual base classifiers selected for construction of the overall ensemble model lead to more accurate predictive performance. In 2011, a method named RPISeq [39] extracted 3-mer and 4-mer sequence features to train RF and SVM models for prediction of protein-RNA interaction. Then, Wang et al. [40] presented a model that predicted interactions between proteins and RNAs based on Naive Bayes (NB) and an extended NB classifier. In 2016, Pan et al. developed IPMiner, a sequence-based method for predicting lncRNA-protein interactions based on stacked autoencoder [41]. Yi et al. proposed RPI-SAN for lncRNA-protein interaction based on stacked autoencoder and RF [42]. lncLocator [43] predicted lncRNA subcellular localizations based on an ensemble of SVM and RF classifiers. A recent tool termed HLPI-Ensemble (human lncRNA-protein interaction) was proposed to predict human lncRNA-protein interactions based on SVM, extreme gradient boost (XGB), and RF [44]. For cancer prediction, a model based on DL with an ensemble of three classifiers was recently proposed by Xiao et al. [45].
In this study, we present a hybrid learning-based method, termed PLRPIM. Our objective is to predict lncRNA-protein interactions for plant species and demonstrate its significance in lncRNA annotation since most plant lncRNAs functions are unknown. Intrinsic high-level features are extracted by stacked sparse AE, a deep generative model. We use k-mer sparse matrices to represent lncRNA and protein sequences. Finally, the extracted features are fed into an ensemble of two classifier models, RF and light gradient boosting machine (LGBM) for prediction. The final outcome is computed by implementing a majority voting mechanism to the individual classification outcomes. The main contributions are: 1) We apply DL to extract high-level abstraction features from lncRNA and protein sequences; 2) adopt dropout, ReLU activation function, and l1/l2-regularization sparsity penalties to significantly reduce overfitting and improve performance; 3) fully exploit sequence information from k-mer and heterogeneous ensemble strategy to further improve prediction accuracy. Experimental results on two datasets including Arabidopsis thaliana and Zea mays show that our method performs better than other state-of-the-art methods such as RPISeq-RF, RPI-SAN, and IPMiner.

2. Materials and Methods

2.1. Dataset

We used sequence data from the plant lncRNA database (PlncRNADB) of lncRNA and their RNA-binding protein partners. The lncRNA-protein interaction data of Arabidopsis thaliana and Zea mays are downloaded from the website: http://bis.zju.edu.cn/PlncRNADB. Arabidopsis thaliana has 390 lncRNAs and 163 RNA-binding proteins whereas Zea mays has 1107 lncRNAs and 190 RNA-binding proteins. The dataset for Arabidopsis thaliana contained 948 positive samples (interactive pairs) and the Zea mays dataset contained 22,133 positive samples. The same number of negative pairs (948 and 22,133) were generated through randomly pairing proteins with lncRNAs and further removing the existing positive pairs [39]. Then, we used 20% of the data as the hold-out test set. Details of the datasets are shown in Table 1.

2.2. Interaction between LncRNA and Protein

We consider sequence features that are likely to be significant in predicting protein-lncRNA interactions. The complex molecular features include the number of atoms, hydrogen bonds, evolutionary conservation score, mutual interaction propensity, and electrostatic charges [46]. The interaction information between lncRNA and protein is derived from biological knowledge. The positive set consists of a concatenation of lncRNA-protein interacting pair li and pj while the negative set consists of non-interacting pairs. To obtain lncRNA-protein interactions, we denote an interaction matrix Y as Y(li, pj). Y(li, pj) ∊ {0,1} where 1 indicates an interaction between li and pj and 0 indicates no interaction. To define interaction profiles of lncRNA and protein, we use variables m = SW(li, lj) and n = SW(pi, pj), where SW refers to Smith-Waterman (SW) algorithm [21]. SW(li, lj) and SW(pi, pj) are the measures of similarity between sequence of lncRNA li and lj and sequences of protein pi and pj. The underlying presumption of m and n is that lncRNAs collaboratively interact with akin protein partners and vice versa.

2.3. Sequence Feature Encoding

Feature encoding is a key task during construction of a sequence-based machine learning model for prediction. Sequence feature encoding in this work is based on n-gram technique, also known as k-mer of length k where lncRNA is represented by (k1 = 1, 2, 3, 4) and proteins (k2 = 1, 2, 3) descriptors. In our experiments, protein and lncRNA sequences are coded using two methods: 1) Autocovariance and 2) conjoint triad. We investigate sequence similarities between lncRNAs and proteins to determine the combination of features. A feature vector of a varied length of size m for lncRNAs and n for proteins is generated. We calculate sequence similarity in lncRNAs using sequence information. Let SlncRNA represent a set of lncRNA sequence of interest. In order to compute similarities between lncRNA li and lj, the following formula is used:
S ln c R N A ( l i , l j ) = S W ( l i , l j ) max ( S W ( l i , l i ) , S W ( l j , l j ) )
We encode amino acid for protein sequences by natural numbers selected randomly. A feature vector of a varied length of size n is generated. Protein sequences are computationally analyzed by mapping amino acid bases into the numeric values. Let Sprotein represent a set of the protein sequence of interest. Given proteins i and j, we find the similarity between them based on SW dynamic programming algorithm using the following formula:
S Pr o t e i n ( p i , p j ) = S W ( p i , p j ) max ( S W ( p i , p i ) , S W ( p j , p j ) )

2.4. The PLRPIM Pipeline

Machine learning is applied in RNA-protein interaction (RPI) prediction to address the challenge of sparsity of known RPIs. The essence is to select effective feature combinations. In this study, we present a stacked AE framework with sparse constraints for lncRNA-protein interaction prediction, namely PLRPIM. PLRPIM is a neural network model described as follows. First, each known lncRNA-protein interaction pair is assigned 1 and 0 to non-interacting pairs. Secondly, AE maps a pair of input protein and lncRNA sequences to representations. Then, the integrated lncRNA and protein feature matrices { X i } i = 1 m and { W i } i = 1 n are transformed into feature vectors. Feature vector for lncRNA is represented as FlncRNA = {Fl1, Fl2,…, Flm} and Fprotein = {Fp1, Fp2,…, Fpn} for protein. Therefore, two subnetworks of lncRNA and protein are generated. Finally, a softmax layer is used to merge the two separate feature vectors and integrate them into a fully connected neural network to predict the label of each pair. lncRNA-protein interaction matrix Mij generated from predicted lncRNA-protein interaction matrix is the expected output matrix. To improve the performance of the proposed method in terms of robustness and accuracy, we adopt an integration strategy. We merge results from two machine-learning methods, RF and LGBM, to construct the prediction framework. The technical workflow of the proposed method is given in Figure 1.

2.4.1. Feature Extraction

Features are the sequence-based consolidated attributes of lncRNAs and proteins encoded into numeric vectors that are used in the prediction. In this work, we selected the k-mer model to extract features from lncRNA and protein, where the length of genetic sequence subset S is represented by an integer k. The k-mer method has been successfully used in extracting features from RNAs and amino acid composition of proteins based on local sequence patterns. RNA sequence consists of symbols from four alphabets, A, G, C, and T. RNA sequences are represented as the occurrence frequencies of k-mers following the principle of complementary pairing [47]. Each k-mer pattern is represented as 4k for each k value, where 4 is the number of RNA sequence nucleotides. For example, 16 patterns (e.g., “AA”,”AC”,”AT”,…) can be obtained when k = 2. We extracted 44 number of lncRNA sequence features by utilizing the k-mer scheme. Each lncRNA sequence is converted into a k-mer sparse matrix from which feature vectors are extracted. A protein chain has amino acids with 20 alphabets (A, C, D, E, F, G, H, I, K, L, M, N, P, Q, R, S, T, V, W, Y).
To obtain highly efficient features, we extracted a set of 599 descriptors from feature vectors encoded by various properties of lncRNAs and proteins. A total of 256 features are obtained from lncRNA sequence and 343 amino acid descriptors from protein sequence. We extract 4-mer sparse matrix of RNA sequence (A, C, G, T) by searching each sequence from left to right and obtain 256 (4 × 4 × 4 × 4) feature map. For the protein sequences, we partition amino acid compositions based on their chemical similarity. Seven groups of physicochemical properties {Val, Gly, Ala}, {Phe, Pro, Leu, Ile}, {Ser, Tyr, Met, Thr}, {His, Asn, Tpr, Gln}, {Arg, Lys}, {Glu, Asp}, and {Cys} of protein sequences were assigned numbers based on dipole moments (<1.0, <1.0, (1.0, 2.0), (2.0, 3.0), >3.0, >3.0, and <1.0) and their chain volumes (<50, >50, >50, >50, >50, >50, and <50) [19,41]. We extracted 3-mer counts of group labels to create a sparse matrix of 343 (7 × 7 × 7) feature map.

2.4.2. Hybrid Learning Method

DL architectures capture complex correlations in feature representations for diverse prediction tasks. The utmost significant property is learning features at multiple abstraction levels and feature concatenation within and across datasets. The core idea of stacked AE via sparse representations is to effectively obtain descriptors of data as linear projections that maximizes the correlation between features. AE is a kind of artificial neural network designed to extract and generate abstract features of high-dimensional data. It applies unsupervised learning algorithm to construct hidden structures from unlabeled data. The process of AE training consists of two components, an encoder and decoder. The encoder (function f) is used for mapping the input data (x, y) into latent representation and the decoder (function g) maps the encoded features to reconstruct input data from the latent representation. In this work, we formulate lncRNA-protein interaction prediction as a binary classification problem. We used constrained stacked AE (CSAE) network for DL and classification of training dataset as shown in Figure 2. The sparsity constraints extract the most informative features for optimal sample selection. Stacked AEs have greater expressive power and the successive layers of representations capture a hierarchical grouping of the input similar to convolution and pooling operations in CNN. Notations W = {W1, W2,, Wnl} and b = {b1, b2, …, bnl} denote the AE parameters. Weight matrix W denoted as W i j l is the weight associated with the connection between neuron unit j in layer l-1 and neuron unit i in layer l, where j = 1,2,…, sl - 1, i = 1,2,…, sl and l = 2,3,…,nl. sl - 1 represents the output of the previous layer while nl represents the number of layers in the network. Bias vector is denoted as b i l where i = 1, 2,…,sl and l = 2, 3,…,nl denotes the bias of neuron unit i in layer l. The training set is defined as {(x1, y1), (x2,y2),…,(xn, yn)} of n data samples. Suppose x is the input data of dimension d, the AE maps x to y as shown in the following formula:
y = f ( W x + b )
where f is the encoding function and Wx is a weight matrix that maps the output of x to a hidden-space. After mapping x to y, the output of the encoder (y) is mapped back to form z, which is the transformation of x with the same shape as x. The reconstruction is formulated as follow:
z = g ( W T y + b )
where g is the decoding function, non-linear function WT is the weighting matrix for transformation, and b′ is the transformation bias. The transformation error is estimated using squared error between x and z. To minimize and optimize the reconstruction error, parameters w and b are adjusted using stochastic gradient descent (SGD) [48]. Cross-entropy, the most common type of loss function, is used to minimize reconstruction error by AE. We use 256, 128, and 64 layers that are fully connected. We model the neurons output f(x) using rectified linear unit (ReLU) activation function:
f ( x ) = max ( x , 0 ) .
We trained the AEs by adding dropout layers. Dropout layers regularize the model to avoid the risk of overfitting by randomly leaving out some neuron units.

2.4.3. Training of PLRPIM Model with Mixed Norm Constraint

Inducing sparse constraints in AEs including weight decay and norm penalties has been successfully implemented to achieve high discriminative power of models. Inspired by [35,36], we implement mixed norm regularizer constraints to train stacked AE. The sparse l1 and l2 penalties for the weights are defined as follows:
l 1 = λ 2 i = 1 m W i , j l ,
l 2 = λ 2 i = 1 m W i , j l 2
where W i , j l denotes the connection between neuron unit i and j in layer l.
The proposed model is multi-layered and is trained in three phases. We use feed forward to optimize weights of the networks and adopt backpropagation in the last phase of training the network. The stacked AE-based generated networks are pre-trained before using backpropagation algorithm. The layers in the model include an input layer x with data points {(x1,y1), (x2,y2),…,(xn,yn) }, hidden layers {h1, h2}, and an output layer z (Figure 2). The features obtained from the hidden layers are used to reconstruct the input data and train the network. The first phase of the training process (pre-training) is realized by minimizing an objective function using cross-entropy (C) to avoid overfitting using Formulas (8) and (9) as follows:
C = ( y log z + ( 1 y ) log ( 1 z ) )
min w , b [ i = 1 m ( z x ) 2 + β j = 1 k C ]
where y are the labels of training examples, z represents the predicted output from input data x, m is the index of the number of training examples, and k is the index of the number of features. β is the trade-off parameter used to control overfitting. The second phase is made up of two hidden layers (h1 and h2). The third phase is the output layer, which receives input from the second hidden layer. In this phase, first, the weight and bias parameters (W3 and b3) are initialized. Then, a backpropagation algorithm is adapted to compute derivatives of the objective function and tune weights.
We apply mixed norm regularizer (l1/l2) with sparsity coefficient λ to minimize reconstruction error. The mixed norm regularizer is based on the assumption that descriptors from a set of training examples exhibit similar sparsity patterns in the feature map [49]. Formula (10) shows the computation of the mixed norm regularizer:
min w , b [ i = 1 m x z 2 + β j = 1 k C + λ 2 W 1 , 2 ] .
The parameter β is used for regularization, λ ≥ 0 is the weights sparsity term, W 1 , 2 represent l1 and l2 norms defined in Formulas (6) and (7) where W 1 = i = 1 m W i j l and W 2 = i = 1 m W i j l 2 . Formula (10) minimizes the average reconstruction error, increase sparsity of latent layer activation, and reduce the number of non-negative weights.

2.4.4. Shallow Classification Models

A vast number of classical machine learning methods have been employed in research and real-world systems to solve classification and regression problems. For RNA-protein interaction problem, SVM and RF shallow classification methods have previously been used [41]. Therefore, we consider these two and four other methods, adaptive boosting (Adaboost), XGB, LGBM, and decision tree (DT) to compare the performance of the single with merged classifiers. We assessed the prediction performance of the six commonly used classification methods. For most artificial intelligence applications, all the classification models are of high accuracy. However, each method may outperform others in different cases. We exemplify determining the most relevant kernel (radial basis function, linear, and polynomial) function challenge for SVM, while RF and gradient boosting decision tree (GBDT) may produce better classification results for the categories with more samples [45]. Therefore, a method that takes advantage of more than one classifier leads to superior prediction performance in practice.
The most widely used method for RNA-protein interaction prediction is SVM [50]. SVM is non-probabilistic classifier that maps input data points into a high-dimensional feature space. Other methods are ensemble methods, which are categorized into three classes: Bagging, boosting, and stacking. They combine several learning methods to obtain a predictive model with improved performance. They include RF, Adaboost, XGB, and LGBM. Given n models, fi is averaged into an ensemble e:
e ( x ) = 1 n i = 1 n f i ( x ) .
RF is a widely used shallow machine learning method that combine decision tree predictors following the bagging technique [45]. In this model, the class that receives majority votes from trees in the forest is considered the output result. This protocol relies on creating n number of models and averaging predictions of all models for a final prediction. In this work, the RF contained 50 decision trees with a minimum leaf size of 3 for each tree.
Adaptive boosting (AdaBoost) is an ensemble method often used to obtain satisfactory results compared with other methods [38]. It aims at converting a set of weak classifiers into a strong one. In this work, AdaBoost included 50 shallow decision trees.
GBDT is a machine learning technique comprised of a collection of decision trees to form a stronger prediction model. GBDT builds the model in a stage-wise fashion and trains it iteratively. It implements generalization by allowing optimization of an arbitrary differentiable loss function, which makes them efficient, accurate, and interpretable. Light gradient boosting (LGBM) is a tree-based learning method that implements GBDT [51]. It is suitable for the large size of data and a large number of features. It trains fast, utilizes low memory, and its accuracy is better. Unlike other methods that follow level-wise training pattern, LGBM and XGB follow a leaf-wise training approach [52].
XGB is an advanced GBDT method designed for speed, flexibility, and accuracy [53]. This tree ensemble model is trained in an adaptive manner. It has been used to build a prediction model for protein-protein interface residues [54]. In this work, the features for XGB were selected by its internal algorithm. Features for the classification models were selected by a backward elimination manner. The labels used by the classification methods for prediction were determined by stacked AE. The classification models were then used to test accuracy and robustness in test sets. Since our dataset is a pair-wise sample, the feature combinations with the highest predictive accuracy were selected and fed into an ensemble of the two models to evaluate the performance. Five-fold cross validation technique was used to evaluate the predictive ability of the proposed method. A hybrid method was employed, which combined RF and LGBM classifiers. These classifiers cannot directly work using raw sequence data. Therefore, an unsupervised DL model is implemented for extraction of features. We employ majority voting, a widely used method for the fusion of multiple classifiers.

2.4.5. Experimental Setup

In this study, we used Arabidopsis thaliana and Zea mays sequence datasets. The source codes for the experiments were written in python 2.7 programming language. For each dataset, we train our models using Keras library (https://github.com/fchollet/keras) with Tensorflow [55] as the backend. We separate the data in the ratio 4:1. Each time, four folds are used as the training set and one is withheld as the test set. We feed the training set into our models and use 20% as a test set. During the training process, test data are used to monitor convergence at the end of each epoch. Adding the number of neurons per layer indicates increasing the number of feature maps. In this experiment, we fix the depth of network to 3. We stack a combination of batch normalization, ReLU layer, and dropout layer to form the encoder. We use three different values for the learning rate (0.5, 1, 2), as shown in Table 2, and the dropping probability 0.6. The momentum update method [56] was employed with a momentum coefficient of 0.9. During the training, we set the maximum number of iterations to 50 for all the training examples. For each iteration, training examples are randomly shuffled to speed up the training process. We set a mini-batch training with a batch size of 64 (m = 64).
The model is fine-tuned by back-propagation using categorical cross-entropy loss function, which is optimized by SGD with momentum. Other hyper-parameter values that were optimized during training were the learning decay rate (1e-6) and a mean squared error (MSE). Mixed-norm constraints (l1 = 0.01 and l2 = 0.01) are used to minimize reconstruction error. We exploit non-linear activation function on hidden neurons to combat nonlinearity in lncRNA-protein interactions. Encoder function uses ReLU activation function while the decoder function uses sigmoid activation function.
Additionally, we employ batch normalization and early stopping regularization techniques to avoid overfitting. More details on parameter settings for our model are shown in Table 2. We observe that initially, the performance of PLRPIM increases with the depth of the network. However, the performance drops when the depth of the network is greater than three. This is due to overfitting since increasing the number of hidden layers decreases the training data. Figure 3 summarizes the steps followed during the testing of the proposed method.

2.4.6. PLRPIM Method

PLRPI predicts based on the harmonious blend of information from concatenation fusion of sparse representations of lncRNA and protein sequences. The pseudocode of the proposed method is presented as follows (Algorithm 1).
  Algorithm 1: Pseudo code of PLRPIM
  Input: Y—adjacency matrix of interacting lncRNA i and protein j; (Y(li, pj)), SlncRNA—set of lncRNA sequence of interest, Sprotein—set of protein sequence of interest, number of stacked AEs = T, number of iterations (epoch) = R
  Output: lncRNA-protein interaction matrix Mij
 1.B = {1, 0}—Binary domain;
 2.Interact: SlncRNA × Sprotein B;
 3.Interact: (S(li), S(pi))- check whether lncRNA li and protein pi interact; if true, return 1; otherwise, return 0;
 4.Concatenate: SlncRNA × Sprotein Y;
 5.Initialize training examples labels (Y(li, pj)) = 0;
 6.Fort = 1 to T do;
 7.Forr = 1 to R do;
 8.Minimize objective function using formula (10);
 9.Compute number of features from similarity matrices { X i } i = 1 m and { W i } i = 1 n ;
 10.Compute hidden vector { Z i } n = 1 N with learned AE;
 11.Update { Z i } n = 1 N ;
 12.End for;
 13.Compute feature vectors {Fl1, Fl2,…, Flm} and {Fp1, Fp2,…, Fpn};
 14.Predict class labels of the test dataset based on ensemble voting process using formula (11);
 15.Update training examples labels (Y(li, pj)) = 1;
 16.End for;
 17.Return:Mij.

2.5. Evaluation of PLRPIM

To empirically assess the prediction capability of the proposed model, we adopt 5-fold cross-validation as well as some commonly used measures. Cross-validation iteratively partitions the dataset into training and test sets to reduce sampling bias. The data were randomly split into five sets of equal size. One fold is withheld as the test set and the remaining four are used as the training set. We use six standard metrics including area under ROC curve (AUC), precision (PRE), sensitivity (SEN), specificity (SPEC), accuracy (ACC), and Mathews correlation coefficient (MCC) to evaluate the average performance. The probability of assigning a high rank to a randomly chosen positive instance over a negative one is plotted on the AUC curve. The greater the value of AUC, the better the predictive power of the model. The outcomes from a classifier can be represented as a confusion matrix. The confusion matrix contains four categories of outcomes, true positive (TP), true negative (TN), false positive (FP), and false negative (FN). TP are actual lncRNA-protein pairs that are predicted correctly, TN are the pairs correctly extracted from the negative samples, FP are negative entities falsely predicted, whereas FN are the cases where actual lncRNA-protein pairs are predicted as non-interacting.
P R E = T P T P + F P
S E N = T P T P + F N
S P E = T N T N + F P
A C C = T P + T N T P + T N + F P + F N
M C C = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )

3. Results

In this section, we present 5-fold cross-validation results of PLRPIM on two datasets. Table 3 shows the results obtained from each of the five folds, the average value, and the standard deviation.
The proposed method has the highest AUC followed by specificity, also referred to as true negative rate. This indicates that our method identifies the percentage of actual non-interactive pairs more accurately for both datasets. Additionally, the standard deviation values showing the variation in the values obtained by each fold are close to zero, indicating low-variance. This shows the uniformity in the results, thus evidencing that our method has reliable performance.

3.1. Comparison between Classification Models

We applied six classification methods individually including SVM, AdaBoost, DT, XGB, LGBM, DT, and RF. Then, we presented the average prediction results obtained from 5-fold cross-validation technique. SVM does not perform well compared with the other methods in our experiments. Therefore, we do not include it in the results presented in Table 4. The strength of our pipeline is based on a heterogeneous ensemble strategy, which effectively reduces the false positive rate and achieves better performance. Integration of baseline models trained by different high-level feature combinations boosted the performance.
We compare our multi-model ensemble method with other classification methods, as presented in Table 4. Results obtained indicate that the proposed method increases prediction accuracy for the datasets compared with single classifiers. From the results presented in Table 4, our method performed better in terms of accuracy, precision, specificity, MCC, and AUC for the Arabidopsis thaliana dataset. LGBM had the highest sensitivity. For the Zea mays dataset, PLRPIM had the highest accuracy, precision, sensitivity, specificity, MCC, and AUC.
Figure 4a shows that our method had better performance in terms of AUC on the Arabidopsis thaliana dataset. In comparison to the other methods, our method had approximately 15% better performance. The contribution of the two models integrated in PLRPIM significantly boosted the performance. Figure 4b shows that our method had better performance in terms of AUC on the Zea mays dataset. In comparison to the other methods, our method had approximately 13% better performance. The performance of all the methods increased in terms of AUC in the Zea mays dataset due to the substantial increase in the size of the dataset.

3.2. Comparison of PLRPIM with Other Existing Tools

To test the reliability of the proposed method, we compare PLRPIM with three other sequence-based computational models, RPISeq-RF, RPI-SAN, and IPMiner. The performance of each tool in terms of accuracy, precision, sensitivity, specificity, and AUC was compared. In terms of accuracy, PLRPIM performed better with 89.98% and 93.44% accuracy for the two plants. The AUC values of PLRPIM, IPMiner, RPISeq-RF, and RPI-SAN for Arabidopsis thaliana are 0.9546, 0.8823, 0.8761, and 0.8164, respectively, as shown in Figure 5a. For the Zea mays dataset, the AUC values are 0.9823, 0.9034, 0.8980, and 0.8792, respectively, as shown in Figure 5b. Table 5 shows the performance of the four methods.
By tapping on the performance benefits of sparsity constraints, our model learns the most informative sequence features. In Table 5, our method performs better that the other methods in terms of accuracy, precision, sensitivity, specificity, MCC, and AUC for both Arabidopsis thaliana and Zea mays datasets.
Figure 5a shows that our method has better performance in terms of AUC on the Arabidopsis thaliana dataset. In comparison to the other methods, our method had approximately 9% better performance in terms of AUC. Figure 5b shows that our method has better performance in terms of AUC on the Zea mays dataset. In comparison to the other methods, our method has approximately 8% increase in AUC. Increasing the amount of data positively affects the performance of the model. The better performance is attributed to data size and employing LGBM. LGBM model boosts prediction power since it is suitable for large size of data. All methods are positively influenced by the increase in data size. Their performance in terms of AUC increase.

3.3. Functional Analysis of lncRNAs

We explore Arabidopsis thaliana and Zea mays lncRNA-protein networks for annotation of lncRNAs. For the Arabidopsis thaliana, TCONS_00011717, TCONS_00008833, and TCONS_00012080 interact with proteins IPR012677 and P41377. IPR012677 functions include oxidoreductase activity, nucleotide binding, and nucleotide acid binding. P41377 functions include ATP binding, mRNA binding, and helicase activity. In our Zea mays dataset, some genes lack annotations and are referred to as “orphan genes”. According to Arendsee et al., orphan genes are defined as genes whose coding sequences are specific to species [57]. Many orphan genes functions are unknown due to lack of homologs and their uniqueness to specific species. Therefore, identifying their protein interaction partners is an alternative way of annotating them. Some orphan genes without functions include GRMZM5G870099, GRMZM2G562000, GRMZM2G046326, GRMZM5G849473, GRMZM2G107204, and GRMZM2G375531.
As an illustration, we provide Table 6 to show the significance of the interaction between lncRNAs and proteins in plant development and regulation of gene expression. For example, the above-mentioned orphan genes and GRMZM2G374777 (orphan gene), GRMZM2G097084 (orphan gene), GRMZM2G078523, GRMZM2G543070, and GRMZM2G147020 interact with protein B6SIF0. This protein has abiotic stress-related functions such as response to cold, response to water deprivation, and response to osmotic stress. We obtain GO annotation terms from EnsemblPlants database (https://plants.ensembl.org) and maize genetics and genomics database (https://maizegdb.org). We observe that the lncRNAs interacting with the same protein share common functions such as two-leaf expansion stage as shown in Table 6.

4. Conclusions

In this paper, we presented PLRPIM, a hybrid method for the prediction of plant lncRNA and protein interactions. The proposed method consists of CSAE for feature selection and reduction of feature space dimension, a feed forward, and ensemble classifiers for prediction. The prediction performances demonstrate its efficiency in comparison to other methods. From our results, the hybrid of heterogeneous ensemble classifiers and unsupervised learning implemented by the proposed method performs well without dataset size limitation. By fully exploiting multiple classifiers, the proposed method is shown to have a high success rate for lncRNA-protein interaction prediction based on genome sequence. Additionally, the proposed method is versatile and can be used to predict lncRNA-protein interaction for other plants. However, the proposed method has several potential limitations that need to be addressed in the future. First, the degree of research for plant lncRNA-related proteins for different plant species is limited; thus, known lncRNA-binding protein partners are sparse. Information bias can mislead the measurement of interaction probability between lncRNAs and proteins in plants. More data sources with experimentally validated data can potentially improve the performance further.

Author Contributions

J.S.W. and J.M. designed the methods and conceived the study. J.S.W. and J.M. implemented the entire method and performed the experiments. M.C. collected datasets and revised the manuscript and comments. J.M., M.C., and Y.L. analyzed the prediction results. J.S.W. and J.M. wrote the manuscript. All authors read and approved the final manuscript.

Funding

This study was supported by the National Natural Science Foundation of China (Nos. 61872055, 31872116).

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. D’Aniello, S.; Spagnuolo, A.; Ceccarelli, M.; Cerulo, L.; Ventola, G.M.M.; Noviello, T.M.R.; D’Aniello, S. Identification of long non-coding transcripts with feature selection: a comparative study. BMC Bioinform. 2017, 18, 187. [Google Scholar]
  2. Lu, X.; Chen, X.; Mu, M.; Wang, J.; Wang, X.; Wang, D.; Yin, Z.; Fan, W.; Wang, S.; Guo, L.; Ye, W. Genome-Wide Analysis of Long Noncoding RNAs and Their Responses to Drought Stress in Cotton (Gossypium hirsutum L.). PLoS ONE 2016, 11, e0156723. [Google Scholar] [CrossRef]
  3. Nejat, N.; Mantri, N. Emerging roles of long non-coding RNAs in plant response to biotic and abiotic stresses. Crit. Rev. Biotechnol. 2018, 38, 93–105. [Google Scholar] [CrossRef]
  4. Kim, D.H.; Xi, Y.; Sung, S. Modular function of long noncoding RNA, COLDAIR, in the vernalization response. PLOS Genet. 2017, 13, e1006939. [Google Scholar] [CrossRef]
  5. Bhatia, G.; Goyal, N.; Sharma, S.; Upadhyay, S.K.; Singh, K. Present Scenario of Long Non-Coding RNAs in Plants. Non-Coding RNA 2017, 3, 16. [Google Scholar] [CrossRef] [PubMed]
  6. Kashi, K.; Henderson, L.; Bonetti, A.; Carninci, P. Discovery and functional analysis of lncRNAs: Methodologies to investigate an uncharacterized transcriptome. Biochim. et Biophys. Acta (BBA)-Gene Regul. Mech. 2016, 1859, 3–15. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Xu, Y.; Wu, W.; Han, Q.; Wang, Y.; Li, C.; Zhang, P.; Xu, H. New Insights into the Interplay between Non-Coding RNAs and RNA-Binding Protein HnRNPK in Regulating Cellular Functions. Cells 2019, 8, 62. [Google Scholar] [CrossRef]
  8. Camborde, L.; Jauneau, A.; Brière, C.; Deslandes, L.; Dumas, B.; Gaulin, E. Detection of nucleic acid-protein interactions in plant leaves using fluorescence lifetime imaging microscopy. Nat. Protoc. 2017, 12, 1933–1950. [Google Scholar] [CrossRef]
  9. Bierhoff, H. Analysis of lncRNA-Protein Interactions by RNA-Protein Pull-Down Assays and RNA Immunoprecipitation (RIP). Cell. Quiescence 2018, 1686, 241–250. [Google Scholar]
  10. Mermaz, B.; Liu, F.; Song, J. RNA Immunoprecipitation Protocol to Identify Protein-RNA Interactions in Arabidopsis thaliana. Plant Chromatin Dyn. 2018, 1675, 331–343. [Google Scholar]
  11. Liu, X.; Hao, L.; Li, D.; Zhu, L.; Hu, S. Long non-coding RNAs and their biological roles in plants. Genom. Proteom. Bioinform. 2015, 13, 137–147. [Google Scholar] [CrossRef]
  12. Han, S.; Liang, Y.; Ma, Q.; Xu, Y.; Zhang, Y.; Du, W.; Wang, C.; Li, Y. LncFinder: An integrated platform for long non-coding RNA identification utilizing sequence intrinsic composition, structural information and physicochemical property. Briefings Bioinform. 2018. [Google Scholar] [CrossRef]
  13. Singh, U.; Khemka, N.; Rajkumar, M.S.; Garg, R.; Jain, M. PLncPRO for prediction of long non-coding RNAs (lncRNAs) in plants and its application for discovery of abiotic stress-responsive lncRNAs in rice and chickpea. Nucleic Acids Res. 2017, 45, e183. [Google Scholar] [CrossRef]
  14. Vieira, L.M.; Grativol, C.; Thiebaut, F.; Carvalho, T.G.; Hardoim, P.R.; Hemerly, A.; Lifschitz, S.; Ferreira, P.C.G.; Walter, M.E.M.T. PlantRNA_Sniffer: A SVM-Based Workflow to Predict Long Intergenic Non-Coding RNAs in Plants. Non-Coding RNA 2017, 3, 11. [Google Scholar] [CrossRef]
  15. Yang, C.; Yang, L.; Zhou, M.; Xie, H.; Zhang, C.; Wang, M.D.; Zhu, H. LncADeep: an ab initio lncRNA identification and functional annotation tool based on deep learning. Bioinformatics 2018, 34, 3825–3834. [Google Scholar] [CrossRef]
  16. Zhang, W.; Yue, X.; Tang, G.; Wu, W.; Huang, F.; Zhang, X. SFPEL-LPI: Sequence-based feature projection ensemble learning for predicting LncRNA-protein interactions. PLOS Comput. Boil. 2018, 14, e1006616. [Google Scholar] [CrossRef]
  17. Bellucci, M.; Agostini, F.; Masin, M.; Tartaglia, G.G. Predicting protein associations with long noncoding RNAs. Nat. Methods 2011, 8, 444–445. [Google Scholar] [CrossRef]
  18. Lu, Q.; Ren, S.; Lu, M.; Zhang, Y.; Zhu, D.; Zhang, X.; Li, T. Computational prediction of associations between long non-coding RNAs and proteins. BMC Genom. 2013, 14, 651. [Google Scholar] [CrossRef]
  19. Suresh, V.; Liu, L.; Adjeroh, D.; Zhou, X. RPI-Pred: Predicting ncRNA-protein interaction using sequence and structural information. Nucleic Acids Res. 2015, 43, 1370–1379. [Google Scholar] [CrossRef] [PubMed]
  20. Hu, H.; Zhu, C.; Ai, H.; Zhang, L.; Zhao, J.; Zhao, Q.; Liu, H. LPI-ETSLP: lncRNA–protein interaction prediction using eigenvalue transformation-based semi-supervised link prediction. Mol. BioSyst. 2017, 13, 1781–1787. [Google Scholar] [CrossRef] [PubMed]
  21. Zhao, Q.; Zhang, Y.; Hu, H.; Ren, G.; Zhang, W.; Liu, H. IRWNRLPI: Integrating Random Walk and Neighborhood Regularized Logistic Matrix Factorization for lncRNA-Protein Interaction Prediction. Front. Genet. 2018, 9, 239. [Google Scholar] [CrossRef] [Green Version]
  22. Zhang, Y.; Wang, M.; Li, A.; Ge, M.; Peng, C. Predicting Long Noncoding RNA and Protein Interactions Using Heterogeneous Network Model. BioMed Res. Int. 2015, 2015. [Google Scholar] [CrossRef]
  23. Ge, M.; Li, A.; Wang, M. A Bipartite Network-based Method for Prediction of Long Non-coding RNA-protein Interactions. Genom. Proteom. Bioinform. 2016, 14, 62–71. [Google Scholar] [CrossRef]
  24. Liu, H.; Ren, G.; Hu, H.; Zhang, L.; Ai, H.; Zhang, W.; Zhao, Q. LPI-NRLMF: lncRNA-protein interaction prediction by neighborhood regularized logistic matrix factorization. Oncotarget 2017, 8, 103975–103984. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, W.; Li, R.; Zeng, T.; Sun, Q.; Kumar, S.; Ye, J.; Ji, S. Deep Model Based Transfer and Multi-Task Learning for Biological Image Analysis. IEEE Trans. Big Data 2016. [Google Scholar] [CrossRef]
  26. Zhang, L.; Lv, C.; Jin, Y.; Cheng, G.; Fu, Y.; Yuan, D.; Tao, Y.; Guo, Y.; Ni, X.; Shi, T. Deep Learning-Based Multi-Omics Data Integration Reveals Two Prognostic Subtypes in High-Risk Neuroblastoma. Front. Genet. 2018, 9, 477. [Google Scholar] [CrossRef] [Green Version]
  27. Fuentes, A.F.; Yoon, S.; Lee, J.; Park, D.S. High-Performance Deep Neural Network-Based Tomato Plant Diseases and Pests Diagnosis System With Refinement Filter Bank. Front. Plant Sci. 2018, 9, 1162. [Google Scholar] [CrossRef] [Green Version]
  28. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using Deep Learning for Image-Based Plant Disease Detection. Front. Plant Sci. 2016, 7, 346. [Google Scholar] [CrossRef]
  29. Liu, G.; Bao, H.; Han, B. A Stacked Autoencoder-Based Deep Neural Network for Achieving Gearbox Fault Diagnosis. Math. Probl. Eng. 2018, 2018, 1–10. [Google Scholar] [CrossRef] [Green Version]
  30. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  31. Alipanahi, B.; Delong, A.; Weirauch, M.T.; Frey, B.J. Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning. Nat. Biotechnol. 2015, 33, 831–838. [Google Scholar] [CrossRef]
  32. Zhou, J.; Troyanskaya, O.G. Predicting effects of noncoding variants with deep learning-based sequence model. Nat. Methods 2015, 12, 931–934. [Google Scholar] [CrossRef]
  33. Wang, Y.; You, Z.-H.; Yang, S.; Li, X.; Jiang, T.-H.; Zhou, X. A High Efficient Biological Language Model for Predicting Protein-Protein Interactions. Cells 2019, 8, 122. [Google Scholar] [CrossRef]
  34. Pan, X.; Shen, H.-B. Predicting RNA-protein binding sites and motifs through combining local and global deep convolutional neural networks. Bioinformatics 2018, 34, 3427–3436. [Google Scholar] [CrossRef]
  35. Hosseini-Asl, E.; Zurada, J.M.; Nasraoui, O. Deep Learning of Part-Based Representation of Data Using Sparse Autoencoders With Nonnegativity Constraints. IEEE Trans. Neural Networks Learn. Syst. 2016, 27, 1–13. [Google Scholar] [CrossRef]
  36. Halkias, X.; Paris, S.; Glotin, H. Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint. Available online: https://arxiv.org/abs/1301.3533 (accessed on 27 December 2018).
  37. Vincent, P.; LaRochelle, H.; Bengio, Y.; Manzagol, P.-A. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th international conference on Machine learning, Helsinki, Finland, 5–9 July 2008; pp. 1096–1103. [Google Scholar]
  38. Van Der Laan, M.J.; Polley, E.C.; Hubbard, A.E. Super learner in prediction. U.C. Berkeley Division of Biostatistics Working Paper Series. Available online: https://biostats.bepress.com/ucbbiostat/paper222/ (accessed on 1 November 2018).
  39. Muppirala, U.K.; Honavar, V.G.; Dobbs, D. Predicting RNA-Protein Interactions Using Only Sequence Information. BMC Bioinform. 2011, 12, 489. [Google Scholar] [CrossRef]
  40. Wang, Y.; Xu, D.; Chen, R.; Chen, L.; Wang, Y.; Chen, X.; Liu, Z.-P.; Huang, Q.; Zhang, X.-S. De novo prediction of RNA–protein interactions from sequence information. Mol. BioSyst. 2013, 9, 133–142. [Google Scholar] [CrossRef]
  41. Pan, X.; Fan, Y.-X.; Yan, J.; Shen, H.-B. IPMiner: hidden ncRNA-protein interaction sequential pattern mining with stacked autoencoder for accurate computational prediction. BMC Genom. 2016, 17, 861. [Google Scholar] [CrossRef]
  42. Yi, H.-C.; You, Z.-H.; Huang, D.-S.; Li, X.; Jiang, T.-H.; Li, L.-P. A Deep Learning Framework for Robust and Accurate Prediction of ncRNA-Protein Interactions Using Evolutionary Information. Mol. Ther. Nucleic Acids 2018, 11, 337–344. [Google Scholar] [CrossRef] [Green Version]
  43. Cao, Z.; Pan, X.; Yang, Y.; Huang, Y.; Shen, H.-B.; Zhen, C. The lncLocator: a subcellular localization predictor for long non-coding RNAs based on a stacked ensemble classifier. Bioinformatics 2018, 34, 2185–2194. [Google Scholar] [CrossRef]
  44. Hu, H.; Zhang, L.; Ai, H.; Zhang, H.; Fan, Y.; Zhao, Q.; Liu, H. HLPI-Ensemble: Prediction of human lncRNA-protein interactions based on ensemble strategy. RNA Boil. 2018, 15, 1–10. [Google Scholar] [CrossRef] [Green Version]
  45. Xiao, Y.; Wu, J.; Lin, Z.; Zhao, X. A deep learning-based multi-model ensemble method for cancer prediction. Comput. Methods Programs Biomed. 2018, 153, 1–9. [Google Scholar] [CrossRef]
  46. Liu, Z.-P.; Miao, H. Prediction of protein-RNA interactions using sequence and structure descriptors. Neurocomputing 2016, 206, 28–34. [Google Scholar] [CrossRef]
  47. Liu, B. BioSeq-Analysis: A platform for DNA, RNA and protein sequence analysis based on machine learning approaches. Briefings Bioinform. 2017, 165. [Google Scholar] [CrossRef] [PubMed]
  48. Le, Q.V. Building high-level features using large scale unsupervised learning. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 8595–8598. [Google Scholar]
  49. Sankaran, A.; Vatsa, M.; Singh, R.; Majumdar, A. Group sparse autoencoder. Image Vision Comput. 2017, 60, 64–74. [Google Scholar] [CrossRef]
  50. Zhang, X.; Liu, S. RBPPred: Predicting RNA-binding proteins from sequence using SVM. Bioinformatics 2017, 33, 854–862. [Google Scholar] [CrossRef] [PubMed]
  51. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. Lightgbm: A highly efficient gradient boosting decision tree. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc: Red Hook, NY, USA, 2017; pp. 3146–3154. [Google Scholar]
  52. Zhan, Z.H.; You, Z.H.; Li, L.P.; Zhou, Y.; Yi, H.C. Accurate Prediction of ncRNA-Protein Interactions From the Integration of Sequence and Evolutionary Information. Front Genet. 2018, 9, 458. [Google Scholar] [CrossRef] [PubMed]
  53. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  54. Wang, H.; Liu, C.; Deng, L. Enhanced Prediction of Hot Spots at Protein-Protein Interfaces Using Extreme Gradient Boosting. Sci. Rep. 2018, 8, 14285. [Google Scholar] [CrossRef]
  55. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. Available online: https://www.usenix.org/conference/osdi16/technical-sessions/presentation/abadi (accessed on 30 October 2018).
  56. Qian, N. On the momentum term in gradient descent learning algorithms. Neural Networks 1999, 12, 145–151. [Google Scholar] [CrossRef]
  57. Arendsee, Z.W.; Li, L.; Wurtele, E.S. Coming of age: orphan genes in plants. Trends Plant. Sci. 2014, 19, 698–708. [Google Scholar] [CrossRef]
Sample Availability: The code and datasets for implementing the proposed method are available at https://github.com/Mjwl/PLRPIM.
Figure 1. The workflow of PLRPIM model.
Figure 1. The workflow of PLRPIM model.
Cells 08 00521 g001
Figure 2. The network structure of constrained stacked autoencoders (AE).
Figure 2. The network structure of constrained stacked autoencoders (AE).
Cells 08 00521 g002
Figure 3. Experimental setup for testing the proposed method. The datasets are split into five folds and hyper parameters are tuned in the training set. The model is learned and applied on the test set based on the hyper parameters. Performance metrics are calculated for all folds.
Figure 3. Experimental setup for testing the proposed method. The datasets are split into five folds and hyper parameters are tuned in the training set. The model is learned and applied on the test set based on the hyper parameters. Performance metrics are calculated for all folds.
Cells 08 00521 g003
Figure 4. ROC curves for PLRPIM, light gradient boosting machine (LGBM), extreme gradient boost (XGB), random forest (RF), adaptive boosting (AdaBoost), and decision tree (DT) on (a) Arabidopsis thaliana dataset and (b) Zea mays dataset.
Figure 4. ROC curves for PLRPIM, light gradient boosting machine (LGBM), extreme gradient boost (XGB), random forest (RF), adaptive boosting (AdaBoost), and decision tree (DT) on (a) Arabidopsis thaliana dataset and (b) Zea mays dataset.
Cells 08 00521 g004
Figure 5. ROC curves for PLRPIM, IPMiner, RPISeq-RF, and RPI-SAN on (a) Arabidopsis thaliana and (b) Zea mays datasets.
Figure 5. ROC curves for PLRPIM, IPMiner, RPISeq-RF, and RPI-SAN on (a) Arabidopsis thaliana and (b) Zea mays datasets.
Cells 08 00521 g005
Table 1. Details of lncRNA–protein interaction datasets for two species.
Table 1. Details of lncRNA–protein interaction datasets for two species.
SpeciesDatasetPositive SamplesNegative SamplesTotal
Arabidopsis thalianaTraining set7587581516
Test set190190380
Zea maysTraining set17,70617,70635,412
Test set442744278854
Table 2. Hyper parameter settings.
Table 2. Hyper parameter settings.
NameSettings
Learning rate0.5, 1, 2
Parameter optimizationSGD, momentum, Adam
Batch size256, 128, 64
ActivationReLU
LossMSE, Cross Entropy
Dropout rate0.6
Table 3. 5-fold results of the proposed method.
Table 3. 5-fold results of the proposed method.
DatasetTest SetACCPRESENSPEMCCAUC
Arabidopsis thaliana10.90000.92500.85060.94170.79950.9527
20.88650.93680.83590.94020.77840.9488
30.89710.93440.86360.93370.79700.9590
40.90500.93530.86410.94360.81170.9490
50.91030.92230.90360.91760.82060.9615
Average0.8998 ± 0.0080.9308 ± 0.0060.8636 ± 0.020.9354 ± 0.0090.8015 ± 0.010.9542 ± 0.005
Zea mays10.93720.93940.93380.94050.87440.9849
20.93400.93800.93130.93690.86810.9846
30.93120.93180.93030.93210.86240.9817
40.93100.93500.92590.93610.86200.9812
50.93870.93670.94080.93660.87730.9833
Average0.9344 ± 0.0030.9362 ± 0.0030.9324 ± 0.0050.9364 ± 0.0030.8689 ± 0.0060.9831 ± 0.001
Table 4. Predictive performance of classifiers and the proposed method.
Table 4. Predictive performance of classifiers and the proposed method.
DatasetMethodACCPRESENSPEMCCAUC
Arabidopsis thalianaPLRPIM0.89980.93080.86360.93540.80150.9542
LGBM0.89500.91820.86680.92300.79120.8949
XGB0.74520.76150.76610.71870.49930.8352
RF0.70880.68510.81640.59510.42940.8171
AdaBoost0.69620.68290.82340.56220.42140.8061
DT0.62330.63310.60600.63590.24780.6284
Zea maysPLRPIM0.93440.93620.93240.93640.86880.9831
LGBM0.93170.93310.93000.93330.86340.9317
XGB0.79360.76890.84260.74460.59090.8862
AdaBoost0.78490.76760.81820.75160.57250.8693
RF0.75360.74070.79720.71030.51110.8641
DT0.65230.65000.66760.63680.30490.6894
Table 5. Predictive performance of other methods and the proposed method.
Table 5. Predictive performance of other methods and the proposed method.
DatasetMethodACCPRESENSPEMCCAUC
Arabidopsis thalianaPLRPIM0.89980.93080.86360.93540.80150.9546
IPMiner0.82750.89300.74480.91070.66460.8823
RPISeq-RF0.80590.81440.79220.82000.61240.8761
RPI-SAN0.75790.79550.69660.81990.52100.8164
Zea maysPLRPIM0.93440.93620.93240.93640.86880.9823
IPMiner0.81270.81420.81060.81480.62580.9034
RPISeq-RF0.80690.79930.81920.79450.61420.8980
RPI-SAN0.78900.79090.78690.79110.57840.8792
Table 6. Some selected Zea mays and Arabidopsis thaliana long non-coding RNAs (lncRNAs) and their biological functions.
Table 6. Some selected Zea mays and Arabidopsis thaliana long non-coding RNAs (lncRNAs) and their biological functions.
SpecieslncRNAsBiological Functions
Arabidopsis thalianaTCONS_00011717GO:0006913—Nucleocytoplasmic transport
TCONS_00008833GO:0006083—Acetate metabolic process
TCONS_00012080GO:0009867—Regulation of jasmonic acid mediated
       signaling pathway
Zea maysGRMZM2G374777PO:0001052—2 leaf expansion stage
PO:0006339—juvenile vascular leaf
PO:0009089—endosperm
GRMZM2G097084GO:0005524—ATP binding
GO:0006183—GTP biosynthetic process
GO:0006952—defense response
GRMZM2G078523PO:0001052—2 leaf expansion stage
PO:0009084—pericarp
PO:0009089—endosperm
GRMZM2G543070PO:0001052—2 leaf expansion stage
PO:0001095—true leaf formation stage
PO:0007016—4 flowering stage
GRMZM2G147020PO:0001052—2 leaf expansion stage
PO:0001095—true leaf formation stage
PO:0007016—4 flowering stage
PO:0001009—D pollen mother cell
       meiosis stage

Share and Cite

MDPI and ACS Style

Wekesa, J.S.; Luan, Y.; Chen, M.; Meng, J. A Hybrid Prediction Method for Plant lncRNA-Protein Interaction. Cells 2019, 8, 521. https://0-doi-org.brum.beds.ac.uk/10.3390/cells8060521

AMA Style

Wekesa JS, Luan Y, Chen M, Meng J. A Hybrid Prediction Method for Plant lncRNA-Protein Interaction. Cells. 2019; 8(6):521. https://0-doi-org.brum.beds.ac.uk/10.3390/cells8060521

Chicago/Turabian Style

Wekesa, Jael Sanyanda, Yushi Luan, Ming Chen, and Jun Meng. 2019. "A Hybrid Prediction Method for Plant lncRNA-Protein Interaction" Cells 8, no. 6: 521. https://0-doi-org.brum.beds.ac.uk/10.3390/cells8060521

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop