Next Article in Journal
Telmisartan Lowers Elevated Blood Pressure in Psoriatic Mice without Attenuating Vascular Dysfunction and Inflammation
Next Article in Special Issue
LDAPred: A Method Based on Information Flow Propagation and a Convolutional Neural Network for the Prediction of Disease-Associated lncRNAs
Previous Article in Journal
Highly Efficient and Heritable Targeted Mutagenesis in Wheat via the Agrobacterium tumefaciens-Mediated CRISPR/Cas9 System
Previous Article in Special Issue
FKRR-MVSF: A Fuzzy Kernel Ridge Regression Model for Identifying DNA-Binding Proteins by Multi-View Sequence Features via Chou’s Five-Step Rule
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CNNDLP: A Method Based on Convolutional Autoencoder and Convolutional Neural Network with Adjacent Edge Attention for Predicting lncRNA–Disease Associations

1
School of Computer Science and Technology, Heilongjiang University, Harbin 150080, China
2
School of Mathematical Science, Heilongjiang University, Harbin 150080, China
3
School of Information Science and Technology, Heilongjiang University, Harbin 150080, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2019, 20(17), 4260; https://doi.org/10.3390/ijms20174260
Submission received: 27 June 2019 / Revised: 28 August 2019 / Accepted: 28 August 2019 / Published: 30 August 2019
(This article belongs to the Special Issue Special Protein or RNA Molecules Computational Identification 2019)

Abstract

:
It is well known that the unusual expression of long non-coding RNAs (lncRNAs) is closely related to the physiological and pathological processes of diseases. Therefore, inferring the potential lncRNA–disease associations are helpful for understanding the molecular pathogenesis of diseases. Most previous methods have concentrated on the construction of shallow learning models in order to predict lncRNA-disease associations, while they have failed to deeply integrate heterogeneous multi-source data and to learn the low-dimensional feature representations from these data. We propose a method based on the convolutional neural network with the attention mechanism and convolutional autoencoder for predicting candidate disease-related lncRNAs, and refer to it as CNNDLP. CNNDLP integrates multiple kinds of data from heterogeneous sources, including the associations, interactions, and similarities related to the lncRNAs, diseases, and miRNAs. Two different embedding layers are established by combining the diverse biological premises about the cases that the lncRNAs are likely to associate with the diseases. We construct a novel prediction model based on the convolutional neural network with attention mechanism and convolutional autoencoder to learn the attention and the low-dimensional network representations of the lncRNA–disease pairs from the embedding layers. The different adjacent edges among the lncRNA, miRNA, and disease nodes have different contributions for association prediction. Hence, an attention mechanism at the adjacent edge level is established, and the left side of the model learns the attention representation of a pair of lncRNA and disease. A new type of lncRNA similarity and a new type of disease similarity are calculated by incorporating the topological structures of multiple bipartite networks. The low-dimensional network representation of the lncRNA-disease pairs is further learned by the autoencoder based convolutional neutral network on the right side of the model. The cross-validation experimental results confirm that CNNDLP has superior prediction performance compared to the state-of-the-art methods. Case studies on stomach cancer, breast cancer, and prostate cancer further show the ability of CNNDLP for discovering the potential disease lncRNAs.

1. Introduction

For the past few years, genetic information has been thought to be stored only in protein-coding genes, while non-coding RNAs (ncRNAs) are only byproducts of the transcription process [1,2]. However, accumulating evidences indicate that ncRNAs play important roles in various biological processes, especially long non-coding RNAs (lncRNAs), with lengths > 200 nucleotides [3,4].
The previous methods have been presented for predicting the lncRNA-disease associations, and they are classified into three categories. The methods in the first category utilize machine learning methods to identify the candidate associations. Chen et al. develop a semi-supervised learning model, LRLSLDA, which uses Laplacian regularized least squares to identify possible associations between lncRNA and disease [5]. A model based on the Bayesian classifier was developed for predicting candidate disease lncRNAs [6]. However, most of the methods in this category fail to achieve the good performances for the lncRNAs with no any known associated diseases.
The second category of methods takes use of the biological premise that lncRNAs with similar functions tend to be associated with similar diseases [7]. First, the similarity between two lncRNAs is calculated by the diseases associated with the lncRNAs, and a network composed of lncRNA is constructed by using the similarities between lncRNAs [8]. Several methods are presented for predicting the lncRNAs related to a given disease based on the lncRNA network, for instance, via random walks on the lncRNA network [9,10] or by utilizing the information of neighboring nodes of lncRNA [11]. These methods are ineffective for the new diseases with no known related lncRNAs, as they rely on a set of seed lncRNAs that have been observed to be related to the disease. Some methods attempt to introduce additional information about diseases to solve this shortcoming. Disease information is incorporated with the lncRNA network to create a heterogeneous lncRNA-disease network that contains information of lncRNA similarities, that of disease similarities and that of lncRNA-disease associations. Several methods exploit the information, but they construct different models within the heterogeneous network to estimate the association scores between the lncRNAs and the diseases. For instance, the association scores are derived by random walks in the lncRNA-disease network [10,12], or by matrix factorization of lncRNA-disease associations [13,14]. Since lncRNAs are often involved in disease processes along with miRNAs, it is necessary to integrate the interactions and associations about the miRNAs. Nevertheless, most of the previous methods overlook these information related to the miRNAs.
The third category of methods integrates multiple biological data sources about the lncRNA, the miRNA, the proteins. MFLDA integrates various information about the genes and the miRNAs interacted with lncRNAs, and about the diseases associated with lncRNAs. The method constructs a matrix factorization based on data fusion model for predicting disease lncRNAs [15]. Zhang et al. introduce the protein information to establish the lncRNA-protein-disease network and predict the candidate associations between lncRNAs and diseases based on propagating information streams in the network [16]. The diverse information available about the lncRNAs, diseases, genes, and proteins reflect the associations of lncRNAs and diseases from the different perspectives. However, it is difficult for these methods to deeply integrate heterogeneous data from multiple sources. Therefore, we present a novel prediction method based on dual convolutional neural networks to learn the latent representations of lncRNA-disease pairs from the multiple-source data.

2. Experimental Evaluations and Discussions

2.1. Evaluation Metrics

Five-fold cross-validation is used to evaluate the prediction performances of CNNDLP and several state-of-the-art methods for predicting lncRNA-disease associations. All the known lncRNA–disease associations are regarded as positive samples, and the unobserved associations are taken as negatives samples. We randomly divided all the positive samples into five subsets, and four of them are used to training the model. As the number of positive samples is far less than that of the negative samples (ratio of positive samples to negative samples is nearly 1:36 in our study), during the training process, we select the negative samples randomly whose number match to the number of the positive training samples, and these negative samples are also used for training the model. The positive samples in the remaining subset and all the negative samples are considered as the testing samples. The number of positive samples and that of negative samples during the cross-validation process are listed in Supplement Table S1. In particular, during each cross-validation, the positive samples used for testing are removed and the lncRNA similarities are recalculated by using the remaining positive samples.
We obtain the association scores of testing samples and prioritize them by their scores. The positive samples are ranked higher, which indicate that the prediction performance is better. The lncRNA-disease node pairs whose scores are greater than a classification threshold θ are identified as positive samples, and the ones that have lower scores are determined as negative samples. The true positive rates (TPRs) and the false positive rates (FPRs) at various θ values are calculated as follows:
T P R = T P T P + F N ,   F P R = F P F P + T N
where T P and T N are the numbers of positive and negative samples that are identified correctly, while F N and F P are the numbers of misidentified positive and negative samples. The receiver operating characteristic (ROC) curve can be drawn according to the TPRs and FPRs at each various θ , while the area under the ROC curve (AUC) is usually used to evaluate the overall performance of a prediction method [17].
A serious imbalance between the positive samples and the negative ones appears since their ratio is 1:36. For such imbalanced cases, precision-recall (PR) curve is confirmed to be more informative than ROC curve [18]. Therefore, the PR curve is used as another important measurement for the prediction performance of each method. Precision and recall are calculated as follows:
p r e c i s i o n = T P T P + F P ,   r e c a l l = T P T P + F N
where precision is the proportion of the real positive samples among the samples that are identified as the positive ones, while recall is the proportion of the real positive samples to the total actual positive ones. The area under the P-R curve (AUPR) is also utilized to evaluate the performance of lncRNA-disease association prediction [19].
In addition, the top candidate lncRNAs are usually selected by the biologists for further experimental verification of their associations with an interested disease. Therefore, we demonstrate the recall rates of the top 30, 60, and 240 candidates, which demonstrates how many of the positive samples are identified correctly within the ranking list of top k.

2.2. Comparison with Other Methods

To assess the prediction performance of CNNDLP, we compare it with several state-of-the-art methods for predicting disease lncRNAs: SIMCLDA [20], Ping’s method [21], MFLDA [15], LDAP [22] and CNNLDA [8]. CNNDLP and the other four methods have specific hyperparameters for fine-tuning to achieve their best association prediction performance. We choose the values of CNNDLP’s hyperparameters, α, β and λ, from {0.1, ..., 0.9}. CNNDLP achieved the best performance of five-fold cross-validation, when α = 0.9, β = 0.8 and λ = 0.3. The prediction performances of CNNDLP at different values of α, β, and λ on CNNDLP in the Supplementary Table S2, Supplementary Table S3, and Supplementary Table S4. In addition, the window size of all convolutional layers and pooling layers in CNNDLP is set as 2 × 2. The number of filters in the first and the second convolutional layers n c o n v 1 and n c o n v 2 are set to 16 and 32 respectively. CNNDLP has a great many parameters, which is easy to make the model overfit all the training samples. Therefore, we adopt dropout strategy and batch normalization to prevent the overfitting. To make a fair comparison, we set the hyperparameters of other methods to the optimal values that are recommended by their respective literatures (i.e., αl = 0.8, αd = 0.6 and λ = 1 for SIMCLDA, α = 0.6 for Ping’s method, α = 105 for MFLDA, Gap open = 10. Gap extend = 0.5 for LDAP).
As shown in Figure 1a, CNNDLP yields the highest average performance on all of the 405 diseases (AUC = 0.969). In particular, its performance is increased SIMCLDA by 21.2%, Ping’s method by 9.3%, MFDDA by 34.4%, LDAP by 10.7%, and CNNLDA by 1.7%. The AUCs of the five methods on 10 well-characterized diseases are also listed in Table 1, and CNNDLP achieves the best performance in all of the 10 diseases. The AUC of CNNDLP is slightly better than CNNLDA, but the AUPR of the former is 3.5% higher than the latter. The possible reason for this is that CNNLDA did not learn the low-dimensional network representation of a lncRNA-disease pair. Ping’s method and the LDAP achieved decent performance as they take advantage of the various similarities of different types of lncRNAs and diseases. MFLDA does not perform as well as the other four methods. A possible reason is that it ignored the lncRNA similarity and the disease similarity, while are exploited by the other methods. The improvement of CNNDLP over the compared methods is primarily due to the fact that it deeply learns the attention representation and low-dimensional network-level representation of the lncRNA-disease node pairs.
As shown in Figure 1b, CNNDLP’s average AUPR is also higher than other methods on 405 diseases (AUPR = 0.286). Its average AUPR is 22.7%, 13.4%, 24.7%, 15.9%, and 3.5% higher SIMCLDA, Ping’s method, MFLDA, LDAP and CNNLDA, respectively. In addition, CNNDLP performs the best performance among nine of the ten well-characterized diseases (Table 2).
A higher recall value in the top k of ranking list indicates that more real lncRNA-disease associations are identified correctly. Figure 2 shows CNNDLP outperforms the other methods at different top k cutoffs, and ranks 88.6% in top 30, ranks 94.6% in top 60, ranks 97.5% in top 90, and ranks 98.3% in top 120. Most of the recall rates of Ping’s method are very close to LDAP. The former ranked 68.9%, 81.3%, 87.5% and 92.7% in top 30, 60, 90 and 120, respectively, and the latter ranked 68.5%, 81.7%, 88.0% and 93.3%. MFLDA is still worse than the other methods, and it ranked 42.0%, 53.9%, 61.0% and 65.6%.
In addition, a paired Wilcoxon test is conducted to confirm whether CNNDLP’s prediction performance is significantly greater than the other methods. The statistical results in Table 3 show that CNNDLP yields better performance than the other methods in terms of not only AUCs but AUPRs, as well for the threshold p-value of 0.05.

2.3. Case Studies: Stomach Cancer, Breast Cancer and Prostate Cancer

To further demonstrate the capability of CNNDLP to discover potential disease-related candidate lncRNAs, we construct the case studies on stomach cancer, breast cancer, and prostate cancer. For each of these three diseases, we prioritize the candidate lncRNA-disease associations based on their association scores and gather their respective 15 candidates.
Stomach cancer is currently the fourth most common malignant tumor in the world and the second leading cause of cancer-related death [23]. First, Lnc2Cancer is a manually curated database that are verified associations between the lncRNAs and the human cancers by the biological experiments [24]. Twelve of 15 candidates are included by Lnc2Cancer (Table 4), which indicates that these lncRNAs are indeed associated with the disease.
Second, LncRNADisease records more than 4564 lncRNA-disease associations that are obtained from experiments, the published literatures or computation, and then the dysregulation of lncRNAs are manually confirmed [25]. There are 14 candidates contained by the LncRNADisease, indicating they are upregulated or downregulated in stomach cancer tissues. In addition, one candidate labeled by “literature” is supported by the literature, and it is confirmed to have dysregulation in the cancer when compared with the normal tissues [26].
Among the top 15 candidates for breast cancer, 11 candidates are reported in Lnc2Cancer with abnormal expression in breast cancer. (Table 5) LncRNADisease contains 12 candidates, which confirms the associations between these candidates and the disease. The remaining 2 candidates are confirmed by the literatures to have desregulation in the breast cancer [27,28].
The top 15 prostate cancer-related candidates and the corresponding evidences are listed in Table 6. Fourteen candidates are included by Lnc2Cancer and 14 ones are contained by LncRNADisease, which indicates that they truly are related to the disease. All the case studies confirm that CNNDLP is effective and impactful for discovering potential candidate disease lncRNAs.

2.4. Prediction of Novel Disease lncRNAs

After five-fold cross validation and case studies to confirm its prediction performance, we further apply CNNDLP to 405 diseases. All the known lncRNA-disease associations are used for training CNNDLP’s to predict potential disease-related lncRNAs. The top 50 potential candidates for each of 405 diseases are demonstrated in Supplementary Table S5.

3. Materials and Methods

3.1. Datasets for lncRNA-Disease Association Prediction

We obtained thousands of lncRNA-disease associations, lncRNA-miRNA interactions and miRNA-disease associations from a published work [15]. The human lncRNA-disease database (LncRNADisease) consists of 2687 lncRNA-disease associations that were verified by the biological experiments, covering 240 lncRNAs and 405 diseases [29]. The disease similarities were calculated based on directed acyclic graphs (DAGs) and the DAGs were constructed based on the disease terms from the U.S. National Library of Medicine (MeSH). The 1002 lncRNA-miRNA interactions were originally extracted from starBasev2.0 and they have been confirmed by biological experiments [30], and were involved 495 miRNAs. The 13,559 miRNA-disease associations were obtained from HMDD database [31].

3.2. Bipartite Graphs about the lncRNAs, Diseases, miRNAs, and Representations

We firstly construct a bipartite graph composed of lncRNAs and diseases by connecting them according to the observed lncRNA-disease associations (Figure 3a). The graph is represented by matrix A = [ A i j ] R N l × N d , where N l and N d are the number of lncRNAs and that of diseases, respectively. Each of rows corresponds to a lncRNA while each of columns represent a disease. If a lncRNA l i has been observed to be associated a disease d j , the A i j in A is set to 1, otherwise A i j is 0.
There are a great many interactions between the lncRNAs and miRNAs that have been confirmed by the biological experiments [32]. A bipartite graph composed of lncRNA and miRNA nodes is established when there are known interactions between them (Figure 3b). B = [ B i j ] R N l × N m is used to represent interaction matrix, the graph including N l lncRNAs and N m miRNAs. If it is known that lncRNA l i is interacted with miRNA m j , B i j = 1 , or B i j = 0 when their interaction has not been observed.
An edge is added to connect a miRNA and a disease, when they are observed to have association (Figure 3c). C = [ C i j ] R N m × N d is a matrix representing a bipartite graph with N m miRNAs and N d diseases. We set C i j to 1 if miRNA m i is associated with disease d j , or 0 when no such association is observed.

3.3. LncRNA-Disease Association Prediction Model Based on CNN

In this section, we describe the prediction model based on convolutional neural networks and attention mechanism for learning the latent representation and predicting the association score of lncRNA l i and disease d j . The embedding layer is firstly constructed by incorporating the associations, the similarities, the interactions about lncRNAs, diseases, miRNAs. A novel prediction model is constructed and it is composed of the left and right parts. The left side of the model learns the attention representation of l i and d j , while the network representation of l i and d j is learned in the right side of model. Each of the two representations goes through a fully connected layer and a softmax layer and the associated possibility between l i and d j is obtained and it is regarded as their association score. The final score is the weighted sum of two association scores.

3.3.1. Embedding Layer on the Left

lncRNA Functional Similarity Measurement

On the basis of the biological premise that lncRNAs with similar functions are more possibly to be associated with similar diseases, the similarity of two lncRNAs is measured by their associated diseases. For instance, lncRNA l a is associated with disease d 1 , d 2 and d 4 and lncRNA l b is associated with diseases d 2 , d 4 and d 5 . The similarity between E a = { d 1 , d 2 , d 4 } and E b = { d 2 , d 4 , d 5 } is regarded as the functional similarity of l a and l b . The lncRNA similarity that are used by us is calculated according to the Xuan’s method [8]. Matrix L = [ L i j ] R N l × N l is the lncRNA similarity matrix (Figure 3d), where L i j is the similarity of lncRNAs l i and l j , L i j value changes between 0 and 1.

Disease Similarity Measurement

All semantic terms related to a disease form its directed acyclic graph (DAG). The semantic similarities between the diseases are successfully calculated by Wang et al. based on their DAGs [33]. We calculate the disease similarities according to Wang’s method, and the similarities can be represented by matrix D = [ D i j ] R N d × N d , where D i j is the similarity of disease d i and d j (Figure 3e). The similarity of two diseases also varies between 0 and 1.

The Left Embedding Layer for Integrating the Original Information

If a lncRNA and a disease have similarity relationships and association relationships with the more common lncRNAs, they are more likely to associated with each other. We take the lncRNA l 1 and the disease d 2 as an example to explain the process of constructing the embedding layer on the left. As shown in Figure 4, let L 1 represents the first row of L which records the similarities between l 1 and all the lncRNAs. The second row of A T , A 2 T , contains the associations between d 2 and all the lncRNAs. For example, as l 1 is similar to l 2 and l 5 , and d 2 has been associated with l 2 , l 4 and l 5 , l 1 is possibly related to d 2 . We stack L 1 and A 2 T together as the first part of the embedding layer. Similarly, l 1 and d 2 are more likely to associate when l 1 and d 2 have the similarity and association connections with more common diseases. Therefore, we stack A 1 and D 2 as the second part of the embedding layer. In addition, when a lncRNA and a disease have interaction and association relationships with the common miRNAs, they are more possibly to have association. For instance, there is a possible association between l 1 and d 2 , since l 1 interacts with miRNAs m 1 and m 3 , and disease d 2 is associated with m 2 and m 3 . The first row of B and the second row G T are stacked as the third part of the embedding layer. The final embedding layer matrix between l 1 and d 2 is denoted as X R 2 × ( N l + N m + N d ) .

Attention at the Adjacent Edge Level

For a lncRNA node or a disease node, not all the adjacent edges of the node have equal contributions for learning the representation of a pair of lncRNA-disease. In order to solve the issue, we establish the attention mechanism to enhance the adjacent edges that are important for predicting the lncRNA-disease associations. In the embedding layer matrix X , X i j represents the connection case between the i-th node and the j-th node, and X i j is assigned an attention weight α i j , which is defined as follows,
F i j = t a n   h ( W X i j + b )  
α i j = e x p ( F i j T u e ) j exp ( F i j T u e )  
X ^ i j = α i j X i j  
where W and b are a weight matrix and a context vector respectively, and u e is a bias vector. F i j is an attention score that represents the importance of X i j . α i j is a normalized importance X i j . X ^ is the enhanced embedding layer matrix after the attention mechanism at the adjacent edge level is applied for X .

3.3.2. Embedding Layer on the Right

First, it is known that two lncRNA nodes are similar if they are associated with some common disease nodes [22]. In the bipartite network of lncRNA-disease (Figure 5a), lncRNA l 1 and l 3 are associated with a common disease node d 2 , so l 1 and l 3 are similar. l 3 and l 5 are also similar because they are related to a disease node d 1 (Figure 5b). Similarly, d 1 is similar to d 2 as they are associated with common lncRNA node l 3 (Figure 5c). Second, if two lncRNA nodes have no common neighboring nodes, while they are related to some similar disease nodes, they are also similar to each other [22]. As shown in Figure 5d, l 1 and l 5 are similar, because their neighboring nodes d 1 and d 2 are similar. Similarly, d 2 and d 3 are similar as they are associated with similar neighboring nodes l 3 and l 5 (Figure 5e). Ping et al. successfully measured the lncRNA similarities and the disease similarities by utilizing the lncRNA-disease bipartite network.
Unlike Ping’s method that focused on a single bipartite network, multiple kinds of lncRNA similarities and disease similarities are calculated by utilizing the bipartite networks from different sources about lncRNA-disease associations, lncRNA-miRNA interactions and miRNA-disease associations. The first kind of lncRNA similarity L ( 1 ) = [ L i j ( 1 ) ] R N l × N l , and the first kind of disease similarity D ( 1 ) = [ D i j ( 1 ) ] R N d × N d are calculated according to Ping’s method. The second kind of lncRNA similarity is measured by exploiting the information of common miRNA nodes and similar ones interacting with two lncRNA nodes in the lncRNA-miRNA bipartite network, and it is denoted as L ( 2 ) = [ L i j ( 2 ) ] R N l × N l . Finally, the second kind of disease similarity D ( 2 ) = [ D i j ( 2 ) ] R N d × N d is calculated based on the miRNA-disease bipartite network.
In order to incorporate two kinds of lncRNA similarities L ( 1 ) and L ( 2 ) , the final lncRNA similarity L ( c ) is defined as follows,
L ( c ) = α L ( 1 ) + ( 1 α ) L ( 2 )
where α is the parameter utilized to control the contributions of L ( 1 ) and L ( 2 ) .
Similarly, the final disease similarity D ( c ) is the weighted sum of D ( 1 ) and D ( 2 ) , as follows,
D ( c ) = β D ( 1 ) + ( 1 β ) D ( 2 )  
where β is a parameter for regulating the weights of D ( 1 ) and D ( 2 ) .
The right embedding layer for integrating the second kinds of lncRNA and disease similarities. The establishment of the right embedding layer matrix Y R 2 × ( N l + N m + N d ) is similar to the left embedding layer matrix X . First, we stack the first row of L ( c ) , L 1 ( c ) , and A 2 T together as the first part of the embedding layer. Second, A 1 and D 2 ( c ) are stacked as the second part of the embedding layer. Finally, the first row of B and the second row G T are stacked as the third part of the embedding layer.

3.4. Convolutional Module on the Left

In this section, we describe a novel model based on convolutional neural networks with adjacent attention for learning latent representations of lncRNA-disease node pairs. The overall architecture is showed in Figure 6. We describe the left convolutional module in detail. Left module includes convolution and activation layer, max-pooling layer, fully connected layer. The embedding matrix X ^ R 2 × ( N l + N m + N d ) is inputted the convolutional module to learn an original representation of a pair of lncRNA-disease node.
For a convolutional layer, the length and the width of a filter are set to w and h respectively, which means the filter is applied on w × h features. In order to learn the marginal information of the embedding matrix X ^ , we pad zeros around X ^ . Let the number of filters is n c o n v . The convolution filters W c o n v R w × h × n c o n v are applied to the embedding matrix X ^ , and obtain the feature maps Z R n c o n v × ( 4 w + 1 ) × ( 2 + N l + N m + N d h + 1 ) . X ^ i j is the element at the i-th row and j-th column of X ^ . X ^ k , i , j represent a region in a filter when the kth filter slides the position X ^ i j .
X ^ k , i , j = X ^ ( i : i + w , j : j + h )
Z k ( i , j ) = g ( W k , i , j X ^ k , i , j + b ( k ) )
i [ 1 ,   4 w + 1 ] j [ 1 ,   2 + N l + N m + N d h + 1 ] ,   k [ 1 ,   n c o n v ]
Z k ( i , j ) is the element at the i-th row and the j-th column of the k-th feature map. g is relu function that it is a nonlinear activation function [34], W k is the weight matrix of the k-th filter and b is a bias vector.
The max-pooling layer is used to down-sample the features of the feature maps Z k ( i , j ) , and it outputs the most important feature in each feature map. Given an input Z k ( i , j ) , the output of pooling layer is shown as follows,
V k ( i , j ) = max ( Z k ( i : i + w p , j + h p ) )
i [ 1 ,   5 w w p + 1 ] , j [ 1 ,   3 + N l + N m + N d h h p + 1 ] , k [ 1 ,   n c o n v ]
where w p is the length of a filter of pooling layer and h p is the width. V k ( i , j ) is the element at the i-th row and the j-th column in the kth feature map. X ^ goes through two convolutional and two max-pooling layers, and we obtain the original representation Z l e f t of l 1 and d 2 from the left convolutional module.
Finally, Z l e f t is flattened to a vector z , which z is feed to fully connected layer. A softmax layer is used to normalized the output of the fully connected layer and we have
s c o r e l = s o f t m a x ( W z + b )
where W is the weight matrix, and b is a bias vector. s c o r e l is an associated probability distribution of C class (C = 2). s c o r e l is the probability that the lncRNA l 1 is associated with the disease d 2 and s c o r e l 0 is the probability that l 1 and d 2 have no association relationship.
Similarly, the embedding matrix Y R 2 × ( N l + N m + N d ) is feed to the convolutional module on the right side of the prediction model for learning the network representation Z r i g h t of l 1 and d 2 . s c o r e r of l 1 and d 2 are obtained when Z r i g h t is feed to the full connection layer and the softmax layer.

3.5. Convalutional Autoencoder Module on the Right

The matrices about lncRNA-disease association, lncRNA-miRNA interaction, and miRNA-disease association are very sparse, resulting in many 0 elements are contained in the embedding matrix Y . An autoencoder based convolutional neural network is constructed to learn important and low-dimensional feature representations of lncRNA-disease pair on the right side of CNNDLP. The encoding and decoding strategies are given as follows,

3.5.1. Encoding Strategy

The embedding layer matrix Y R 2 × ( N l + N d + N m ) is mapped into the low-dimensional feature space through encoding based on convolutional neural network. Y e n c o d e , k ( n 1 ) is inputted to the n-th convolution layer to obtain Z e n c o d e , k ( n ) . Y e n c o d e , k ( n ) is formed after Z e n c o d e , k ( n ) passes the n-th max-pooling layer. They are defined as follows,
Z e n c o d e , k ( n ) ( i , j ) = g ( W e n c o d e , k ( n ) Y e n c o d e , k ( n 1 ) ( i , j ) + b e n c o d e ( n ) ( k ) )
Y e n c o d e , k ( n ) ( i , j ) = m a x ( Z e n c o d e , k ( n ) ( i : i + w e , j + h e ) )
i [ 1 ,   4 w e + 1 ] ,   j [ 1 ,   2 + N l + N m + N d h e + 1 ] ,   k [ 1 ,   n e n c o d e ] ,   n [ 1 , H e ]
where H e is the total number of encoding layers, and Y e n c o d e ( 0 ) = Y . k represents the k-th filter and n e n c o d e is the number of filters during encoding process. Z e n c o d e , k ( n ) ( i , j ) and Y e n c o d e , k ( n ) ( i , j ) are the elements at the i-th row and the j-th column of the k-th feature map, respectively. W e n c o d e , k ( n ) is a weight matrix and b e n c o d e ( n ) ( k ) is a bias vector.

3.5.2. Decoding Strategy

The output of the H e -th encoding layer Y e n c o d e ( H e ) is used as the input of the decoder. It is a matrix that is similar to Y by decoding. The decoding process includes both the transpose convolution layer and transpose pooling layer, and they are respectively defined as,
Z d e c o d e , k ( n ) ( i , j ) = g ( W d e c o d e , k ( n ) Y d e c o d e , k , i , j ( n 1 ) + b d e c o d e ( n ) ( k ) )
Y d e c o d e , k ( n ) ( i , j ) = M a x u n p o o l ( W d e c o d e , k ( n ) Z d e c o d e , k ( n ) ( i , j ) + b d e c o d e ( n ) ( k ) )
  k [ 1 ,   n d e c o d e ] ,   n [ 1 , H d ]
Y d e c o d e , k ( n ) and Z d e c o d e , k ( n ) are the outputs of the n-th transpose convolution layer and transpose max-pooling layer, respectively. H d   is the total number of decoding layers, and n d e c o d e is the number of filters for decoding. As Y d e c o d e ( H d ) should be consistent with Y , we defined the loss function as follows,
l o s s a u t o = i   =   1 T ( Y i ( Y d e c o d e ) i ) 2 T
where Y d e c o d e is the output of decoding and Y is the input of encoding; Y i is corresponding to the i-th training sample (lncRNA-disease pair), and T is the number of training samples. The s c o r e r of l 1 and d 2 is obtained after the Y e n c o d e is feed to the full connection layer and the softmax layer.

3.6. Combined Strategy

In our model, the cross-entropy is used as the loss function, for the left and right parts of the prediction model loss functions are defined as follows,
l o s s 1 = i   =   1 T [ y l a b e l log ( s c o r e l ) + ( 1 y l a b e l ) l o g ( 1 s c o r e l ) ]
l o s s 2 = i   =   1 T [ y l a b e l log ( s c o r e r ) + ( 1 y l a b e l ) l o g ( 1 s c o r e r ) ]
where y l a b e l denotes the actual association label between a lncRNA and a disease. y l a b e l is 1 when the lncRNA is indeed associated with the disease, otherwise y l a b e l is 0. T is the number of training samples.
The final score of our model is a weighted sum of s c o r e l and s c o r e r as follows,
s c o r e = λ s c o r e l + ( 1 λ ) s c o r e r
where the parameter λ ( 0 , 1 ) is used to adjust the importance of s c o r e l and s c o r e r .

4. Conclusions

A novel method based on the convolutional neural network with adjacent edge attention and convolutional autoencoder, entitled CNNDLP, is developed for inferring potential candidate lncRNA-disease associations. Two embedding layers are constructed from the biological perspective for integrating heterogeneous data about lncRNAs, diseases, and miRNAs from multiple sources. We construct the attention mechanism at the adjacent edge level to discriminate the different contributions of edges and the latent representation of a lncRNA-disease pair is learned from the more informative edges by the left side of CNNDLP’s prediction model. On the basis of calculating the new type of lncRNA similarity and that of disease similarity, the right side of CNNDLP’s model captures the complex relationships among these similarities and the lncRNA-disease associations, as well as the topological structures of multiple heterogeneous networks. The novel prediction model based on the convolutional neural network learns the attention representation and the low-dimensional network one of the lncRNA-disease pair. The experimental results demonstrated that CNNDLP outperforms the other methods in terms of not only AUCs but AUPRs as well. In particular, CNNDLP is more beneficial for the biologists as the top part of its ranking list may retrieve more real lncRNA-disease associations. Case studies on three diseases further confirm that CNNDLP is able to discover the potential candidate disease-related lncRNAs. CNNDLP may serve as a powerful prioritization tool that screens prospective candidates for the subsequent discovery of actual lncRNA-disease associations through wet-lab experiments.

Supplementary Materials

Supplementary materials can be found at https://www.mdpi.com/1422-0067/20/17/4260/s1.

Author Contributions

Author Contributions: P.X. and N.S. conceived the prediction method, and they wrote the paper. N.S. and Y.L. developed the computer programs. T.Z. and Y.G. analyzed the results and revised the paper.

Funding

The work was supported by the Natural Science Foundation of China (61972135), the Natural Science Foundation of Heilongjiang Province (LH2019F049, LH2019A029), the China Postdoctoral Science Foundation (2019M650069), the Heilongjiang Postdoctoral Scientific Research Staring Foundation (BHL-Q18104), the Fundamental Research Foundation of Universities in Heilongjiang Province for Technology Innovation (KJCX201805), the Innovation Talents Project of Harbin Science and Technology Bureau (2017RAQXJ094), and the Fundamental Research Foundation of Universities in Heilongjiang Province for Youth Innovation Team (RCYJTD201805).

Acknowledgments

We would like to thank Editage (www.editage.com) for English language editing.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fu, X.-D. Non-coding RNA: A new frontier in regulatory biology. Natl. Sci. Rev. 2014, 1, 190–204. [Google Scholar] [CrossRef] [PubMed]
  2. Derrien, T.; Johnson, R.; Bussotti, G.; Tanzer, A.; Djebali, S.; Tilgner, H.; Guernec, G.; Martin, D.; Merkel, A.; Knowles, D.G.; et al. The GENCODE v7 catalog of human long noncoding RNAs: Analysis of their gene structure, evolution, and expression. Genome Res. 2012, 22, 1775–1789. [Google Scholar] [CrossRef] [Green Version]
  3. Guttman, M.; Rinn, J.L. Modular regulatory principles of large non-coding RNAs. Nature 2012, 482, 339–346. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Wapinski, O.; Chang, H.Y. Long noncoding RNAs and human disease. Trends Cell Biol. 2011, 21, 354–361. [Google Scholar] [CrossRef] [PubMed]
  5. Chen, X.; Yan, G.-Y. Novel human lncRNA-disease association inference based on lncRNA expression profiles. Bioinformatics 2013, 29, 2617–2624. [Google Scholar] [CrossRef] [PubMed]
  6. Zhao, T.; Xu, J.; Liu, L.; Bai, J.; Xu, C.; Xiao, Y.; Li, X.; Zhang, L. Identification of cancer-related lncRNAs through integrating genome, regulome and transcriptome features. Mol. Biosyst. 2015, 11, 126–136. [Google Scholar] [CrossRef] [PubMed]
  7. Lan, W.; Huang, L.; Lai, D.; Chen, Q. Identifying Interactions Between Long Noncoding RNAs and Diseases Based on Computational Methods. Methods Mol. Biol. 2018, 1754, 205–221. [Google Scholar]
  8. Xuan, P.; Cao, Y.; Zhang, T.; Kong, R.; Zhang, Z. Dual Convolutional Neural Networks with Attention Mechanisms Based Method for Predicting Disease-Related lncRNA Genes. Front. Genet. 2019, 10, 416. [Google Scholar] [CrossRef] [PubMed]
  9. Yu, G.; Fu, G.; Lu, C.; Ren, Y.; Wang, J. BRWLDA: Bi-random walks for predicting lncRNA-disease associations. Oncotarget 2017, 8, 60429–60446. [Google Scholar] [CrossRef]
  10. Chen, X.; You, Z.-H.; Yan, G.-Y.; Gong, D.-W. IRWRLDA: Improved random walk with restart for lncRNA-disease association prediction. Oncotarget 2016, 7, 57919–57931. [Google Scholar] [CrossRef]
  11. Xiao, X.; Zhu, W.; Liao, B.; Xu, J.; Gu, C.; Ji, B.; Yao, Y.; Peng, L.; Yang, J. BPLLDA: Predicting lncRNA-Disease Associations Based on Simple Paths with Limited Lengths in a Heterogeneous Network. Front. Genet. 2018, 9, 411. [Google Scholar] [CrossRef] [PubMed]
  12. Sun, J.; Shi, H.; Wang, Z.; Zhang, C.; Liu, L.; Wang, L.; He, W.; Hao, D.; Liu, S.; Zhou, M. Inferring novel lncRNA-disease associations based on a random walk model of a lncRNA functional similarity network. Mol. Biosyst. 2014, 10, 2074–2081. [Google Scholar] [CrossRef] [PubMed]
  13. Ding, L.; Wang, M.; Sun, D.; Li, A. TPGLDA: Novel prediction of associations between lncRNAs and diseases via lncRNA-disease-gene tripartite graph. Sci. Rep. 2018, 8, 1065. [Google Scholar] [CrossRef] [PubMed]
  14. Xuan, Z.; Li, J.; Yu, J.; Feng, X.; Zhao, B.; Wang, L. A Probabilistic Matrix Factorization Method for Identifying lncRNA-disease Associations. Genes 2019, 10, 126. [Google Scholar] [CrossRef] [PubMed]
  15. Fu, G.; Wang, J.; Domeniconi, C.; Yu, G. Matrix factorization-based data fusion for the prediction of lncRNA–disease associations. Bioinformatics 2017, 34, 1529–1537. [Google Scholar] [CrossRef] [PubMed]
  16. Zhang, J.; Zhang, Z.; Chen, Z.; Deng, L. Integrating Multiple Heterogeneous Networks for Novel LncRNA-Disease Association Inference. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019, 16, 396–406. [Google Scholar] [CrossRef] [PubMed]
  17. Xuan, P.; Ye, Y.; Zhang, T.; Zhao, L.; Sun, C. Convolutional Neural Network and Bidirectional Long Short-Term Memory-Based Method for Predicting Drug-Disease Associations. Cells 2019, 8, 705. [Google Scholar] [CrossRef] [PubMed]
  18. Saito, T.; Rehmsmeier, M. The Precision-Recall Plot Is More Informative than the ROC Plot When Evaluating Binary Classifiers on Imbalanced Datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef] [PubMed]
  19. Xuan, P.; Shen, T.; Wang, X.; Zhang, T.; Zhang, W. Inferring disease-associated microRNAs in heterogeneous networks with node attributes. IEEE/ACM Trans. Comput. Biol. Bioinform. 2018. [Google Scholar] [CrossRef] [PubMed]
  20. Lu, C.; Yang, M.; Luo, F.; Wu, F.-X.; Li, M.; Pan, Y.; Li, Y.; Wang, J. Prediction of lncRNA–disease associations based on inductive matrix completion. Bioinformatics 2018, 34, 3357–3364. [Google Scholar] [CrossRef]
  21. Ping, P.; Wang, L.; Kuang, L.; Ye, S.; Iqbal, M.F.B.; Pei, T. A Novel Method for LncRNA-Disease Association Prediction Based on an lncRNA-Disease Association Network. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019, 16, 688–693. [Google Scholar] [CrossRef] [PubMed]
  22. Lan, W.; Li, M.; Zhao, K.; Liu, J.; Wu, F.-X.; Pan, Y.; Wang, J. LDAP: A web server for lncRNA-disease association prediction. Bioinformatics 2016, 33, 458–460. [Google Scholar] [CrossRef] [PubMed]
  23. Ferlay, J.; Shin, H.-R.; Bray, F.; Forman, D.; Mathers, C.; Maxwell Parkin, D. Estimates of worldwide burden of cancer in 2008: GLOBOCAN. Int. J. Cancer 2010, 127, 2893–2917. [Google Scholar] [CrossRef] [PubMed]
  24. Ning, S.; Zhang, J.; Wang, P.; Zhi, H.; Wang, J.; Liu, Y.; Gao, Y.; Guo, M.; Yue, M.; Wang, L.; et al. Lnc2Cancer: A manually curated database of experimentally supported lncRNAs associated with various human cancers. Nucleic Acids Res. 2016, 44, D980–D985. [Google Scholar] [CrossRef] [PubMed]
  25. Bao, Z.; Yang, Z.; Huang, Z.; Zhou, Y.; Cui, Q.; Dong, D. LncRNADisease 2.0: An updated database of long non-coding RNA-associated diseases. Nucleic Acids Res. 2019, 47, D1034–D1037. [Google Scholar] [CrossRef] [PubMed]
  26. He, Y.; Luo, Y.; Liang, B.; Ye, L.; Lu, G.; He, W. Potential applications of MEG3 in cancer diagnosis and prognosis. Oncotarget 2017, 8, 73282–73295. [Google Scholar] [CrossRef] [Green Version]
  27. Xu, Y.; Chen, M.; Liu, C.; Zhang, X.; Li, W.; Cheng, H.; Zhu, J.; Zhang, M.; Chen, Z.; Zhang, B. Association Study Confirmed Three Breast Cancer-Specific Molecular Subtype-Associated Susceptibility Loci in Chinese Han Women. Oncologist 2017, 22, 890–894. [Google Scholar] [CrossRef] [Green Version]
  28. Lv, X.-B.; Jiao, Y.; Qing, Y.; Hu, H.; Cui, X.; Lin, T.; Song, E.; Yu, F. miR-124 suppresses multiple steps of breast cancer metastasis by targeting a cohort of pro-metastatic genes in vitro. Chin. J. Cancer 2011, 30, 821–830. [Google Scholar] [CrossRef] [Green Version]
  29. Chen, G.; Wang, Z.; Wang, D.; Qiu, C.; Liu, M.; Chen, X.; Zhang, Q.; Yan, G.; Cui, Q. LncRNADisease: A database for long-non-coding RNA-associated diseases. Nucleic Acids Res. 2013, 41, D983–D986. [Google Scholar] [CrossRef]
  30. Li, J.H.; Liu, S.; Zhou, H.; Qu, L.H.; Yang, J.H. starBase v2.0: Decoding miRNA-ceRNA, miRNA-ncRNA and protein-RNA interaction networks from large-scale CLIP-Seq data. Nucleic Acids Res. 2014, 42, D92–D97. [Google Scholar] [CrossRef]
  31. Li, Y.; Qiu, C.; Tu, J.; Geng, B.; Yang, J.; Jiang, T.; Cui, Q. HMDD v2.0: A database for experimentally supported human microRNA and disease associations. Nucleic Acids Res. 2014, 42, D1070–D1074. [Google Scholar] [CrossRef] [PubMed]
  32. Paraskevopoulou, M.D.; Hatzigeorgiou, A.G. Analyzing MiRNA-LncRNA Interactions. Methods Mol. Biol. 2016, 1402, 271–286. [Google Scholar] [PubMed]
  33. Wang, D.; Wang, J.; Lu, M.; Song, F.; Cui, Q. Inferring the human microRNA functional similarity and functional network based on microRNA-associated diseases. Bioinformatics 2010, 26, 1644–1650. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
Figure 1. ROC curves and PR curves of CNNDLP and other methods for all diseases.
Figure 1. ROC curves and PR curves of CNNDLP and other methods for all diseases.
Ijms 20 04260 g001
Figure 2. Recall values of top k candidates of CNNDLP and other four methods.
Figure 2. Recall values of top k candidates of CNNDLP and other four methods.
Ijms 20 04260 g002
Figure 3. Construction and representation of multiple bipartite graphs. (a) Construct a lncRNA-disease association bipartite graph based on the known associations between lncRNAs and diseases, and its’ matrix representation A. (b) Construct lncRNA-miRNA interactions bipartite graph based on the known lncRNA-miRNA interactions, and its’ matrix representation B. (c) Construct miRNA-disease association bipartite graph based on known miRNA-disease associations, and its’ matrix representation C. (d) Calculate the lncRNA similarity, and construct the matrix representation L. (e) Calculate the disease similarity, and construct the matrix representation D.
Figure 3. Construction and representation of multiple bipartite graphs. (a) Construct a lncRNA-disease association bipartite graph based on the known associations between lncRNAs and diseases, and its’ matrix representation A. (b) Construct lncRNA-miRNA interactions bipartite graph based on the known lncRNA-miRNA interactions, and its’ matrix representation B. (c) Construct miRNA-disease association bipartite graph based on known miRNA-disease associations, and its’ matrix representation C. (d) Calculate the lncRNA similarity, and construct the matrix representation L. (e) Calculate the disease similarity, and construct the matrix representation D.
Ijms 20 04260 g003
Figure 4. Construction of the left embedding layer matrix of l 1 and d 2 , X . (a) Construct the first part of X by exploiting the lncRNA similarities and the lncRNA-disease associations. (b) Construct the second part of X by integrating the lncRNA-disease associations and the disease similarities. (c) Construct the third part of X by incorporating the lncRNA-miRNA interactions and the miRNA-disease associations. (d) Concatenate the three parts of X .
Figure 4. Construction of the left embedding layer matrix of l 1 and d 2 , X . (a) Construct the first part of X by exploiting the lncRNA similarities and the lncRNA-disease associations. (b) Construct the second part of X by integrating the lncRNA-disease associations and the disease similarities. (c) Construct the third part of X by incorporating the lncRNA-miRNA interactions and the miRNA-disease associations. (d) Concatenate the three parts of X .
Ijms 20 04260 g004
Figure 5. Calculation of the first type of lncRNA similarity and the first type of disease similarity. (a) The lncRNA-disease association bipartite network. (b) Calculate the lncRNA similarities based on the common associated disease nodes. (c) Computer the disease similarities based on their common related lncRNA nodes. (d) Calculate the lncRNA similarities according to their associated similar disease nodes. (e) The disease similarity calculation based on their related similar lncRNA nodes.
Figure 5. Calculation of the first type of lncRNA similarity and the first type of disease similarity. (a) The lncRNA-disease association bipartite network. (b) Calculate the lncRNA similarities based on the common associated disease nodes. (c) Computer the disease similarities based on their common related lncRNA nodes. (d) Calculate the lncRNA similarities according to their associated similar disease nodes. (e) The disease similarity calculation based on their related similar lncRNA nodes.
Ijms 20 04260 g005
Figure 6. Construction of the prediction model based on the convolutional neural network and convolutional autoencoder for learning the attention representation and the low-dimensional network representation.
Figure 6. Construction of the prediction model based on the convolutional neural network and convolutional autoencoder for learning the attention representation and the low-dimensional network representation.
Ijms 20 04260 g006
Table 1. AUCs of CNNDLP and other methods on all the diseases and 10 well-characterized diseases.
Table 1. AUCs of CNNDLP and other methods on all the diseases and 10 well-characterized diseases.
Disease NameCNNDLPPing’s MethodAUC
LDAP
SIMCLDAMFLDACNNLDA
Prostate cancer0.9510.8260.7100.8740.5530.897
Stomach cancer0.9470.9300.9280.8640.4670.958
Lung cancer0.9760.9110.8820.7900.6760.940
Breast cancer0.9560.8720.8300.7420.5170.836
Reproduce organ cancer0.9430.8180.7420.7070.7400.922
Ovarian cancer0.9700.9130.8570.7860.5580.942
Hematologic cancer0.9890.9080.9030.8280.7160.934
Kidney cancer0.9840.9790.9770.7280.6770.956
Liver cancer0.9560.9100.8980.7990.6340.918
Thoracic cancer0.9210.8600.7920.7920.6490.890
Average AUC of 405 diseases0.9690.8700.7450.7450.6260.952
The bold values indicate the higher AUCs.
Table 2. AUPRs of CNNDLP and other methods on all the diseases and 10 well-characterized diseases.
Table 2. AUPRs of CNNDLP and other methods on all the diseases and 10 well-characterized diseases.
Disease NameCNNDLPPing’s MethodAUPR
LDAP
SIMCLDAMFLDACNNLDA
Prostate cancer0.5380.3330.2970.1760.0920.390
Stomach cancer0.3730.3640.0940.1380.0080.286
Lung cancer0.6660.4370.3630.1310.1710.058
Breast cancer0.4850.4030.3960.0470.0310.964
Reproduce organ cancer0.4980.2810.2400.1300.1030.091
Ovarian cancer0.5520.4830.4270.0270.0230.526
Hematologic cancer0.6670.4030.3700.2160.1210.523
Kidney cancer0.5690.6630.4620.0300.0340.584
Liver cancer0.6300.4980.5110.1400.1100.666
Thoracic cancer0.3990.3830.3640.1550.1020.890
Average AUC of 405 diseases0.2860.1520.1270.0590.0390.251
The bold values indicate the higher AUPRs.
Table 3. Comparing of different methods based on AUCs with the paired Wilcoxon test.
Table 3. Comparing of different methods based on AUCs with the paired Wilcoxon test.
SIMCLDAPing’s MethodMFLDALDAPCNNLDA
p-value of ROC curve9.2454 × 10−60.000485.9940 × 10−70.001210.00773
p-value of PR curve8.3473 × 10−70.041743.5037 × 10−80.001260.00024
Table 4. The top 15 stomach cancer-related candidate lncRNAs.
Table 4. The top 15 stomach cancer-related candidate lncRNAs.
Rank lncRNA NameDescription Rank lncRNA NameDescription
1SPRY4-IT1 Lnc2Cancer, LncRNADisease9CDKN2B-AS1LncRNADisease
2TINCR Lnc2Cancer, LncRNADisease10CCAT1Lnc2Cancer, LncRNADisease
3H19 Lnc2Cancer, LncRNADisease11HOTAIRLnc2Cancer, LncRNADisease
4TUSC7 Lnc2Cancer, LncRNADisease12GACAT2LncRNADisease
5BANCRLnc2Cancer, LncRNADisease13UCA1Lnc2Cancer, LncRNADisease
6MEG3Lnc2Cancer, LncRNADisease14PVT1Lnc2Cancer, LncRNADisease
7GAS5Lnc2Cancer, LncRNADisease15MEG8literature
8GHET1Lnc2Cancer, LncRNADisease
Table 5. The top 15 breast cancer-related candidate lncRNAs.
Table 5. The top 15 breast cancer-related candidate lncRNAs.
Rank lncRNA NameDescription Rank lncRNA NameDescription
1SOX2-OTLnc2Cancer, LncRNADisease9CCAT1Lnc2Cancer, LncRNADisease
2HOTAIR Lnc2Cancer, LncRNADisease10GAS5Lnc2Cancer, LncRNADisease
3LINC00472Lnc2Cancer, LncRNADisease11MIR124-2HGliterature
4BCYRN1LncRNADisease12XISTLnc2Cancer, LncRNADisease
5LINC-PINTliterature13LINC-RORLnc2Cancer, LncRNADisease
6MALAT1Lnc2Cancer, LncRNADisease14PANDARLnc2Cancer, LncRNADisease
7CDKN2B-AS1LncRNADisease15AFAP1-AS1Lnc2Cancer
8SPRY4-IT1Lnc2Cancer, LncRNADisease
Table 6. The top 15 prostate cancer-related candidate lncRNAs.
Table 6. The top 15 prostate cancer-related candidate lncRNAs.
Rank lncRNA NameDescription Rank lncRNA NameDescription
1CDKN2B-AS1LncRNADisease9HOTAIRLnc2Cancer, LncRNADisease
2PCGEM1Lnc2Cancer, LncRNADisease10LINC00963Lnc2Cancer, LncRNADisease
3PVT1Lnc2Cancer, LncRNADisease11H19Lnc2Cancer, LncRNADisease
4GAS5Lnc2Cancer, LncRNADisease12MEG3Lnc2Cancer, LncRNADisease
5HOTTIPLnc2Cancer, LncRNADisease13TUG1Lnc2Cancer, LncRNADisease
6NEAT1Lnc2Cancer, LncRNADisease14PCA3Lnc2Cancer, LncRNADisease
7PCAT5Lnc2Cancer15DANCRLnc2Cancer, LncRNADisease
8PRINSLnc2Cancer, LncRNADisease

Share and Cite

MDPI and ACS Style

Xuan, P.; Sheng, N.; Zhang, T.; Liu, Y.; Guo, Y. CNNDLP: A Method Based on Convolutional Autoencoder and Convolutional Neural Network with Adjacent Edge Attention for Predicting lncRNA–Disease Associations. Int. J. Mol. Sci. 2019, 20, 4260. https://doi.org/10.3390/ijms20174260

AMA Style

Xuan P, Sheng N, Zhang T, Liu Y, Guo Y. CNNDLP: A Method Based on Convolutional Autoencoder and Convolutional Neural Network with Adjacent Edge Attention for Predicting lncRNA–Disease Associations. International Journal of Molecular Sciences. 2019; 20(17):4260. https://doi.org/10.3390/ijms20174260

Chicago/Turabian Style

Xuan, Ping, Nan Sheng, Tiangang Zhang, Yong Liu, and Yahong Guo. 2019. "CNNDLP: A Method Based on Convolutional Autoencoder and Convolutional Neural Network with Adjacent Edge Attention for Predicting lncRNA–Disease Associations" International Journal of Molecular Sciences 20, no. 17: 4260. https://doi.org/10.3390/ijms20174260

APA Style

Xuan, P., Sheng, N., Zhang, T., Liu, Y., & Guo, Y. (2019). CNNDLP: A Method Based on Convolutional Autoencoder and Convolutional Neural Network with Adjacent Edge Attention for Predicting lncRNA–Disease Associations. International Journal of Molecular Sciences, 20(17), 4260. https://doi.org/10.3390/ijms20174260

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop