Skip to main content

METHODS article

Front. Cell Dev. Biol., 17 December 2021
Sec. Molecular and Cellular Pathology
Volume 9 - 2021 | https://doi.org/10.3389/fcell.2021.753027

GBDTLRL2D Predicts LncRNA–Disease Associations Using MetaGraph2Vec and K-Means Based on Heterogeneous Network

  • School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha, China

In recent years, the long noncoding RNA (lncRNA) has been shown to be involved in many disease processes. The prediction of the lncRNA–disease association is helpful to clarify the mechanism of disease occurrence and bring some new methods of disease prevention and treatment. The current methods for predicting the potential lncRNA–disease association seldom consider the heterogeneous networks with complex node paths, and these methods have the problem of unbalanced positive and negative samples. To solve this problem, a method based on the Gradient Boosting Decision Tree (GBDT) and logistic regression (LR) to predict the lncRNA–disease association (GBDTLRL2D) is proposed in this paper. MetaGraph2Vec is used for feature learning, and negative sample sets are selected by using K-means clustering. The innovation of the GBDTLRL2D is that the clustering algorithm is used to select a representative negative sample set, and the use of MetaGraph2Vec can better retain the semantic and structural features in heterogeneous networks. The average area under the receiver operating characteristic curve (AUC) values of GBDTLRL2D obtained on the three datasets are 0.98, 0.98, and 0.96 in 10-fold cross-validation.

1 Introduction

In the human genome, more than 98% of the genes are noncoding protein sequences. The remaining 2% can only be transcribed into noncoding RNAs (ncRNAs). The ncRNAs can be divided into microRNA (miRNA), long ncRNA (lncRNA), etc. NcRNAs between 200 and 100,000 in length are lncRNAs.

At first, lncRNAs are considered useless RNAs without any biological function (Mercer and Mattick, 2013). This is because they are expressed at a lower level than protein-coding RNAs. However, with the development of experimental methods and computing power, the change of lncRNA has been found to be associated with many diseases, such as colorectal cancer (Xiong et al., 2021), lung adenocarcinoma (Hou et al., 2021), and gastrointestinal cancer (Abdi et al., 2021). With the deepening of research on lncRNAs, there are many pieces of evidence that lncRNAs play a key role in many important biological processes, including transcription, translation, splicing, differentiation, epigenetic regulation, immune response, and cell cycle control. For example, lncRNA loc105377478 promotes NPs-Nd2O3 in 16HBE cells and thus induces inflammation in human bronchial epithelial cells (Yu et al., 2021). LncRNA HOTAIR is considered a potential biomarker (Maass et al., 2014). The expression level of HOTAIR in breast cancer tissues is 100 to approximately 2,000 times higher than normal tissues and is associated with the proliferation and survival of colorectal cancer (Ge et al., 2013). Some research has shown that lncRNA BCAR4 is expressed in 27% of primary breast tumors. LncRNA Braveheart has also been demonstrated to control heart development by interacting with the epigenetic modifier PRC2. And increasing the expression of lncRNA Linc-MD1 can promote muscle differentiation (Cesana et al., 2011). LncRNA NEAT1 can regulate the development of Parkinson's. Therefore, the research of the potential lncRNA–disease association can better comprehend the potential mechanism of human diseases and help diagnose and treat diseases. This research has important practical implications.

Biological experiments to identify potential associations are time-consuming, labor-intensive, and very expensive. Therefore, in order to effectively reduce the time consumed by biological experiments and economic costs, there has been much research based on bioinformatics and computational power. For example, the method KATZLGO is proposed by Zhang et al. (2017) to predict the interaction of lncRNA–lncRNA. The PLPIHS is proposed by Xiao et al. (2017) to predict lncRNA–protein interactions using HeteSim score. A computational framework for predicting lncRNA–protein interactions is proposed by Liu (2020). An approach to explore miRNA sponge networks in breast cancer is proposed by Tian and Wang (2021). A method to predict the subcellular localization of lncRNAs is proposed by Yang et al. (2020). The potential roles of oral squamous cell carcinoma (OSCC)-related mRNA and lncRNA are revealed by Li et al. (2021) through protein interaction network and co-expression network analysis. The model GBDTL2E is proposed by Wang et al. (2020) to predict the association between lncRNA and environmental factors. With the deepening of research, research on the prediction of lncRNA–disease association is mainly divided into the following categories:

1) Based on machine learning methods, the main idea of these methods is to prioritize candidate lncRNAs by training known and unknown lncRNA–disease correlation. The semi-supervised learning framework LRLSLDA is proposed by Chen and Yan (2013). A graph regularization non-negative matrix factorization (LDGRNMF) is proposed by Wang M.-N. et al. (2021). Based on the weight algorithm and the improved projection algorithm, LDAP-WMPS is proposed by Wang B. et al. (2021). A model proposed by Zhou et al. (2021) uses high-order proximity reserved embedding to embed nodes into the network. The model VGAELDA, which integrates variational reasoning and graph autoencoder, is proposed by Shi et al. (2021). A multi-label fusion collaborative matrix decomposition (MLFCMF) method is proposed by Gao et al. (2021) to predict lncRNA–disease associations. The model PSPA-LA-PCRA is proposed by Wang and Zhang (2021), which uses the data of pathological stages. The random distribution logical regression framework (RDLRF) is proposed by Sun et al. (2021), and the RDLRF combines simboost feature extraction with logistic regression (LR). The method FVTLDA is proposed by Xiao et al. (2020), which combines multiple linear regression and artificial neural network. The BLM-NPA is proposed by Cui et al. (2019) to predict based on the nearest neighbor. The alternate least squares method of matrix factor factorization (ALSBMF) is proposed by Zhu et al. (2020). A computational method based on graphical autoencoder matrix completion (GAMCLDA) is proposed by Wu et al. (2020). A deep matrix factorization method (DMFLDA) is proposed by Zeng et al. (2020). A graph-based method (PANDA) is proposed by Silva and Spinosa (2021). The PANDA takes the association prediction of lncRNAs and diseases as a link prediction problem.

2) Based on network methods, the main idea of these methods is using a similarity network to predict lncRNA–disease association. Based on the combination of incremental principal component analysis (IPCA) and random forest (RF), a lncRNA–disease association prediction method IPCARF is proposed by Zhu et al. (2021). The prediction method for lncRNA–disease associations (PCSLDA) based on Point Cut Set is proposed by Kuang et al. (2019). The model GAERF is proposed by Wu et al. (2021); GAREF uses graph autocoding (GAE) and RF to identify disease-related lncRNAs. A random walk-based multi-similarity fusion and bidirectional label propagation method RWSF-BLP is proposed by Xie et al. (2021). Based on the assumption that there is a potential association between an lncRNA and a disease, if they are associated with the same set of miRNAs, similar diseases tend to be closely related to IneRNAs with similar functions; the method LDLMD is proposed by (Wang et al., 2019). A method of internal confidence-based local radial basis biological network (ICLRBBN) is proposed by Wang Y. et al. (2021). A two-stage prediction model (DRW-BNSP) is proposed by Zhang et al. (2021). HAUBRW algorithm is proposed by Xie et al. (2020b), which combines thermal diffusion algorithm and probabilistic diffusion algorithm to redistribute resources. An lncRNA–disease association prediction model based on RF and feature selection, RFLDA, is proposed by Yao et al. (2020). A predictive lncRNA–disease prediction model based on heterogeneous networks is proposed by Song et al. (2020). The LDAMAN is proposed by Zhang et al. (2020), which uses a structural deep network embedding model. The method based on linear neighborhood similarity and unbalanced double random walk (LDA-LNSUBRW) is proposed by Xie et al. (2020a). The MHRWR is proposed by Zhao et al. (2020) to integrate the similarity network of lncRNAs, diseases, and genes, with the known lncRNA–disease association network, lncRNA–gene network, and disease–gene network. The method LDAH2V is proposed by Deng et al. (2021), which uses HIN2Vec to calculate the meta path and eigenvector of each lncRNA–disease pair in heterogeneous information networks.

It can be seen that the association prediction of lncRNAs and diseases has become a research hotspot. Currently, the existing methods simply regard all objects in the network as the same type. However, in heterogeneous networks, there are many types of nodes, and the relationship between nodes is very complex, which is not considered by traditional methods. At the same time, unknown correlation is far greater than known correlation, which brings great challenges to model training. To solve these problems, a method based on the Gradient Boosting Decision Tree (GBDT) and LR to predict the lncRNA–disease association (GBDTLRL2D) is proposed in this paper. The GBDTLRL2D uses MetaGraph2Vec for feature learning and the K-means clustering method to select negative sample sets. The contributions of our method are included:

• The GBDTLRL2D comprehensively considers the topological structure characteristics and meta-path characteristics of nodes in heterogeneous networks. MetaGraph2Vec is used to learn more information by capturing more semantic relationships between remote nodes.

• The GBDTLRL2D uses the K-means to get the clustering of the unknown correlation. The same number of negative samples as the positive samples is selected from the clusters.

• The GBDTLRL2D combines GBDT and LR. The GBDT + LR is a special classification algorithm. Its ability to find features and combine features is very powerful. The classification accuracy is high.

2 Materials and Methods

The known lncRNA and disease-associated data used in this paper are downloaded from the lncRNADisease (Chen et al., 2012), which includes three versions, namely, the version of June 2012, the version of January 2014, and the version of June 2015. Table 1 shows the data after deduplication.

TABLE 1
www.frontiersin.org

TABLE 1. LncRNA–disease association relationship dataset.

In this section, a method based on the GBDT and LR to predict the lncRNA–disease association (GBDTLRL2D) is proposed. The GBDTLRL2D uses MetaGraph2Vec for feature learning and the K-means clustering method to select negative sample sets. The main steps of GBDTLRL2D are as follows: 1) according to the downloaded data, the set of lncRNAs and diseases as well as the association matrix A of lncRNA–diseases is obtained after deduplication. 2) The disease semantic similarity matrix SSD and lncRNA functional similarity matrix FSL are calculated, and then the Gaussian interaction profile kernel similarity matrix of disease (GSD) and lncRNA (GSL) are calculated. 3) LncRNA similarity matrix SL is constructed according to GSL and FSL, and disease similarity matrix SD is constructed according to GSD and SSD. 4) The association matrix A of lncRNA–disease, lncRNA similarity matrix SL, and the disease similarity matrix SD are integrated to construct the global heterogeneous network G. On the G, the feature of each node is learned by MetaGraph2Vec to obtain the feature representation of each node. 5) K-means is used to select negative samples to obtain train sets. 6) The GBDT and LR classifier are used to predict the lncRNA–disease association. Figure 1 is a flowchart of the GBDTLRL2D. Each step of GBDTLRL2D is detailed in the next section.

FIGURE 1
www.frontiersin.org

FIGURE 1. Flowchart of the GBDTLRL2D. (A) Obtained the association matrix A. (B) Calculated the disease semantic similarity matrix SSD. (C) Calculated the long noncoding RNA (lncRNA) functional similarity matrix FSL. (D) Calculated the disease Gaussian interaction profile kernel similarity GSD. (E) Calculated the lncRNA Gaussian interaction profile kernel similarity GSL. (F) Obtained the lncRNA similarity SL. (G) Obtained the disease similarity SD. (H) Integrated three subnets A, SL, and SD to construct a global heterogeneous network. (I) Obtained the embedded features of nodes. (J) Selected the negative sample and obtained the training set. (K) Trained the Gradient Boosting Decision Tree combined with logistic regression classifier (GBDT + LR).

2.1 Calculate Disease Semantic Similarity

The computing method of disease semantic similarity SSD in this experiment is based on the Disease Ontology. The method presents the disease tissue as a directed acyclic graph (DAG). As shown in Figure 2, the relationships between diseases are described in a DAG. Each node is a disease, and the arrow points from a disease to its ancestor disease. It can be seen from Figure 3 that if one lncRNA is associated with a disease, then this lncRNA may be associated with sub-diseases of the disease. From this perspective, the correct identification of these new associations may help to understand the mechanisms underlying RNA levels and improve the speed of accurate diagnosis and treatment of diseases. Therefore, the SSD between diseases is calculated according to the DAG.

FIGURE 2
www.frontiersin.org

FIGURE 2. The disease DAG of breast neoplasms. DAG, directed acyclic graph.

FIGURE 3
www.frontiersin.org

FIGURE 3. Diseases associated with lncRNA H19. lncRNA, long noncoding RNA.

For disease di, the semantic value is obtained. The contribution of each ancestral disease in the DAG of disease du to the semantic value of di is firstly calculated as shown in formula (1):

Ci(u)=1, if u=imaxΔ×Ciuu children of u, if ui(1)

where Δ is the weight of the edge connecting disease du and its sub-diseases, namely, semantic contribution factor. According to the above formula, as the distance between disease di and other diseases increases, semantic contributions decrease. Therefore, Δ should be selected between 0 and 1. In this paper, Δ= 0.5. Then the semantic value of di is calculated as the sum of the contributions of di’s ancestor disease and di itself, as shown in formula (2):

C(i)=uZ(i)Ci(u)(2)

where Z(i) represents the node set in the DAG of disease di.

For the disease di and dj, when the DAG of disease di and dj has more overlapping nodes, their semantic similarity is higher. Therefore, the semantic similarity matrix SSD of diseases can be obtained as shown in formula (3):

SSD(i,j)=uZ(i)Z(j)Ci(u)+Cj(u)C(i)+C(j)(3)

SSD (i, j) is denoted as the semantic similarity value between disease di and dj.

2.2 Calculate Long Noncoding RNA Functional Similarity

According to the LNCSIM (Chen et al., 2015), the lncRNA functional similarity is described as follows. Diseases associated with the same lncRNA are grouped into a set. The DL1 is the disease set related to lncRNA lm, including x diseases. The DL2 is the disease set related to lncRNA ln, including y diseases. When the semantic similarity between diseases in DS1 and DS2 is higher, the functional similarity between lncRNA lm and ln may be higher, as shown in formula (4):

FSLlm,ln=dDL2maxSd,DL1+dDL1maxSd,DL2x+y(4)
maxSd,DL1=maxdDL1SSd,d1(5)

where maxS (d, DL1 (lm)) is the maximum semantic similarity of all diseases in the set DL1 related to lncRNA lm.

2.3 Calculate Gaussian Interaction Profile Kernel Similarity

In this section, the adjacency matrix A is constructed according to the known lncRNA–disease association. The A (li, dj) indicates whether lncRNA li and disease dj are related. The A as shown in formula (6):

Ali,dj=1li is associated with dj0 otherwise (6)

In lncRNA functional similarity matrix FSL and the disease semantic similarity matrix SSD, many elements are 0. In order to make the similarity network not sparse, Gaussian kernel similarity is calculated, as shown in formula (7):

GSLlm,ln=expδlAm,:An,:2(7)
GSDdi,dj=expδdA:,iA:,j2(8)

where GSL (lm, ln) is the Gaussian interaction profile kernel similarity score of lncRNA lm and ln, and GSD (di, dj) is the Gaussian interaction profile kernel similarity of diseases di and dj. The A (m,) is mth row of A, and A (. i) is ith col of A. Parameters δl and δd are obtained as shown in Eq. 9 and Eq. 10:

δl=δl/1pm=1pAm,:2(9)
δd=δd/1qi=1qA:,i2(10)

where p is the number of lncRNAs and q is the number of diseases.

2.4 Obtain Similarity Network

In this section, lncRNA similarity network and disease similarity network are constructed. The lncRNA similarity network is represented as SL. SL is fused by FSL and GSL. For lncRNA lm and ln, if FSL (lm, ln) = 0, then SL (lm, ln) = GSL (lm, ln); otherwise, SL (lm, ln) = FSL (lm, ln), as shown in formula (11):

SLlm,ln=GSLlm,ln if FSLlm,ln=0FSLlm,ln otherwise (11)
SDdi,dj=GSDdi,dj if SSDdi,dj=0SSDdi,dj otherwise (12)

Similarly, the disease similarity network is expressed as SD. The SD is fused by SSD and GSD, as shown in formula (12):

2.5 Obtain Node Features Through MetaGraph2Vec

In this section, the heterogeneous network is constructed by integrating the adjacency matrix A of lncRNA–disease association, lncRNA similarity network SL, and the disease similarity network SD. Because there are many types of nodes and complex relationships among nodes in the heterogeneous network, embedding networks is difficult to implement. In GBDTLRL2D algorithm, MetaGraph2Vec (Zhang et al., 2018) is used for feature learning, which preserves both structural and semantic features in heterogeneous networks. MetaGraph2Vec uses metagraph to guide the generation of a random walk and learn about the potential embedding of nodes in a heterogeneous network of multiple types. Metagraph can represent features of sparse heterogeneous networks based on meta-paths. The specific substeps are shown below.

2.5.1 Building Heterogeneous Networks

The heterogeneous network G = (V, E) is constructed, where V represents the set of nodes and E represents the set of edges. The G is denoted as U. The dimension of the matrix U is (nl + nd)∗(nl + nd), where nl is the number of lncRNAs and nd is the number of diseases, as shown in formula (13):

U=SLAATSD(13)

where AT is the transpose of A.

2.5.2 Random Walk Guided by metagraph

Given a metaGraph g = (N, M, ns, nt) with ns = nt, the metaGraph is defined as a DAG on G, where ns represents the source node, nt represents the target node, N is the node set, and M is the edge set. The g=(N,M,ns,nt) is recursive metaGraph of g. The g is constructed by any number of g’s connected tail to head. A node of type ns is selected to start a random walk guided by the metatype.

In step i, node vi−1 is selected as the start of a random walk guided by the metaGraph. The random walk obtains the edge types of node vi−1 in a heterogeneous network that meets the constraints in the metagraph with all neighboring nodes. One edge type is randomly selected. Then, an edge of the selected edge type is randomly selected to get the next node vi. The random walk terminates when there is no edge type that satisfies the constraint.

The transition probability of step i guided by metaGraph g is denoted as T (vivi−1; g), vi−1 is the current node, and vi is the next hop node. If the node vi−1 and its neighbors in the heterogeneous network G do not satisfy the edge type of the constraint of the g, T (vivi−1; g) = 0. Otherwise, T (vivi−1; g) is shown in formula (14):

Tvivi1;g=1NUMgvi1×1μvi1,μE,ϕvi=ϕ(μ)(14)

where μvi1,μE,ϕvi=ϕ(μ) is the number of neighbor nodes of the same type as vi−1. NUMg(vi1) is the number of edge types that satisfy the constraint in the g starting from vi−1, as shown in formula (15):

NUMgvi1=jϕvi1,ϕ(μ)MNlϕvi1×N[j],vi1,μE(15)

After several walks, a node sequence Sg = v1, v2, …, vL of length L is finally obtained.

2.5.3 Obtain Node Features Through MetaGraph2Vec

By learning the mapping function Ψ, the nodes of heterogeneous networks are embedded into a d-dimensional space to obtain the embedding feature. The network G has a large number of nodes with different semantics. The nodes with similar semantics in heterogeneous networks are guaranteed to have similar low-dimensional representations Ψ(v).

The node sequence Sg = v1, v2, …, vL of length L is obtained by a random walk guided by metaGraph g. The embedding function Ψ(⋅) is learned by maximizing the occurrence probability of nodes before and after vi in the window, and the window size is b. Ψ(⋅) is shown in formula (16):

minΨlogTvib,,vi+b\/viΨvi(16)

where Tvib,,vi+b\/viΨvi=j=ib,jii+bTvjΨvi.

Following MetaPath2Vec, the T (vj∣Ψ(vi)) is related to the type of vj, as shown in formula (17):

TvjΨvi=TvjΨvi,ψvjTψvjΨvi(17)

where the probability TvjΨvi,ψvj is shown in formula (18):

TvjΨvi,ψvj=expΦvjΨviμV,ψ(μ)=ψvjexpΦ(μ)Ψvi(18)

After that, stochastic gradient descent is used to learn the parameters. At each iteration, a node context pair (vi, vj) is sampled according to the distribution of P (vi, vj), and the P (vi, vj) is the occurrence frequency of each node context pair (vi, vj) within b window size. The gradients are updated to minimize the following objective:

Oij=logTvjΨvi(19)

To speed up training, negative sampling is used to approximate the objective function:

Oij=logρΦvjΨvi+u=1UlogρΦvNj,uΨvi(20)

where ρ(⋅) is the sigmoid function, vNj,u is the uth negative node sampled for node vj, and U is the number of negative samples, vNj,u sampled from nodes with type Ψ(vj). Formally, parameters Ψ and Φ are updated as follows:

Ψ=ΨλOijΨ;Φ=ΦλOijΦ(21)

where λ is the learning rate.

The embedding function Ψ embeds the nodes of a heterogeneous network into a low-dimensional space, embedding each node and obtaining a low-dimensional representation Φ(v). Finally, the d-dimensional matrix X is obtained.

2.6 Obtain Negative Samples Using K-Means Clustering

Since the number of negative samples is far greater than that of positive samples in the set, it is necessary to balance the train set. The proposed method uses a novel and advanced data balancing method K-means clustering. The K-means is a segmentation technique based on centroid. The centroid of a cluster is used to represent the cluster. The centroid of a cluster is defined as the mean value of points within the cluster. The K-means is relatively simple and easy to implement. The specific implementation steps are as follows:

1. The initial k cluster centers are randomly selected from the unknown sample.

2. According to the distance from the point to the center of each cluster, the point is assigned to the closest cluster center category.

3. All points are assigned, and k cluster centers are recalculated.

4. If the recalculated k cluster centers are not the same as the previous cluster centers, we go to 2; otherwise, go to 5.

5. The clustering center is not changed, and the clustering results are output.

In this method, the data feature input into the K-means clustering method is composed of the fusion of SL, SD, and A. For example, the embedding matrix for sample lncRNA l2 and disease d4 pairs is shown in Figure 4.

FIGURE 4
www.frontiersin.org

FIGURE 4. An example of K-means embedding matrix.

As shown in Figure 4, the embedding matrix of lncRNA l2 and disease d4 pairs includes the following parts: 1) the first part is the second row of lncRNA similarity matrix SL; 2) the second part is composed of the vector corresponding to the adjacency matrix A of d4; 3) the third part is composed of the vector corresponding to l2 of the adjacency matrix A; and 4) the fourth part is the second row of disease similarity matrix SD. Combined with the representations in the first part, second part, third part, and fourth part, the final lncRNA l2 and disease d4 samples are constructed to carry out the K-means embedding matrix.

2.7 Train the Gradient Boosting Decision Tree Combined With Logistic Regression Classifier

After the sample and features are obtained, the GBDT + LR classifier is trained. Parameters of the model are initialized. The training data are regressed through the GBDT model and generate a decision tree. The leaf nodes of the decision tree are combined to find the new feature. The feature is used as input to the LR classifier model. Thus, the training process of GBDT + LR classifier is completed.

GBDT + LR is a process of feature crossing, and the path of GBDT can be directly used as the input feature of LR, avoiding the process of manual combination of cross features. Its algorithm structure is shown in Figure 5. The two trees in the figure are regression tree models trained by GBDT. The left tree has three leaf nodes, and the right tree has two leaf nodes. The final feature is a five-dimensional vector. For input x, it is assumed that it falls on the first node of the left tree and encodes (1, 0, 0); if it falls on the second node of the right tree, it encodes (0, 1), so the overall code is (1, 0, 0, 0, 1). Such codes are input into LR for classification. The steps of GBDT + LR for the algorithm are as follows:

FIGURE 5
www.frontiersin.org

FIGURE 5. Algorithm structure of GBDT + LR. GBDT, Gradient Boosting Decision Tree; LR, logistic regression.

Step 1) The original training data are trained with GDBT to generate a decision tree, and grid search is used to find the best parameter combination.

A: The initialization parameter of GDBT is shown in formula (22):

Θ0(x)=12logi=1NUMyii=1NUM1yi(22)

There are NUM samples to be trained, and yi is the label for sample i. The loss function J (y, Θt(x)) is defined as shown in formula (23):

Jy,Θt(x)=log1+expyΘt(x)(23)

where y is the label and Θt(x) is the weak model in the tth round.

B: Cycle t, in turn, where t = 1, 2, …T.

1) The negative gradient of the loss function of sample ith in wheel tth is calculated, as follows:

rt,i=Jyi,Θt1xiΘt1xi=yi1+expyiΘxi,i=1,2,,NUM(24)

where i = 1, 2, 3, …., NUM.

2) Construct the tth decision tree, and then get the corresponding leaf node area as Rtn, where n = 1, 2, …, N. N is the number of leaf nodes of the tree.

3) For the samples in each leaf node, we calculated the ctn, which minimizes the loss function, as shown in formula (25):

ctn=arg mincxRtnlog1+expyiΘxi+c(25)

4) Update the tth weak model as shown in formula (26):

Θt(x)=Θt1(x)+αn=1NctnIxRtn(26)

where I (xRtn) means that if x falls on a leaf node corresponding to Rtn, then the corresponding term is 1, and α means the learning rate.

5) Determine whether t is greater than T. If t is less than T, t = t + 1 and jump to 1) for the next iteration. Otherwise, it means that all T weak learners have been constructed and jump to C to end the training.

C: The final strong learner model is shown in formula (27):

Θ(x)=Θ0(x)+αt=1Tn=1NctnIxRtn(27)

where α is the learning rate.

Step 2) After the training of GDBT, for each tree in the model, the calculated probability value of the leaf node is denoted as 1, and new training data are constructed. In this paper, One-Hot Encoding is used to process the results of GDBT and construct a new training dataset.

One-hot Encoding is also known as one-bit Efficient coding. Single-hot coding encodes N states by using an N-bit status register. Each of the N states has its own independent register bit, and only one is valid at any time such as the following.

General status Encoder: 000, 001, 010, 011, 100, 101

One-Hot Encoder: 000 001, 000 010, 000 100, 001 000, 010 000, 100 000

Step 3) The new features obtained and the label data of the original training data are input into the LR classifier for the training. The hypothesis function of LR is shown in formula (28). Given x and θ, the possibility that x belongs to a positive sample is shown in formula (29). θ is obtained by training to minimize the loss function in formula (30).

hθ(x)=gθTx,g(z)=11+ez(28)
Pr(y=1x;θ)=hθ(x)=11+eθTx(29)
L(θ)=1mi=1my(i)loghθx(i)+1y(i)log1hθx(i)(30)

3 Result and Discussion

3.1 Dataset

The data are downloaded from the lncRNA–disease-associated data from the lncRNADisease database, including the data of three versions, namely, the version of June 2012, the version of January 2014, and the version of June 2015, labeled as DS1, DS2, and DS3, respectively. The training samples are obtained by all positive samples and randomly selecting negative samples by K-means clustering.

3.2 Performance Measures

In this paper, the 10-fold cross-validation is selected to measure the performance of the proposed method. The parameters of GBDTLRL2D are shown in Table 2. The main steps of 10-fold cross-validation are as follows: the training set is randomly divided into 10 subsets of the same size, nine of which are used as training data, and the remaining one is used as validation data in each training. After ten times of the above process training, each of the ten subsets, in turn, is used as validation data to obtain ten performance results. The final performance evaluation is obtained by averaging the ten performance results. Various evaluation indexes are used in this experiment, including Recall (REC), F1-score, Accuracy (ACC), Matthews correlation coefficient (MCC), and area under the receiver operating characteristic (ROC) curve (AUC). Their definition is as follows:

 Recall =TPTP+FN(31)
 Accuracy =TP+TNTP+TN+FP+FN(32)
F1score =2TP2TP+FP+FN(33)
MCC=TPTNFPFN(TP+FP)(TP+FN)(TN+FP)(TN+FN)(34)

where TP represents the number of correct prediction of positive samples as positive samples, TN represents the number of correct prediction of negative samples as negative samples, FN represents the number of incorrect prediction of positive samples as negative samples, and FP represents the number of prediction of negative samples as positive samples. We plot the ROC based on the true positive rate (TPR) and false positive rate (FPR), and we calculate the AUC as an important index to measure the model.

TABLE 2
www.frontiersin.org

TABLE 2. The partial experimental parameters of GBDTLRL2D.

3.3 Performance Comparison With Existing Machine Learning Methods

In order to prove the advantages of GBDT combined with LR classifier, we carried out several experiments to compare with GBDTLRL2D, including using RF + LR as the classifier, using GBDT only as the classifier, and using LR only as a classifier. It can be seen that GBDTLRL2D obtains the best performance among these methods. The 10-fold cross-validation is selected to measure the performance of the proposed method. Table 3 shows the predictive performance of GBDT + LR compared with other methods. The ROC curves of 10-fold cross-validation of GBDTLRL2D, RF + LR, GBDT, and LR are shown in Figures 69, respectively. The AUCs of GBDTLRL2D, RF + LR, GDBBT, and LR in DS1 are 0.976, 0.880, 0.619, and 0.649, respectively. The AUCs of GBDTLRL2D, RF + LR, GDBBT, and LR in DS2 are 0.983, 0.898, 0.654, and 0.705, respectively. The AUCs of GBDTLRL2D, RF + LR, GDBBT, and LR in DS3 are 0.961, 0.889, 0.647, and 0.667, respectively. It can be seen that GBDTLRL2D obtains the best performance among these methods.

TABLE 3
www.frontiersin.org

TABLE 3. Comparison of prediction performance using other machine learning methods.

FIGURE 6
www.frontiersin.org

FIGURE 6. ROC curve of GBDTLRL2D on three datasets. ROC, receiver operating characteristic.

FIGURE 7
www.frontiersin.org

FIGURE 7. ROC curves on three datasets using RF + LR. ROC, receiver operating characteristic; RF, random forest; LR, logistic regression.

FIGURE 8
www.frontiersin.org

FIGURE 8. ROC curves on three datasets using GBDT. ROC, receiver operating characteristic.

FIGURE 9
www.frontiersin.org

FIGURE 9. ROC curves on three datasets using LR. ROC, receiver operating characteristic; LR, logistic regression.

3.4 Performance Comparison With Different Topological Features

In order to demonstrate the performance of the experimental features, different feature groups (not using MetaGraph2Vec for representation learning, but using MetaGraph2Vec for representation learning) and different negative samples (not using K-means for clustering, but using K-means for clustering) are used for performance comparison in this section. Table 4 and Table 5 show the performance comparison with different topological features. In Table 4, on the same dataset, the result shows that the features obtained through MetaGraph2Vec embedding learning are trained to achieve better performance. Similarly, in Table 5, the performance of negative samples obtained through K-means cluster screening is better than that of negative samples randomly selected for training.

TABLE 4
www.frontiersin.org

TABLE 4. Performance comparison of representation learning without MetaGraph2vec.

TABLE 5
www.frontiersin.org

TABLE 5. Performance comparison without K-means.

3.5 Performance Comparison With Existing Methods

To further illustrate the advantages of the proposed model, several existing methods based on embedding are compared with GBDTLRL2D, such as LDAH2V, VGAELDA (Shi et al., 2021), and GCNMDA (Long et al., 2020). The 10-fold cross-validation is selected to measure the performance.

LDAH2V: The LDAH2V uses the HIN2Vec to calculate the meta-path and feature vector for each lncRNA–disease pair in the heterogeneous information network (HIN), which consists of lncRNA similarity network, disease similarity network, miRNA similarity network, and the associations between them. Then, a Gradient Boosting Tree (GBT) classifier to predict lncRNA–disease associations is built with the feature vectors.

VGAELDA: The VGAELDA integrates graph embedding learning and the alternate training via variational inference. Variational graph autoencoders (VGAEs) infer representations from features of lncRNAs and diseases, while graph autoencoders propagate labels via known lncRNA–disease associations. These two kinds of autoencoders are trained alternately by adopting variational expectation–maximization algorithm.

GCNMDA: The graph convolution network is used for network embedding in GCNMDA. The GCNMDA exploited the Conditional Random Field (CRF), which can ensure that similar nodes have similar representations. At the same time, the attention mechanism is designed in CRF layer.

Figure 10 shows the comparison results. Among these methods, the proposed model GBDTLRL2D achieves the best performance. There are several reasons: 1) the features learned by MetaGraph2Vec can better preserve node information and semantic information in a heterogeneous information network. 2) K-means clustering is used to select more representative negative samples. 3) The GBDTLRL2D uses the combined machine learning method of GBDT + LR with good performance to make predictions.

FIGURE 10
www.frontiersin.org

FIGURE 10. ROC curve comparison with existing methods. ROC, receiver operating characteristic.

Despite that our method is obviously superior to previous methods in all aspects, there are some limitations to GBDTLRL2D. The number of lncRNA–disease associations confirmed by biological experimental methods is limited. In addition, it is important to select classifiers. Currently, GBDT + LR is the best classifier for our model. In the future, we will have to try to combine other classifiers to achieve more accurate predictions.

3.6 Case Study

In this section, to further show the performance of the proposed model GBDTLRL2D in predicting the lncRNA–disease association, a case study is conducted on lncRNA “PVT1.” A proven association between “PVT1” and many diseases has been found in biology. In this paper, the proposed model GBDTLRL2D is used to predict the association between “PVT1” and disease. After processing by our algorithm, the list of diseases associated with lncRNA “PVT1” and their predicted scores is obtained. Ranking the diseases according to the predicted score from large to small, we can find that all the diseases in the top 10 associated with lncRNA “PVT1” are confirmed to be associated with “PVT1” in the lncRNADisease database. The top 10 diseases associated with lncRNA “PVT1” and their predicted scores are shown in Table 6.

TABLE 6
www.frontiersin.org

TABLE 6. The top 10 predicted diseases related to “PVT1.”

Conclusions

LncRNAs have been found by biologists to be closely related to diseases. Predicting the lncRNA–disease associations is conducive to research on the pathogenesis of a disease. But traditional biological methods have a large amount of data and are expensive, labor-intensive, and time-consuming. In recent years, there has been much research on computational models of biological experiments. In this paper, a method for predicting lncRNA–disease association is proposed. The proposed method uses MetaGraph2Vec to learn the features of nodes in a heterogeneous network and then uses K-means to select representative negative samples to solve the problem of imbalance between positive and negative samples, and the GBDT combined with LR is used as a classifier to predict lncRNA–disease associations. At last, the average AUCs of GBDTLRL2D obtained on the three datasets are 0.98, 0.98, and 0.96 in 10-fold cross-validation. Compared with the SIMCLDA, IIRWR, NCPLDA, and other experiments, the GBDTLRL2D greatly improves accuracy and performance.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: http://www.cuilab.cn/lncrnadisease.

Author Contributions

TD, ZK, JW, and ZM conceived this work and designed the experiments. TD, JW, and ZK carried out the experiments. TD and ZM collected the data and analyzed the results. TD and ZK wrote, revised, and approved the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant Nos. 62072477, 61309027, 61702562, and 61702561; the Hunan Provincial Natural Science Foundation of China under Grant No. 2018JJ3888; the Scientific Research Fund of Hunan Provincial Education Department under Grant No. 18B197; the National Key R&D Program of China under Grant No. 2018YFB1700200; the Open Research Project of Key Laboratory of Intelligent Information Perception and Processing Technology (Hunan Province) under Grant No. 2017KF01; and the Hunan Key Laboratory of Intelligent Logistics Technology 2019TP1015.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Acknowledgments

We would like to thank the Experimental Center of School of Computer and Information Engineering, Central South University of Forestry and Technology, for providing computing resources.

References

Abdi, E., Latifi-Navid, S., Latifi-Navid, H., and Safaralizadeh, R. (2021). Lncrna Polymorphisms and Upper Gastrointestinal Cancer Risk. Pathol. - Res. Pract. 218, 153324. doi:10.1016/j.prp.2020.153324

PubMed Abstract | CrossRef Full Text | Google Scholar

Cesana, M., Cacchiarelli, D., Legnini, I., Santini, T., Sthandier, O., Chinappi, M., et al. (2011). A Long Noncoding Rna Controls Muscle Differentiation by Functioning as a Competing Endogenous Rna. Cell 147, 358–369. doi:10.1016/j.cell.2011.09.028

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, G., Wang, Z., Wang, D., Qiu, C., Liu, M., Chen, X., et al. (2012). Lncrnadisease: a Database for Long-Non-Coding Rna-Associated Diseases. Nucleic Acids Res. 41, D983–D986. doi:10.1093/nar/gks1099

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, X., Yan, C. C., Luo, C., Ji, W., Zhang, Y., and Dai, Q. (2015). Constructing Lncrna Functional Similarity Network Based on Lncrna-Disease Associations and Disease Semantic Similarity. Sci. Rep. 5, 11338–11412. doi:10.1038/srep11338

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, X., and Yan, G.-Y. (2013). Novel Human lncRNA-Disease Association Inference Based on lncRNA Expression Profiles. Bioinformatics 29, 2617–2624. doi:10.1093/bioinformatics/btt426

PubMed Abstract | CrossRef Full Text | Google Scholar

Cui, Z., Liu, J. X., Gao, Y. L., Zhu, R., and Yuan, S. S. (2019). Lncrna-disease Associations Prediction Using Bipartite Local Model with Nearest Profile-Based Association Inferring. IEEE J. Biomed. Health Inform. 24, 1519–1527. doi:10.1109/JBHI.2019.2937827

PubMed Abstract | CrossRef Full Text | Google Scholar

Deng, L., Li, W., and Zhang, J. (2021). Ldah2v: Exploring Meta-Paths across Multiple Networks for Lncrna-Disease Association Prediction. Ieee/acm Trans. Comput. Biol. Bioinf. 18, 1572–1581. doi:10.1109/TCBB.2019.2946257

CrossRef Full Text | Google Scholar

Gao, M.-M., Cui, Z., Gao, Y.-L., Wang, J., and Liu, J.-X. (2021). Multi-label Fusion Collaborative Matrix Factorization for Predicting Lncrna-Disease Associations. IEEE J. Biomed. Health Inform. 25, 881–890. doi:10.1109/JBHI.2020.2988720

CrossRef Full Text | Google Scholar

Ge, X., Chen, Y., Liao, X., Liu, D., Li, F., Ruan, H., et al. (2013). Overexpression of Long Noncoding Rna Pcat-1 Is a Novel Biomarker of Poor Prognosis in Patients with Colorectal Cancer. Med. Oncol. 30, 588. doi:10.1007/s12032-013-0588-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Hou, M., Wu, N., and Yao, L. (2021). LncRNA CBR3-AS1 Potentiates Wnt/β-Catenin Signaling to Regulate Lung Adenocarcinoma Cells Proliferation, Migration and Invasion. Cancer Cel Int 21, 36–12. doi:10.1186/s12935-020-01685-y

CrossRef Full Text | Google Scholar

Kuang, L., Zhao, H., Wang, L., Xuan, Z., and Pei, T. (2019). A Novel Approach Based on point Cut Set to Predict Associations of Diseases and Lncrnas. Cbio 14, 333–343. doi:10.2174/1574893613666181026122045

CrossRef Full Text | Google Scholar

Li, J., Zhang, C., Shi, Y., Li, Q., Li, N., and Mi, Y. (2021). Identification of Key Lncrnas and Mrnas Associated with Oral Squamous Cell Carcinoma Progression. Cbio 16, 207–215. doi:10.2174/1573411016999200729125745

CrossRef Full Text | Google Scholar

Liu, Z.-P. (2020). Predicting Lncrna-Protein Interactions by Machine Learning Methods: a Review. Curr. Bioinformatics 15, 831–840.

Google Scholar

Long, Y., Wu, M., Kwoh, C. K., Luo, J., and Li, X. (2020). Predicting Human Microbe-Drug Associations via Graph Convolutional Network with Conditional Random Field. Bioinformatics 36, 4918–4927. doi:10.1093/bioinformatics/btaa598

PubMed Abstract | CrossRef Full Text | Google Scholar

Maass, P. G., Luft, F. C., and Bähring, S. (2014). Long Non-coding Rna in Health and Disease. J. Mol. Med. 92, 337–346. doi:10.1007/s00109-014-1131-8

CrossRef Full Text | Google Scholar

Mercer, T. R., and Mattick, J. S. (2013). Structure and Function of Long Noncoding Rnas in Epigenetic Regulation. Nat. Struct. Mol. Biol. 20, 300–307. doi:10.1038/nsmb.2480

PubMed Abstract | CrossRef Full Text | Google Scholar

Shi, Z., Zhang, H., Jin, C., Quan, X., and Yin, Y. (2021). A Representation Learning Model Based on Variational Inference and Graph Autoencoder for Predicting Lncrna-Disease Associations. BMC bioinformatics 22, 1–20. doi:10.1186/s12859-021-04073-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Silva, A. B. O. V., and Spinosa, E. J. (2021). Graph Convolutional Auto-Encoders for Predicting Novel Lncrna-Disease Associations. Ieee/acm Trans. Comput. Biol. Bioinf. 1, 1. doi:10.1109/TCBB.2021.3070910

CrossRef Full Text | Google Scholar

Song, X.-Y., Liu, T., Qiu, Z.-Y., You, Z.-H., Sun, Y., Jin, L.-T., et al. (2020). Prediction of Lncrna-Disease Associations from Heterogeneous Information Network Based on Deepwalk Embedding Model. In International Conference on Intelligent Computing. Springer, 291–300. doi:10.1007/978-3-030-60796-8_25

CrossRef Full Text | Google Scholar

Sun, Y., Zhao, H., Zhou, G., Guan, T., Wang, Y., and Gao, J. (2021). Random Distributed Logistic Regression Framework for Predicting Potential lncRNA‒disease Association. J. Mol. Cel Biol. 13, 386–388. doi:10.1093/jmcb/mjab005

CrossRef Full Text | Google Scholar

Tian, L., and Wang, S.-L. (2021). Exploring Mirna Sponge Networks of Breast Cancer by Combining Mirna-Disease-Lncrna and Mirna-Target Networks. Cbio 16, 385–394. doi:10.2174/1574893615999200711171530

CrossRef Full Text | Google Scholar

Wang, B., Zhang, C., Du, X.-x., and Zhang, J.-f. (2021a). Lncrna-Disease Association Prediction Based on Weight Matrix and Projection Score. BMC Bioinformatics. doi:10.21203/rs.3.rs-428221/v1

CrossRef Full Text | Google Scholar

Wang, B., and Zhang, J. (2021). Principal Component Regression Analysis for Lncrna-Disease Association Prediction Based on Pathological Stage Data. IEEE Access 9, 20629–20640. doi:10.1109/access.2021.3053839

CrossRef Full Text | Google Scholar

Wang, J., Kuang, Z., Ma, Z., and Han, G. (2020). Gbdtl2e: Predicting lncrna-ef associations using diffusion and hetesim features based on a heterogeneous network. Front. Genet. 11, 272. doi:10.3389/fgene.2020.00272

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, L., Xuan, Z., Zhou, S., Kuang, L., and Pei, T. (2019). A Novel Model for Predicting Lncrna-Disease Associations Based on the Lncrna-Mirna-Disease Interactive Network. Cbio 14, 269–278. doi:10.2174/1574893613666180703105258

CrossRef Full Text | Google Scholar

Wang, M.-N., You, Z.-H., Wang, L., Li, L.-P., and Zheng, K. (2021b). Ldgrnmf: Lncrna-Disease Associations Prediction Based on Graph Regularized Non-negative Matrix Factorization. Neurocomputing 424, 236–245. doi:10.1016/j.neucom.2020.02.062

CrossRef Full Text | Google Scholar

Wang, Y., Li, H., Kuang, L., Tan, Y., Li, X., Zhang, Z., et al. (2021c). Iclrbbn: a Tool for Accurate Prediction of Potential Lncrna Disease Associations. Mol. Ther. - Nucleic Acids 23, 501–511. doi:10.1016/j.omtn.2020.12.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, Q. W., Xia, J. F., Ni, J. C., and Zheng, C. H. (2021). GAERF: Predicting lncRNA-Disease Associations by Graph Auto-Encoder and Random forest. Brief Bioinform 22. doi:10.1093/bib/bbaa391.Bbaa391

CrossRef Full Text | Google Scholar

Wu, X., Lan, W., Chen, Q., Dong, Y., Liu, J., and Peng, W. (2020). Inferring Lncrna-Disease Associations Based on Graph Autoencoder Matrix Completion. Comput. Biol. Chem. 87, 107282. doi:10.1016/j.compbiolchem.2020.107282

PubMed Abstract | CrossRef Full Text | Google Scholar

Xiao, Y., Xiao, Z., Feng, X., Chen, Z., Kuang, L., and Wang, L. (2020). A Novel Computational Model for Predicting Potential Lncrna-Disease Associations Based on Both Direct and Indirect Features of Lncrna-Disease Pairs. BMC bioinformatics 21, 555–622. doi:10.1186/s12859-020-03906-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Xiao, Y., Zhang, J., and Deng, L. (2017). Prediction of Lncrna-Protein Interactions Using Hetesim Scores Based on Heterogeneous Networks. Sci. Rep. 7, 3664–3712. doi:10.1038/s41598-017-03986-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Xie, G., Huang, B., Sun, Y., Wu, C., and Han, Y. (2021). Rwsf-blp: a Novel Lncrna-Disease Association Prediction Model Using Random Walk-Based Multi-Similarity Fusion and Bidirectional Label Propagation. Mol. Genet. Genomics, 1–11. doi:10.1007/s00438-021-01764-3

CrossRef Full Text | Google Scholar

Xie, G., Jiang, J., and Sun, Y. (2020a). Lda-lnsubrw: Lncrna-Disease Association Prediction Based on Linear Neighborhood Similarity and Unbalanced Bi-random Walk, Ieee/acm Trans. Comput. Biol. Bioinf., 1, 1. doi:10.1109/TCBB.2020.3020595

CrossRef Full Text | Google Scholar

Xie, G., Wu, C., Gu, G., and Huang, B. (2020b). Haubrw: Hybrid Algorithm and Unbalanced Bi-random Walk for Predicting Lncrna-Disease Associations. Genomics 112, 4777–4787. doi:10.1016/j.ygeno.2020.08.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Xiong, M., Wu, M., Peng, D., Huang, W., Chen, Z., Ke, H., et al. (2021). Lncrna Dancr Represses Doxorubicin-Induced Apoptosis through Stabilizing Malat1 Expression in Colorectal Cancer Cells. Cel Death Dis. 12, 1–17. doi:10.1038/s41419-020-03318-8

CrossRef Full Text | Google Scholar

Yang, X.-F., Zhou, Y.-K., Zhang, L., Gao, Y., and Du, P.-F. (2020). Predicting Lncrna Subcellular Localization Using Unbalanced Pseudo-k Nucleotide Compositions. Cbio 15, 554–562. doi:10.2174/1574893614666190902151038

CrossRef Full Text | Google Scholar

Yao, D., Zhan, X., Zhan, X., Kwoh, C. K., Li, P., and Wang, J. (2020). A Random forest Based Computational Model for Predicting Novel Lncrna-Disease Associations. BMC bioinformatics 21, 126–218. doi:10.1186/s12859-020-3458-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, F., Zhang, X., Gao, L., Xue, H., Liu, L., Wang, S., et al. (2021). LncRNA Loc105377478 Promotes NPs-Nd2O3-Induced Inflammation in Human Bronchial Epithelial Cells through the ADIPOR1/NF-Κb axis. Ecotoxicology Environ. Saf. 208, 111609. doi:10.1016/j.ecoenv.2020.111609

CrossRef Full Text | Google Scholar

Zeng, M., Lu, C., Fei, Z., Wu, F., Li, Y., Wang, J., et al. (2020). DMFLDA: A Deep Learning Framework for Predicting IncRNA-Disease Associations, Ieee/acm Trans. Comput. Biol. Bioinf., 1. doi:10.1109/TCBB.2020.2983958

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, D., Yin, J., Zhu, X., and Zhang, C. (2018). “MetaGraph2vec: Complex Semantic Path Augmented Heterogeneous Network Embedding,” in Pacific-Asia conference on knowledge discovery and data mining (Springer), 196–208. doi:10.1007/978-3-319-93037-4_16

CrossRef Full Text | Google Scholar

Zhang, P., Zhao, B.-W., Wong, L., You, Z.-H., Guo, Z.-H., and Yi, H.-C. (2020). “A Novel Computational Method for Predicting Lncrna-Disease Associations from Heterogeneous Information Network with Sdne Embedding Model,” in International Conference on Intelligent Computing (Springer), 505–513. doi:10.1007/978-3-030-60802-6_44

CrossRef Full Text | Google Scholar

Zhang, Y., Chen, M., Xie, X., Shen, X., and Wang, Y. (2021). Two-stage Inference for Lncrna-Disease Associations Based on Diverse Heterogeneous Information Sources. IEEE Access 9, 16103–16113. doi:10.1109/ACCESS.2021.3053030

CrossRef Full Text | Google Scholar

Zhang, Z., Zhang, J., Fan, C., Tang, Y., and Deng, L. (2017). Katzlgo: Large-Scale Prediction of Lncrna Functions by Using the Katz Measure Based on Multiple Networks. Ieee/acm Trans. Comput. Biol. Bioinform 16, 407–416. doi:10.1109/TCBB.2017.2704587

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, X., Yang, Y., and Yin, M. (2020). Mhrwr: Prediction of Lncrna-Disease Associations Based on Multiple Heterogeneous Networks, Ieee/acm Trans. Comput. Biol. Bioinf., 1. doi:10.1109/TCBB.2020.2974732

CrossRef Full Text | Google Scholar

Zhou, J.-R., You, Z.-H., Cheng, L., and Ji, B.-Y. (2021). Prediction of Lncrna-Disease Associations via an Embedding Learning hope in Heterogeneous Information Networks. Mol. Ther. - Nucleic Acids 23, 277–285. doi:10.1016/j.omtn.2020.10.040

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu, R., Wang, Y., Liu, J. X., and Dai, L. Y. (2021). Ipcarf: Improving Lncrna-Disease Association Prediction Using Incremental Principal Component Analysis Feature Selection and a Random forest Classifier. BMC bioinformatics 22, 175–217. doi:10.1186/s12859-021-04104-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhu, W., Huang, K., Xiao, X., Liao, B., Yao, Y., and Wu, F.-X. (2020). Alsbmf: Predicting Lncrna-Disease Associations by Alternating Least Squares Based on Matrix Factorization. IEEE Access 8, 26190–26198. doi:10.1109/access.2020.2970069

CrossRef Full Text | Google Scholar

Keywords: long noncoding RNA, heterogeneous network, MetaGraph2Vec, K-means, Gradient Boosting Decision Tree, logistic regression

Citation: Duan T, Kuang Z, Wang J and Ma Z (2021) GBDTLRL2D Predicts LncRNA–Disease Associations Using MetaGraph2Vec and K-Means Based on Heterogeneous Network. Front. Cell Dev. Biol. 9:753027. doi: 10.3389/fcell.2021.753027

Received: 04 August 2021; Accepted: 22 November 2021;
Published: 17 December 2021.

Edited by:

Liang Cheng, Harbin Medical University, China

Reviewed by:

Yongjun Tang, Central South University, China
Bingbo Wang, Xidian University, China

Copyright © 2021 Duan, Kuang, Wang and Ma. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zhufang Kuang, zfkuangcn@163.com

Download