Next Article in Journal
Mitochondrial Aging and Senolytic Natural Products with Protective Potential
Next Article in Special Issue
Assessing the Differential Methylation Analysis Quality for Microarray and NGS Platforms
Previous Article in Journal
Effects of Cardiac Stem Cell on Postinfarction Arrhythmogenic Substrate
Previous Article in Special Issue
DLm6Am: A Deep-Learning-Based Tool for Identifying N6,2′-O-Dimethyladenosine Sites in RNA Sequences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

iADRGSE: A Graph-Embedding and Self-Attention Encoding for Identifying Adverse Drug Reaction in the Earlier Phase of Drug Development

Department of Computer, Jingdezhen Ceramic University, Jingdezhen 333403, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2022, 23(24), 16216; https://doi.org/10.3390/ijms232416216
Submission received: 22 November 2022 / Revised: 15 December 2022 / Accepted: 16 December 2022 / Published: 19 December 2022

Abstract

:
Adverse drug reactions (ADRs) are a major issue to be addressed by the pharmaceutical industry. Early and accurate detection of potential ADRs contributes to enhancing drug safety and reducing financial expenses. The majority of the approaches that have been employed to identify ADRs are limited to determining whether a drug exhibits an ADR, rather than identifying the exact type of ADR. By introducing the “multi-level feature-fusion deep-learning model”, a new predictor, called iADRGSE, has been developed, which can be used to identify adverse drug reactions at the early stage of drug discovery. iADRGSE integrates a self-attentive module and a graph-network module that can extract one-dimensional sub-structure sequence information and two-dimensional chemical-structure graph information of drug molecules. As a demonstration, cross-validation and independent testing were performed with iADRGSE on a dataset of ADRs classified into 27 categories, based on SOC (system organ classification). In addition, experiments comparing iADRGSE with approaches such as NPF were conducted on the OMOP dataset, using the jackknife test method. Experiments show that iADRGSE was superior to existing state-of-the-art predictors.

1. Introduction

Adverse drug reactions (ADRs) or side effects are substantially harmful or distressing reactions, and are described as adverse responses to drugs beyond their anticipated therapeutic effects [1]. In the United States, it is estimated that ADRs result in over 100,000 patient deaths per year [2] and the cost of ADRs-related morbidity was USD 528.4 billion in 2016 [3]. The process of drug-development involves a lot of monetary resources because it involves a lot of clinical trials and tests [4]. Many ADRs are not detected in the early stages of drug development, owing to restricted trial samples and time [5]. Thus, ADRs not only jeopardize patient health but also result in wasted healthcare costs, and are considered as a major global public-health problem. Traditional laboratory experiments to identify potential ADRs are not merely cumbersome and low cost-effective, but also less effective in the earlier phase. In recent years, algorithms in silico have been employed to speed up the prediction process and reduce drug-development costs.
Among the existing studies, some utilize data mining to analyze potential ADRs from large amounts of data and various sources of information; others adopts machine learning methods to predict ADRs.
The available databases of ADRs have some limitations at present. The data collected by the spontaneous reporting systems (SRS) and FDA Adverse Event Reporting System (FAERS) are not comprehensive enough, and there are problems such as repeated declaration. Drugs in the Side Effect Resource (SIDER) are limited to FDA-approved drugs only. The content of the European Medicines Agency (EMA) and other large-scale databases is complicated, and has no special retrieval of ADRs, which cause a lot of inconvenience for the use of data. Considering the limitations of the existing database, some researchers have mined the relationship between drugs and ADRs from texts, including SRS (covering spontaneous reports of adverse drug-events by healthcare professionals or patients), clinical narratives written by healthcare professionals, and electronic health records where diagnostic records are stored [6]. Other valuable big-data sources include social media posts such as health-related tweets, blogs, and forums [7]. Jagannatha and Hong trained the recurrent-neural-network framework (RNN) at the sentence- and document-levels to extract medical events and their attributes from unstructured electronic health records, and revealed that all RNN models outperform the counterfactual regret minimization (CRF) models [8]. An RNN model based on bi-directional long short-term memory (BiLSTM) networks has been proposed to treat text in social media posts as a sequence of words, and two BiLSTM models (BiLSTM-M2 and BiLSTM-M3) initialized with pre-trained embeddings perform significantly better than the BiLSTM-M1 model using random initialized embedding, because the pre-trained word embeddings are more effective in capturing the semantic similarity of words [9]. Ding et al. adopted character embedding and word embedding and combined them via an embedding-level attention mechanism to permit the model to determine how much information was used from the character-level or word-level component [10]. Although the previous attention methods have obtained good results in predicting ADRs, they only extract the individual semantic information entailed in a single sentence representation. In order to capture the different semantic information represented by different parts of the sentence, Zhang et al. developed a multi-hop self-attention mechanism (MSAM) model, in which each attention step aims to obtain different attention weights for different segments, in an attempt to capture multifaceted semantic information for ADR detection [11]. A weighted online recurrent extreme-learning-machine (WOR-ELM) method has been exploited to discriminate the boundaries of adverse drug reactions mentioned in biomedical texts [12]. It can be concluded from the above studies that both LSTM and the gate recurrent unit (GRU) are valuable tools for extracting ADRs from textual data. However, the methods of mining the ADRs from the text can only be used after the drug has been introduced onto the market, and cannot be used for the drugs in the research process.
Machine learning methods used to identify ADRs can be divided into three categories: similarity-based, network-topology-based, and matrix-decomposition-based.
The similarity-based methods are based on the fact that similar drugs have similar properties. It has been recognized that drugs with similar chemical structures exhibit similar biological activities; similar drug targets induce similar signal-cascade reactions, so they have similar ADRs [13]. Zhang et al. proposed a new method of measuring drug–drug similarity named “linear neighborhood similarity” to predict potential ADRs of drugs [14]. The diversification of drug information can enhance the predictive capability of such methods. In addition to drug chemical structures, drug target proteins, and drug substituents, Zheng et al. also used drug treatment information to identify ADRs [15]. Seo et al. applied the similarity of single-nucleic-acid polymorphism, side-effect anatomical hierarchy, drug–drug interaction, and target, and finally achieved better results by integrating the four predictors, random forest, logic regression, XGBOOST, and naïve bayes, using neural networks [13]. Liang et al. used a novel computational framework based on a multi-view and multi-label learning method to construct important drug features to improve predictor accuracy [16]. These predictors are similar in the use of learning classification models, and the key difference lies in the vectorized representation of drugs and ADRs.
The associations between drugs and other entities in the above methods are not integrated into the vector, so useful information may be lost. For this reason, the network-based method is also used to predict ADRs, and the new ADRs are inferred from the constructed network. Emir established a structural similarity network of drug chemical formulas and an ADRs transmission network to predict the potential ADRs [17]. Knowledge graphs (KGs) and their embedding process have become a useful tool in recent years, as they can not only represent the rich relationships between entities, but also directly encode these complex relationships as vectors. Using KG embedding to vectorize drugs and other entities is expected to better characterize drugs and other nodes. Bean et al. constructed a KG with four nodes, and vectorized it using an adjacent matrix of drug nodes to predict ADRs [18]. Emir et al. used KG to unify heterogeneous data from multiple databases, and the prediction of ADRs was regarded as a multi-label classification [19]. Zhang et al. designed a novel knowledge-graph-embedding method based on the Word2Vec model, and constructed a logistic-regression classification model to detect potential ADRs [20].
The matrix-decomposition algorithm decomposes the adjacency matrix of drug-ADRs pairs into multiple matrixes, and reconstructs an adjacency matrix to identify new drug-ADRs pairs. Liu et al. proposed a method based on structural matrix-decomposition named as LP-SDA, which is a label communication framework that links the chemical structure of a drug with the FDA Adverse Event Reporting System to predict ADRs [21]. Timilsina et al. integrated the bipartite graph which expressed the drugs and ADRs’ interactive relationship and the drug–drug network, where the edges represent semantic similarity between drugs using a matrix factorization method and a diffusion-based model [22]. DivePred was developed by Xuan et al., based on non-negative matrix factorization using disease similarity, various drug features of drug chemical substructures, and drug target-protein domains [23].
In recent years, graph neural networks (GNN) have been widely applied in various fields, and focus on mining potential information from the network structure. GNN has demonstrated its outstanding capability in the representation of biomolecular structures and relationships between molecules, and has received wide attention in the life sciences [24]. Withnall et al. introduced attentional and limbic memory-schemes into an existing message-passing neural network framework, and the prediction performance of molecular properties has been improved [25]. Furthermore, self-attentive mechanisms are frequently utilized in the field of natural language processing and are capable of efficiently processing text sequence-data [26]. In training the DDI prediction model, Schwarz et al. found that the model with an attention mechanism performed better than deep-neural-network models without attention [27].
In the early stage of drug design, there is no other information except the chemical-structure information of the drug. If the above methods relied only on the molecular formula structure to predict ADRs, the performance was very poor. For example, Dey et al. achieved an AUC of only 0.72 when using only the chemical structure of the drug to predict side effects [28]. Inspired by the advantages of GNN and self-attentive mechanisms, we propose an ADR multi-label prediction model called iADRGSE, which includes a self-attentive module based on drug substructure sequences and a graph network module on drug chemical-structure maps. The structure of this dual-channel model can effectively adapt to the different structural information of drugs, and improve the ability to predict ADRs. To verify the performance of the model, we collected data from the adverse drug reaction classification system (ADRECS), and classified the types of ADRs into 27 categories, in accordance with system organ type (SOC). The iADRGSE demonstrated better performance than other state-of-the-art methods in a multi-label ADRs prediction task.

2. Results and Discussion

2.1. Evaluation Metrics

ADRs prediction is a multi-label classification problem. The quality of multi-label learning is evaluated as more complex than single-label classification, because each sample is a label set. The metrics such as accuracy, precision, recall, AUC, and AUPR are frequently used. The last four metrics set the parameter average = ‘macro’, which represents the average of the metrics independently calculated over the 27 labels. Their formulas are as follows:
A c c u r a c y = 1 N j = 1 N T P j + T N j T P j + F N j + F P j + T N j
P r e c i s i o n m a c r o = 1 L i = 1 L j = 1 N T P i j T P i j + F P i j
R e c a l l m a c r o = 1 L i = 1 L j = 1 N T P i j T P i j + F N i j
where TP, TN, FP, and NP denote true positive, true negative, false positive, and false negative, respectively. N stands for the count of samples, and L represents the number of labels. Accuracy represents the proportion of drugs that are correctly predicted. Precision stands for the fraction of drugs that are predicted to be positive which is actually correct. Recall means the fraction of drugs that are truly labeled as positive which is correctly predicted; AUC is the area under the receiver operating characteristic curve; AUPR indicates the area under the precision recall curve.

2.2. Parameter Setting

We randomly selected 90% of the collated 2248 drugs as a training dataset for constructing and training the prediction model, and the remaining 10% as an independent-testing dataset, to test the constructed model. The selection of hyperparameter and feature-evaluation experiments were all optimized by a five-fold cross-validation test.
There are four parameters which have the greatest impact on the performance of the iADRGSE deep learning model: parameter h for the number of heads in attention, parameter ε for the dropout rate, parameter ξ for the learning rate in the model training, and parameter δ for the L2-regularization. It was observed from Figure 1 that when h = 2, ε = 0.5, ξ = 0.001, δ = 0.001, the performance reached its optimal value. Generally speaking, multiple heads are preferable to single heads, but more heads are not necessarily better. As shown in Figure 1a, the performance of the model is similar when the heads are set to 4 or 8, and the AUC is increased by 4.41% when the heads are 2. The effect of L2-regularization on the model is illustrated in Figure 1b, where the model works better with this hyperparameter of 0.001.

2.3. Feature Evaluation

We assessed the impact of various combinations of drug features on the forecasting of ADRs, and used the model’s metric scores as an indicator of the usability of the feature combinations. The results for different hierarchical feature combinations are displayed in Table 1.
In this study, the graph channel generated the mutual information, the edge information, and the context information of the drug molecule graph. We found that removal of one or two of the graph channel features had little effect on the performance of the model, but if the graph channel features were all removed, the performance of the model decreased significantly, and the AUC especially was reduced by 5%. Consequently, graph-embedding features are fairly significant in the model. In addition, in order to demonstrate that self-attentive coding of sequential channels can extract task-related features, we carried out experiments without encoders. The experiments demonstrated that the performance of the model without the encoders decreased in all metrics, and the AUC was reduced by close to 3%. We also compared feature-fusion methods by applying the mean value algorithm with our method, and the results revealed that our model feature- fusion method was more competitive.

2.4. Comparison of Feature-Extraction Methods of iADRGSE and Several Classic Feature-Extraction Methods

We compared the feature-extraction methods of iADRGSE with other feature-extraction methods, such as CNN_FP2, BERT_smiles [29], and Attentive_FP [30]; the hyperparameter settings for these three baseline methods are given in Supplementary Table S1. CNN is the most frequently used deep learning method in the field of vision. CNN_FP2 is the method of mining FP2 information through the convolution neural network. BERT_smiles use pre-trained bidirectional encoder representations from transformers (BERT) to extract SMILES sequence features. Attentive is the method of mining molecular fingerprint information though the graph attention-based approach. The results of the proposed iADRGSE predictor and the above three feature-extraction methods are also presented in Table 1. It is not difficult to find that the CNN_FP2 approach is better than the BERT_smiles and Attentive_FP. Moreover, the iADRGSE_no_Gin method, which also uses FP2 as input, remarkably outperforms CNN_FP2 by approximately 4% in AUC and AUPR. This demonstrates the superiority of sequence-based self-attentive encoding methods in feature extraction. The proposed iADRGSE predictor in this paper remarkably outperformed the above three feature-extraction methods in all metrics from Table 1, and outperformed CNN_FP2 by approximately 3.15% in accuracy, 6.86% in precision, 9.19% in AUC, and 7.23% in AUPR.
Independent tests can better verify the robustness of the prediction models. We also tested the performance of the iADRGSE and the above three feature-extraction methods using the independent test set, and the results are listed in Table 2. The predictive results of iADRGSE are very stable, at approximately 0.8196, 0.7632, 0.7461, 0.7735 and 0.7950 for accuracy, precision, recall, AUC and AUPR, respectively. It can be observed from the table that the accuracy score obtained by the current iADRGSE is significantly higher than that of the other three models, as are the other three indicators, except that recall is slightly lower than for BERT_smiles.

2.5. Comparison with Existing Predictor

Our model only used the chemical structure of the drug, which is helpful for the detection of ADRs in the preclinical stage of drug development. To further illustrate our approach, we compared the performance of iADRGSE with those of other models employing only chemical structures (NFP [28], circular fingerprinting [31]), and two drug-safety signal-detection algorithms (MGPS [32] and MCEM [33]), using the jackknife test method. For convenience of comparison, the scores of the five indexes obtained by these five predictors based on the OMOP dataset are listed in Figure 2. It can be observed from the figure that the AUC obtained by the iADRGSE model is significantly higher than that of the existing predictors, and remarkably outperforms the finest result of the comparison method by approximately 7%, in AUC.
In addition, using only the chemical-structure information of the drug, our model achieved good performance on the drug-safety signal-detection task (AUC = 0.7877), which provides a favorable complementary approach for toxicity detection in the early stages of drug design.

2.6. Case study

In this section, we undertake a case study to demonstrate the usability of iADRGSE in practice. In accordance with the loss value, the top 100 drugs were selected for case analysis to verify the ability of the model to predict potential ADRs. Next, comparing the predicted results of these 100 drugs with the true values, we found 21 drugs with potential adverse effects, whose predicted values are found in Supplementary Table S2. These 21 drugs had a total of 23 pairs of potential adverse reactions, as shown in Figure 3. Finally, we analyzed the predicted results in detail, and mainly focused on 23 pairs of potential adverse reactions, in Table 3.
For these 23 pairs of potential adverse reactions, we applied the search tools provided by medthority.com (accessed on 2 December 2022), reports from clinicaltrial.gov, and the related literature in PubMed and et al., to find the supporting evidence for them. From Table 3, we can observe that 17 of the 23 pairs of potential adverse reactions have evidence for them, indicating that the accuracy of the model iADRGSE has been further improved. For instance, the drug Pomalidomide carries a risk of social circumstances, which was reported in the literature [34].

3. Materials and Methods

3.1. Dataset

ADRECS [35] is an adverse-drug-reaction database that contains 2526 drugs and 9375 types of ADRs. To guarantee the quality, the drugs data were screened strictly according to the following criteria: (1) drugs without PubChem_ID were removed because PubChem_ID should be used to acquire the drug SMILES in the PubChem database; (2) drugs having no SMILES were removed. After following strictly the above two procedures, we finally obtained an ADRs dataset that contained 2248 drugs. We classified the 9375 adverse-drug-reaction types into 27 categories according to system organ classification (SOC). Finally, we obtained 2248 drugs, of which 27 belong to one ADR attribute, 32 to two different ADR attributes and so on. Detailed information is shown in Figure 4.
For other details of the dataset, please see Figure 5 and Figure 6. It is apparent that the data is unbalanced. The label base and density of the dataset are 16.5 and 0.6111, respectively. The base and density are relevant to the learning hardness of the multi-label classifier, i.e., the lower the density and the larger the base, the more difficult the multi-label learning process [36]. The dataset can be downloaded from the website https://github.com/cathrienli/iADRGSE (accessed on 10 December 2022).
In this study, the same dataset as that investigated in Harpaz et al. [37] was adopted for demonstration. The reason we chose it as a comparison dataset for the current study is that the OMOP dataset is derived from real-world data, such as the FDA Adverse Event Reporting System (FAERS) and data reported in recent papers. This dataset consisted of 171 drugs and four ADRs (acute kidney injury, acute liver injury, acute myocardial infarction, and gastrointestinal bleeding). Dataset statistics are provided in Table 4.

3.2. Problem Formulation

The core of our work is to construct a one-to-many mapping F : d i l i j j = 1 N l between a set of drugs D = d i | 1 i N d and a set of ADR labels L = l j | l j 0 , 1 ,   1 j N l , where N d is the number of drugs and N l represents the number of labels. For the multi-label ADR task, we define the label l j = 1 if the drug belongs to the j-th ADR class; otherwise, l j = 0 . In this study, each drug is expressed by two parts: molecular structure maps and substructure sequences.

3.3. Overview of iADRGSE

The system architecture of iADRGSE is shown in Figure 7. The architecture can be divided into feature extraction and prediction modules. In the feature-extraction module, a dual-channel network with sequence channel and graph channel is constructed, to learn the various structural features of drugs. In the graph channel, drug molecules are represented as chemical structure graphs, and we use the pre-trained graph isomorphism network (GIN) [38] to obtain various physicochemical properties of drugs. The sequence channel is connected by three units of preprocessing, encoder and feedforward in tandem, which aims to extract the substructural features of drug molecules. In the preprocessing unit, word embedding is applied to generate dense vectors from drug substructure sequences, and these vectors are fed to a downstream module for feature mining. The encoder mainly utilizes the multi-head self-attention mechanism from the transformer [26] network to perform a weighted combination of substructure embeddings. The feedforward unit reduces the dimensionality of the encoded features to adapt to the subsequent prediction task. Finally, in the prediction module, we concatenate diverse structural features learned from the upstream phase and input them into the deep neural networks (DNN) to predict the ADR labels.

3.4. Drug Molecular Representation

The simplified molecular-input line entry system (SMILES) is a specification in the form of a line notation for describing the structure of chemical species, using short ASCII strings [35]. We invoked the RDKit [36] tool to convert the SMILES of the drug, d i , into a molecular structure graph g i = D , E , where node set D represents atoms and edge set E represents chemical bonds. Node information carries atomic attributes such as atom type, atomic number, atom degree, electrons, hybridization, aromatic, etc. Edge information involves bond type, conjugated, ring, etc. Inspired by MUFFIN [37], in this study, we adopted the information of the number and chirality of the atom and the type and direction of the bond.
The FP2 fingerprint format Is a path-based fingerprint, which can be generated by Open Babel [38]. FP2 can represent drugs as 1024-dimensional binary vectors according to chemical substructures, with each dimension indicating the presence or absence of the corresponding substructure. To avoid sparsity, the drug representation thus obtained is a 256-digit hexadecimal string.

3.5. Feature Learning

3.5.1. Graph Channel

In order to parse graph-structured data, the GIN model is used because of its powerful function in the field of the graph neural network. The GIN model is pre-trained on individual nodes as well as the entire graph, to learn local and global representations. The iADRGSE model applies the pre-trained models of deep-graph information maximum [39] (Infomax), raw edge-prediction [40] (Edge), and context prediction [41] (Context), based on the GIN architecture, to generate the mutual information, X m , the edge information, X e , and the context information, X c , of the drug molecule graph, respectively.
The information-extraction process includes a message-passing stage and a readout stage. The message passing is to conduct the aggregation function, M t , to collect the information of neighboring nodes and edges, and to fuse the aggregation information to the current node through the update function, U t . Therefore, message passing can be described as below:
m u t + 1 = v ϵ N u M t h u t ,   h v t ,   e u v
h u t + 1 = U t h u t , m u t + 1  
where t is the number of iterations, u denotes the uth node in the graph, g i , N u represents the adjacent nodes of node u , h u t stands for the intermediate state of node u at time, t , e u v   ϵ   E indicates the attributes of edge between u and v . In particular, both M t and U t are inherently linear layers with the same dimension weight matrix W, which have been used to transfer information between nodes. Finally, the eigenvector, X i , of graph g i is calculated by M a x P o o l i n g on the node representation at the t step. The readout phrase can be formulated as:
X i * = M a x P o o l i n g ( h u t   |   u     g i ) , m , e , c

3.5.2. Sequence Channel

Taking into account the sparsity of the substructure sequences, a word-embedding layer is used to preprocess the sequences. The substructure sequence is a hexadecimal string in which each of four substructures is represented by a hexadecimal digit. The word-embedding module assigns a dense learnable embedding-representation for each hexadecimal number, which are stored in a simple lookup table. The model retrieves the homologous word-embedding in accordance with the index (i.e., hexadecimal number) of the substructure. The layer-normalization [42] module re-standardizes the substructure-embedding vectors using the mean, μ , and variance, σ 2 , across the embedding dimension. Scale, γ , and bias, β , are learnable affine-transformation parameters and ϵ is a value added to the denominator for numerical stability. Therefore, the preprocessing feature, E i , of the drug, d i , is calculated as follows:
E i = E m b e d d i n g q i μ σ 2 + ϵ * γ + β
where E i     256 × d i m , d i m is the word-embedding dimension.
To explore the different types of relevance that may exist between molecular substructures, we employ a multi-head self-attention mechanism, consisting of multiple parallel self-attention layers to encode substructure sequences. Each input vector, E i , s , can be calculated out three new vectors, Q i , s , K i , s , and   V i , s based on three different linear transformation matrices, W q u e r y   , W k e y ,   and   W v a l u e , respectively:
Q i , s | K i , s | V i , s = E i , s W q u e r y | W k e y | W v a l u e
where s indexes the sth substructure embeddings in E i , s     0 , , 255 , [ Q i ,   s , K i ,   s , V i ,   s ]     1 × d v , [ W q u e r y ,   W k e y ,   W v a l u e ] d i m × d v ,   d v is the vector dimension. Based on the aforementioned three intermediate vector matrices, we perform an attention-weighted sum over all substructures. The attention scores refer to the influence coefficients between the upstream and downstream substructures of the sequence, and are computed by the scaled dot-product of each substructure vector. For each substructure, s , the attention score, α i , s , j , of it and the substructure j 0 , , 255 can be calculated as follows:
α i , s , j = S o f t m a x Q i , s T K i , j d v
These attention scores, α i , s , j , and the vector, V i , j , are weighted sum to generate a new vector, O i , s , to represent the substructure, s . All the substructure vectors are simultaneously operated in parallel to obtain the new latent feature, O i , of the drug, d i .
O i , s = j = 0 255 α i , s , j V i , s
O i = O i , 0 , .. , O i , s , .. , O i , 255
For the head number of self-attention H > 1 , the model actually runs the single-head self-attention function with H times based on the different parameter matrices W q u e r y h , W k e y h , W v a l u e h in parallel, and a new feature representation, O i h , of the drug can be acquired, based on the hth self-attention head. These output values are concatenated and once again linearly transformed by the parameter matric W o     h d v × d i m to obtain output O i     256 × d i m . The multi-head process is depicted as:
O i = c o n c a t O i 1 , .. , O i h , .. , O i H W o
Note that we set d v = d i m H . To avoid the gradient problem, we add residual connection [43] to the input and output of the multi-head self-attention layer. The connection trick is to element-wise sum the output, E i , of the previous preprocessing unit and the output, O i , of the current multi-head self-attention layer. Finally, the residual features are transmitted through a layer-normalization module.
The fully connected feedforward-network consists of two linear layers, a batch-normalization [44] layer and a non-linear activation function, in order to further abstract and compress the latent encoded representation from the previous unit. Note that the output, O i ,   of the encoder unit is flattened before the linear transformation. The algorithm of batch normalization is the same as layer normalization, and the difference lies in which dimension is biased. The mean and standard-deviation are calculated per dimension, over the mini-batches. Ultimately, the substructure feature representation of the output of the sequence channel can be formulated as follows:
X i f = L e a k y R e L U B N F l a t t e n O i W l a y e r 1 W l a y e r 2

3.5.3. Multi-Label Classification

We spliced four structural features as drug representation, including mutual information X i m , edge information X i e , context information X i c and substructure information X i f . That is, drug d i can be marked as:
X i = c o n c a t X i m , X i e , X i c , X i f
where X i 4 d i m , d i m denotes the dimension of each structural feature.
Subsequently, X i is fed into a single-layer linear network parameterized by W p r e d , and an activation function is employed to output a predicted probability vector, P i , where each component is deemed as the likelihood of a label. The process can be defined as follows:
P i =   σ X i W p r e d
where W p r e d   ϵ   4 d i m × N l , σ refers to the s i g m o i d function for each P i component.

4. Conclusions

In this study, we design a fast and effective prediction framework based on the fusion of graph embedding and self-attentive encoder features, named iADRGSE, to predict ADRs. Based on feature analysis and various kinds of experiments, the robustness and performance of iADRGSE is testified. The case study is conducted, in which the top 100 drugs are selected for analysis, and the study demonstrates that the model is competent in predicting the potential ADRs. For practical applications, a user-friendly online web server for iADRGSE is built at http://121.36.221.79/iADRGSE (accessed on 10 December 2022), which allows users to easily obtain results and brings great convenience to researchers.
iADRGSE obtains a better prediction performance than that of pervious methods. The primary reason is that iADRGSE fuses the graph-embedding and self-attentive encoder features of the drug, and these features are closely related to the prediction of ADRs.
It is anticipated that predictor iADRGSE will become a very useful tool for predicting ADRs at the early stage of drug discovery.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijms232416216/s1.

Author Contributions

X.C., X.X. and L.Y. conceptualized the methodology. M.C., X.C. and L.Y. did the analysis. M.C., X.C. and X.X. wrote the manuscript. X.X. supervised the study. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the grants from the National Natural Science Foundation of China (Nos. 32260154, 31860312), the China-Montenegro Intergovernmental S&T Cooperation (NO. 2018-3-3).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Edwards, I.R.; Aronson, J.K. Adverse drug reactions: Definitions, diagnosis, and management. Lancet 2000, 356, 1255–1259. [Google Scholar] [CrossRef] [PubMed]
  2. Lazarou, J.; Pomeranz, B.H.; Corey, P.N. Incidence of adverse drug reactions in hospitalized patients: A meta-analysis of prospective studies. JAMA 1998, 279, 1200–1205. [Google Scholar] [CrossRef] [PubMed]
  3. Pirmohamed, M.; James, S.; Meakin, S.; Green, C.; Scott, A.K.; Walley, T.J.; Farrar, K.; Park, B.K.; Breckenridge, A.M. Adverse drug reactions as cause of admission to hospital: Prospective analysis of 18 820 patients. BMJ 2004, 329, 15. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Dickson, M.; Gagnon, J.P. The cost of new drug discovery and development. Discov. Med. 2009, 4, 172–179. [Google Scholar]
  5. Whitebread, S.; Hamon, J.; Bojanic, D.; Urban, L. Keynote review: In vitro safety pharmacology profiling: An essential tool for successful drug development. Drug Discov. Today 2005, 10, 1421–1433. [Google Scholar] [CrossRef]
  6. Harpaz, R.; Vilar, S.; Dumouchel, W.; Salmasian, H.; Haerian, K.; Shah, N.H.; Chase, H.S.; Friedman, C. Combing signals from spontaneous reports and electronic health records for detection of adverse drug reactions. J. Am. Med. Inform. Assoc. 2013, 20, 413–419. [Google Scholar] [CrossRef] [Green Version]
  7. Tutubalina, E.; Alimova, I.; Miftahutdinov, Z.; Sakhovskiy, A.; Malykh, V.; Nikolenko, S. The Russian Drug Reaction Corpus and neural models for drug reactions and effectiveness detection in user reviews. Bioinformatics 2021, 37, 243–249. [Google Scholar] [CrossRef]
  8. Jagannatha, A.N.; Yu, H. Bidirectional RNN for medical event detection in electronic health records. In Proceedings of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; p. 473. [Google Scholar]
  9. Cocos, A.; Fiks, A.G.; Masino, A.J. Deep learning for pharmacovigilance: Recurrent neural network architectures for labeling adverse drug reactions in Twitter posts. J. Am. Med. Inform. Assoc. 2017, 24, 813–821. [Google Scholar] [CrossRef]
  10. Ding, P.; Zhou, X.; Zhang, X.; Wang, J.; Lei, Z. An attentive neural sequence labeling model for adverse drug reactions mentions extraction. IEEE Access 2018, 6, 73305–73315. [Google Scholar] [CrossRef]
  11. Zhang, T.; Lin, H.; Ren, Y.; Yang, L.; Xu, B.; Yang, Z.; Wang, J.; Zhang, Y. Adverse drug reaction detection via a multihop self-attention mechanism. BMC Bioinform. 2019, 20, 479. [Google Scholar] [CrossRef] [Green Version]
  12. El-Allaly, E.-D.; Sarrouti, M.; En-Nahnahi, N.; El Alaoui, S.O. An adverse drug effect mentions extraction method based on weighted online recurrent extreme learning machine. Comput. Methods Programs Biomed. 2019, 176, 33–41. [Google Scholar] [CrossRef] [PubMed]
  13. Seo, S.; Lee, T.; Kim, M.-H.; Yoon, Y. Prediction of Side Effects Using Comprehensive Similarity Measures. Biomed Res. Int. 2020, 2020, 1357630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Zhang, W.; Yue, X.; Liu, F.; Chen, Y.; Tu, S.; Zhang, X. A unified frame of predicting side effects of drugs by using linear neighborhood similarity. BMC Syst. Biol. 2017, 11, 101. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Zheng, Y.; Peng, H.; Ghosh, S.; Lan, C.; Li, J. Inverse similarity and reliable negative samples for drug side-effect prediction. BMC Bioinform. 2019, 19, 554. [Google Scholar] [CrossRef] [Green Version]
  16. Liang, X.; Zhang, P.; Li, J.; Fu, Y.; Qu, L.; Chen, Y.; Chen, Z. Learning important features from multi-view data to predict drug side effects. J. Cheminformatics 2019, 11, 79. [Google Scholar] [CrossRef] [Green Version]
  17. Emir Muñoz, M.; Nováček, V.; Vandenbussche, P.-Y. Using Drug Similarities for Discovery of Possible Adverse Reactions. AMIA Annu Symp Proc 2017, 2016, 924–933. [Google Scholar]
  18. Bean, D.M.; Wu, H.; Iqbal, E.; Dzahini, O.; Ibrahim, Z.M.; Broadbent, M.; Stewart, R.; Dobson, R.J. Knowledge graph prediction of unknown adverse drug reactions and validation in electronic health records. Sci. Rep. 2017, 7, 16416. [Google Scholar] [CrossRef] [Green Version]
  19. Emir, M.; Nováček, V.; Vandenbussche, P.-Y. Facilitating prediction of adverse drug reactions by using knowledge graphs and multi-label learning models. Brief. Bioinform. 2019, 20, 190–202. [Google Scholar]
  20. Zhang, F.; Sun, B.; Diao, X.; Zhao, W.; Shu, T. Prediction of adverse drug reactions based on knowledge graph embedding. BMC Med. Inform. Decis. Mak. 2021, 21, 38. [Google Scholar] [CrossRef]
  21. Liu, R.; Zhang, P. Towards early detection of adverse drug reactions: Combining pre-clinical drug structures and post-market safety reports. BMC Med. Inform. Decis. Mak. 2019, 19, 279. [Google Scholar] [CrossRef] [Green Version]
  22. Timilsina, M.; Tandan, M.; d’Aquin, M.; Yang, H. Discovering Links Between Side Effects and Drugs Using a Diffusion Based Method. Sci. Rep. 2019, 9, 10436. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Xuan, P.; Song, Y.; Zhang, T.; Jia, L. Prediction of Potential Drug–Disease Associations through Deep Integration of Diversity and Projections of Various Drug Features. Int. J. Mol. Sci. 2019, 20, 4102. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Xiong, J.; Xiong, Z.; Chen, K.; Jiang, H.; Zheng, M. Graph neural networks for automated de novo drug design. Drug Discov. Today 2021, 26, 1382–1393. [Google Scholar] [CrossRef] [PubMed]
  25. Withnall, M.; Lindelöf, E.; Engkvist, O.; Chen, H. Building attention and edge message passing neural networks for bioactivity and physical–chemical property prediction. J. Cheminformatics 2020, 12, 1. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; p. 30. [Google Scholar]
  27. Schwarz, K.; Allam, A.; Perez Gonzalez, N.A.; Krauthammer, M. AttentionDDI: Siamese attention-based deep learning method for drug–drug interaction predictions. BMC Bioinform. 2021, 22, 412. [Google Scholar] [CrossRef]
  28. Dey, S.; Luo, H.; Fokoue, A.; Hu, J.; Zhang, P. Predicting adverse drug reactions through interpretable deep learning framework. BMC Bioinform. 2018, 19, 476. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  30. Xiong, Z.; Wang, D.; Liu, X.; Zhong, F.; Wan, X.; Li, X.; Li, Z.; Luo, X.; Chen, K.; Jiang, H.; et al. Pushing the Boundaries of Molecular Representation for Drug Discovery with the Graph Attention Mechanism. J. Med. Chem. 2020, 63, 8749–8760. [Google Scholar] [CrossRef] [PubMed]
  31. Rogers, D.; Hahn, M. Extended-connectivity fingerprints. J. Chem. Inf. Modeling 2010, 50, 742–754. [Google Scholar] [CrossRef]
  32. DuMouchel, W. Bayesian data mining in large frequency tables, with an application to the FDA spontaneous reporting system. Am. Stat. 1999, 53, 177–190. [Google Scholar]
  33. Xiao, C.; Li, Y.; Baytas, I.M.; Zhou, J.; Wang, F. An MCEM Framework for Drug Safety Signal Detection and Combination from Heterogeneous Real World Evidence. Sci. Rep. 2018, 8, 1806. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Park, H.; Byun, J.M.; Yoon, S.-S.; Koh, Y.; Yoon, S.-W.; Shin, D.-Y.; Hong, J.; Kim, I. Cyclophosphamide addition to pomalidomide/dexamethasone is not necessarily associated with universal benefits in RRMM. PLoS ONE 2022, 17, e0260113. [Google Scholar] [CrossRef] [PubMed]
  35. Toropov, A.A.; Toropova, A.P.; Mukhamedzhanova, D.V.; Gutman, I. Simplified molecular input line entry system (SMILES) as an alternative for constructing quantitative structure-property relationships (QSPR). Indian J. Chemistry. Sect. A Inorg. Phys. Theor. Anal. 2005, 44, 1545–1552. [Google Scholar]
  36. Landrum, G. RDKit: Open-Source Cheminformatics and Machine Learning. Available online: https://www.rdkit.org (accessed on 12 September 2022).
  37. Chen, Y.; Ma, T.; Yang, X.; Wang, J.; Song, B.; Zeng, X. MUFFIN: Multi-scale feature fusion for drug–drug interaction prediction. Bioinformatics 2021, 37, 2651–2658. [Google Scholar] [CrossRef]
  38. O’Boyle, N.M.; Banck, M.; James, C.A.; Morley, C.; Vandermeersch, T.; Hutchison, G.R. Open Babel: An open chemical toolbox. J. Cheminformatics 2011, 3, 33. [Google Scholar] [CrossRef]
  39. Veličković, P.; Fedus, W.; Hamilton, W.L.; Liò, P.; Bengio, Y.; Hjelm, R.D. Deep Graph Infomax. arXiv 2018, arXiv:1809.10341. [Google Scholar]
  40. Hamilton, W.L.; Ying, R.; Leskovec, J. Representation learning on graphs: Methods and applications. arXiv 2017, arXiv:1709.05584. [Google Scholar]
  41. Hu, W.; Liu, B.; Gomes, J.; Zitnik, M.; Liang, P.; Pande, V.; Leskovec, J. Strategies for Pre-training Graph Neural Networks. arXiv 2020, arXiv:1905.12265. [Google Scholar]
  42. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
  43. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  44. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning 2015, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
Figure 1. Model performance with different hyperparameter settings. (a)The performance of models under different number of heads. (b) The performance of models under different L2-Regularization. (c) The performance of models under different dropout. (d) The performance of models under different Learning rate.
Figure 1. Model performance with different hyperparameter settings. (a)The performance of models under different number of heads. (b) The performance of models under different L2-Regularization. (c) The performance of models under different dropout. (d) The performance of models under different Learning rate.
Ijms 23 16216 g001
Figure 2. Results of the comparison method and iADRGSE on the OMOP dataset.
Figure 2. Results of the comparison method and iADRGSE on the OMOP dataset.
Ijms 23 16216 g002
Figure 3. Drug and ADR association graph, Ijms 23 16216 i001 represents the drug node, while Ijms 23 16216 i002 represents the adverse reaction node; the more drugs connected to the ADR, the larger the node of the ADR; the line connecting the drug node and the ADR node indicates that the drug has these adverse reactions, and the thickened line indicates the potential adverse reactions of the drug.
Figure 3. Drug and ADR association graph, Ijms 23 16216 i001 represents the drug node, while Ijms 23 16216 i002 represents the adverse reaction node; the more drugs connected to the ADR, the larger the node of the ADR; the line connecting the drug node and the ADR node indicates that the drug has these adverse reactions, and the thickened line indicates the potential adverse reactions of the drug.
Ijms 23 16216 g003
Figure 4. Number of drugs with one or more ADR types in the ADRs benchmark dataset.
Figure 4. Number of drugs with one or more ADR types in the ADRs benchmark dataset.
Ijms 23 16216 g004
Figure 5. Sample distribution plot: horizontal axis represents the sample size, vertical axis represents the 27 ADR labels; Ijms 23 16216 i003 represents positive samples, while Ijms 23 16216 i004 is negative samples.
Figure 5. Sample distribution plot: horizontal axis represents the sample size, vertical axis represents the 27 ADR labels; Ijms 23 16216 i003 represents positive samples, while Ijms 23 16216 i004 is negative samples.
Ijms 23 16216 g005
Figure 6. Label co-occurrence diagram, referring to the simultaneous occurrence of two labels; the green circle represents the label, and the size of the circle is the frequency of that ADR label; the red line connecting the two circles represents the simultaneous occurrence of these two ADR labels; the color shade of the edges indicates the number of times this group of labels appears; the darker the color, the more often this group of labels appears.
Figure 6. Label co-occurrence diagram, referring to the simultaneous occurrence of two labels; the green circle represents the label, and the size of the circle is the frequency of that ADR label; the red line connecting the two circles represents the simultaneous occurrence of these two ADR labels; the color shade of the edges indicates the number of times this group of labels appears; the darker the color, the more often this group of labels appears.
Ijms 23 16216 g006
Figure 7. iADRGSE framework. (a) Graph channel. We perform the RDKit tool to convert the drug SMILES into chemical structure graphs and feed them into a pretrained GIN network, to learn graph-based structural information. (b) Sequence channel. The preprocessing unit utilizes Open Babel software to generate molecular substructure sequences from the SMILES of drugs. Then, the substructure sequences are represented as one-dimensional sequence vectors through the embedding layer. Next, the correlation information of each substructure is extracted further, using the encoder unit with a multi-head self-attention mechanism. Finally, the feed-forward unit (a multi-fully connected layer) receives encoded data from the upper layer to obtain the final sequence-based structural information of drugs. (c) Prediction module. These two types of structural information are concatenated and then mapped to the size of the labels, through an affine transformation for multi-label prediction.
Figure 7. iADRGSE framework. (a) Graph channel. We perform the RDKit tool to convert the drug SMILES into chemical structure graphs and feed them into a pretrained GIN network, to learn graph-based structural information. (b) Sequence channel. The preprocessing unit utilizes Open Babel software to generate molecular substructure sequences from the SMILES of drugs. Then, the substructure sequences are represented as one-dimensional sequence vectors through the embedding layer. Next, the correlation information of each substructure is extracted further, using the encoder unit with a multi-head self-attention mechanism. Finally, the feed-forward unit (a multi-fully connected layer) receives encoded data from the upper layer to obtain the final sequence-based structural information of drugs. (c) Prediction module. These two types of structural information are concatenated and then mapped to the size of the labels, through an affine transformation for multi-label prediction.
Ijms 23 16216 g007
Table 1. Results for different combinations of baseline, iADRGSE, and iADRGSE features.
Table 1. Results for different combinations of baseline, iADRGSE, and iADRGSE features.
Features SetAccuracyPrecision (Macro)Recall (Macro)AUC (Macro)AUPR (Macro)
CNN_FP20.7802 ± 0.00890.6474 ± 0.02130.7255 ± 0.01250.6726 ± 0.01450.7037 ± 0.0156
BERT_smiles0.7754 ± 0.00840.6266 ± 0.02510.7246 ± 0.00910.6587 ± 0.02020.6987 ± 0.0150
Attentive_FP0.7638 ± 0.00990.6748 ± 0.01300.7431 ± 0.01530.5669 ± 0.02340.6362 ± 0.0137
E + S0.8074 ± 0.00830.7241 ± 0.02800.7350 ± 0.01450.7519 ± 0.02100.7590 ± 0.0156
C + S0.8008 ± 0.00710.7251 ± 0.01200.7342 ± 0.02600.7545 ± 0.01630.7605 ± 0.0810
I + S0.8044 ± 0.00810.7136 ± 0.02860.7282 ± 0.01720.7479 ± 0.01720.7533 ± 0.0172
E + C + S0.8100 ± 0.00810.7369 ± 0.02560.7384 ± 0.01360.7628 ± 0.01480.7709 ± 0.0135
E + I + S0.8078 ± 0.00810.7251 ± 0.02340.7342 ± 0.01380.7545 ± 0.01810.7605 ± 0.0138
C + I + S0.8065 ± 0.00830.7245 ± 0.03050.7393 ± 0.01230.7580 ± 0.01250.7682 ± 0.0145
iADRGSE_Gin0.7992 ± 0.00220.7450 ± 0.01030.7235 ± 0.00630.7358 ± 0.01130.7526 ± 0.0088
iADRGSE_no_Gin0.7900 ± 0.00570.6888 ± 0.03230.7506 ± 0.01760.7098 ± 0.01790.7428 ± 0.0120
iADRGSE_
no_attention
0.8028 ± 0.0110.7451 ± 0.02570.7302 ± 0.01170.7410 ± 0.01920.7619 ± 0.0139
iADRGSE_mean0.7938 ± 0.00620.6793 ± 0.03520.7441 ± 0.01320.7206 ± 0.01610.7426 ± 0.0156
iADRGSE (ours)0.8117 ± 0.00890.7434 ± 0.02660.7421 ± 0.01050.7674 ± 0.01470.7760 ± 0.0130
Note: E: Gin_Edge; C: Gin_Context; I: Gin_Infomax; S: based on sequence channel. iADRGSE_Gin: no sequence channel; iADRGSE_no_Gin: no graph channels; iADRGSE_no_attention: sequence channel has no self-attention; iADRGSE_mean: use the mean operation to fuse features. The best performance for each metric is shown in bold.
Table 2. Results of the baseline and iADRGSE on independent test sets.
Table 2. Results of the baseline and iADRGSE on independent test sets.
Features SetAccuracyPrecision (Macro)Recall (Macro)AUC (Macro)AUPR (Macro)
CNN_FP20.80210.69600.73910.69900.7566
BERT_smiles0.79490.64360.75230.65470.7196
Attentive_FP0.77940.57910.73140.53980.6507
iADRGSE0.81960.76320.74610.77350.7950
Table 3. Potential adverse-drug-reactions.
Table 3. Potential adverse-drug-reactions.
Drug NameADREvidence
PomalidomideSurgical and medical proceduresclinicaltrials.gov (all accessed on 2 December 2022)
PomalidomideSocial circumstancesPMID: 35085238
KetorolacSurgical and medical procedurescdek.liu.edu
Prochlorperazine edisylateInfections and infestationsbaxter.ca
Trametinib dimethyl sulfoxideEar and labyrinth disordersclinicaltrials.gov
TrifluoperazineInfections and infestationshealthline.com
Desipramine hydrochlorideRespiratory, thoracic and mediastinal disordersrxlist.com
Chlorpromazine hydrochlorideInfections and infestationsUnconfirmed
EletriptanCongenital, familial and genetic disordersUnconfirmed
2-[1-methyl-5-(4-methylbenzoyl)
3-pyrrol-2-yl]acetate
Musculoskeletal and connective-tissue disordersmedthority.com
AlpelisibEar and labyrinth disordersclinicaltrials.gov
ImipramineMusculoskeletal and connective-tissue disorderscchr.org.au
Delavirdine mesylateCongenital, familial and genetic disordersrochecanada.com
Delavirdine mesylateDelavirdine mesylateUnconfirmed
TiagabineCongenital, familial and genetic disordersUnconfirmed
Minocycline anionEndocrine disordersmedthority.com
Naratriptan hydrochlorideHepatobiliary disordersmedthority.com
SertralineSocial circumstancesmedthority.com
Amlodipine besylateInjury, poisoning and procedural complicationsPMID: 25097362
PalonosetronPregnancy, puerperium and perinatal conditionsUnconfirmed
RufinamideNeoplasms: benign, malignant and unspecified (incl cysts and polyps)clinicaltrials.gov
Carvedilol phosphateNeoplasms: benign, malignant and unspecified (incl cysts and polyps)clinicaltrials.gov
MaprotilineNeoplasms: benign, malignant and unspecified (incl cysts and polyps)Unconfirmed
Table 4. Dataset statistics.
Table 4. Dataset statistics.
DatasetsDrugADRS Labels
ADRECS224827
OMOP1714
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cheng, X.; Cheng, M.; Yu, L.; Xiao, X. iADRGSE: A Graph-Embedding and Self-Attention Encoding for Identifying Adverse Drug Reaction in the Earlier Phase of Drug Development. Int. J. Mol. Sci. 2022, 23, 16216. https://doi.org/10.3390/ijms232416216

AMA Style

Cheng X, Cheng M, Yu L, Xiao X. iADRGSE: A Graph-Embedding and Self-Attention Encoding for Identifying Adverse Drug Reaction in the Earlier Phase of Drug Development. International Journal of Molecular Sciences. 2022; 23(24):16216. https://doi.org/10.3390/ijms232416216

Chicago/Turabian Style

Cheng, Xiang, Meiling Cheng, Liyi Yu, and Xuan Xiao. 2022. "iADRGSE: A Graph-Embedding and Self-Attention Encoding for Identifying Adverse Drug Reaction in the Earlier Phase of Drug Development" International Journal of Molecular Sciences 23, no. 24: 16216. https://doi.org/10.3390/ijms232416216

APA Style

Cheng, X., Cheng, M., Yu, L., & Xiao, X. (2022). iADRGSE: A Graph-Embedding and Self-Attention Encoding for Identifying Adverse Drug Reaction in the Earlier Phase of Drug Development. International Journal of Molecular Sciences, 23(24), 16216. https://doi.org/10.3390/ijms232416216

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop