Next Article in Journal
Effects of Environmental and Electric Perturbations on the pKa of Thioredoxin Cysteine 35: A Computational Study
Previous Article in Journal
Facile Synthesis of NiCo2O4 Nanowire Arrays/Few-Layered Ti3C2-MXene Composite as Binder-Free Electrode for High-Performance Supercapacitors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adera2.0: A Drug Repurposing Workflow for Neuroimmunological Investigations Using Neural Networks

by
Marzena Lazarczyk
1,
Kamila Duda
2,
Michel Edwar Mickael
1,3,*,
Onurhan AK
4,
Justyna Paszkiewicz
5,
Agnieszka Kowalczyk
2,
Jarosław Olav Horbańczuk
6 and
Mariusz Sacharczuk
1,7,*
1
Department of Experimental Genomics, Institute of Genetics and Animal Biotechnology of the Polish Academy of Sciences, ul. Postepu 36A, Jastrzebiec, 05-552 Magdalenka, Poland
2
Centre for Preclinical Research and Technology, Department of Pharmacodynamics, Faculty of Pharmacy with the Laboratory Medicine Division, Medical University of Warsaw, Banacha 1B, 02-091 Warsaw, Poland
3
PM Research Center, Väpnaregatan 22, 58649 Linköping, Sweden
4
Department of Sociology, Queen’s University at Kingston, 99 University Ave, Kingston, ON K7L 3N6, Canada
5
Department of Health, John Paul II University of Applied Sciences in Biala Podlaska, Sidorska 95/97, 21-500 Biała Podlaska, Poland
6
Institute of Genetics and Animal Biotechnology of the Polish Academy of Sciences, ul. Postepu 36A, Jastrzebiec, 05-552 Magdalenka, Poland
7
Department of Pharmacodynamics, Faculty of Pharmacy with the Laboratory Medicine Division, Medical University of Warsaw, Banacha 1B, 02-091 Warsaw, Poland
*
Authors to whom correspondence should be addressed.
Molecules 2022, 27(19), 6453; https://doi.org/10.3390/molecules27196453
Submission received: 9 August 2022 / Revised: 25 September 2022 / Accepted: 26 September 2022 / Published: 30 September 2022

Abstract

:
Drug repurposing in the context of neuroimmunological (NI) investigations is still in its primary stages. Drug repurposing is an important method that bypasses lengthy drug discovery procedures and focuses on discovering new usages for known medications. Neuroimmunological diseases, such as Alzheimer’s, Parkinson’s, multiple sclerosis, and depression, include various pathologies that result from the interaction between the central nervous system and the immune system. However, the repurposing of NI medications is hindered by the vast amount of information that needs mining. We previously presented Adera1.0, which was capable of text mining PubMed for answering query-based questions. However, Adera1.0 was not able to automatically identify chemical compounds within relevant sentences. To challenge the need for repurposing known medications for neuroimmunological diseases, we built a deep neural network named Adera2.0 to perform drug repurposing. The workflow uses three deep learning networks. The first network is an encoder and its main task is to embed text into matrices. The second network uses a mean squared error (MSE) loss function to predict answers in the form of embedded matrices. The third network, which constitutes the main novelty in our updated workflow, also uses a MSE loss function. Its main usage is to extract compound names from relevant sentences resulting from the previous network. To optimize the network function, we compared eight different designs. We found that a deep neural network consisting of an RNN neural network and a leaky ReLU could achieve 0.0001 loss and 67% sensitivity. Additionally, we validated Adera2.0’s ability to predict NI drug usage against the DRUG Repurposing Hub database. These results establish the ability of Adera2.0 to repurpose drug candidates that can shorten the development of the drug cycle. The workflow could be download online.

1. Introduction

Drug repurposing represents a lifeline for the neuroimmunological drug industry. The rate of development of new drugs has dramatically slowed in recent years [1]. This phenomenon was accompanied by a significant increase in the cost of drug development [1]. A single novel drug might cost 1 billion USD to produce [2]. Drug repurposing constitutes a viable alternative to conventional drug development techniques. Repurposing drugs that have already been legalized for human usage can substantially reduce the costs accompanying the primary stages of drug discovery [3]. Moreover, repurposing novel drugs eliminates the delay faced by de novo drug development [4]. At present, numerous repurposed medications are being extensively used, such as azathioprine, which is commonly used for rheumatoid arthritis; currently, it is used for renal transplants [5]. Minoxidil, originally intended to treat ulcers, was repurposed for the treatment of severe hypertension when early studies revealed that it was a vasodilator [6]. Another example is thalidomide, which was initially licensed in the 1950s to treat morning sickness in expectant mothers before being repurposed in 2006 to treat multiple myeloma [7]. In addition, although chlorpromazine was initially tested as an antimalarial medication, it was shown to be more effective in the treatment of mania [8]. We previously repurposed zileuton, which was originally used as an inhibitor of the 5-lipoxygenase, as an Nrf2 activator to treat depression [4]. Thus, drug repurposing techniques have proven essential for improving the drug production cycle.
Several computational approaches aim to extract drug information from medical literature using question–answer drug repurposing systems. These approaches could be applied separately or as a part of a workflow. These approaches include data mining, network analysis, and machine learning [9]. (I) The data mining approach is based on text mining. Text mining is utilized to identify drug names, using part-of-speech tagging or the bag of words algorithms. Text mining approaches include polysearch [10], DNorm [11], and MEDIC [12,13]. The main drawback of this approach is the need for a large-sized dictionary to predict textual relationships. The dictionary used requires various resources to perform several processing steps, such as normalization, mapping, and maintenance. An alternative approach is to directly text mine databases (e.g., DrugBank) [14]. However, many of the chemical compounds currently investigated are not yet accessible through DrugBank. Another text mining approach is based on the use of semantics technologies. Semantic text mining substitutes the need for stored vocabulary with a semantic concept built of multi-word terms (ngrams) [15]. This system could prove useful in filtering out less relevant documents. However, it suffers from the need to store the semantics database. An alternative method of text mining is based on ABC occurrence (i.e., if A and B occur together, B and C occur together, then A and C may occur together) [16]. However, this system suffers from low sensitivity (a high degree of false positives). (II) Networks analysis techniques are being widely employed in drug repurposing approaches. These approaches do not only detect drug compounds but can also mine drug characteristics, such as usages, indications, actions, and targets [12]. A bipartite graph is a relevant example that demonstrates this approach. Hu and Agarwal combined the analysis of microarray expression profiles with CMap data to produce a network of drugs–genes [17]. The drawback of this workflow is that future investigations are limited to the microarray datasets investigated. Clustering techniques based on the calculated Jaccard coefficient for Kegg Medicus have also been utilized [18]. The disadvantage of this system is that it is limited by the Kegg database annotation [19] (III) Machine learning techniques utilized to extract drug information from immunological texts could be generically grouped into classification or sequence labeling while using regression-based approaches is less common [12]. Classification models employed to recognize drug entities include maximum entropy (ME) and support vector machine (SVM) [20,21]. One of the drawbacks of these classification methods is that they are unaware of the order of the tokens (e.g., words).
Combining natural language processing (NLP) techniques with machine learning could be used to increase the awareness of the order and relationships between entities in any given text [22]. The primary objective of NLP techniques is to recognize concepts and relationships. A typical NLP workflow includes text tokenization to detect words and sentence boundaries. Part-of-speech algorithms could be used to tag the type of words using different tags (e.g., nouns, verbs, and prepositions). Extracted words are then mapped into a biological category. Finally, a tree is built based on the text syntax [23]. Additional NLP functionality includes several methods, such as (i) assertion status detection, which classifies a medical term according to its status into one of four groups, namely “present”, “absent”, “conditional”, or “associated with someone else” [24]. Assertion status detection methods have been combined with AI methods, such as convolutional neural networks (CNNs) and long-short-term memory networks (LSTMs) [23]. (ii) Entity resolution is an intriguing concept that attempts to comprehend the relationships between words within a sentence (e.g., links possessive articles to the correct nouns) [25]. This method is used with LTSM and achieves high accuracy. However, it does not extract compound names. (iii) Relationship extraction techniques aim to extract semantic associations between two or more entities (e.g., drug–disease relationships) [26]. (iv) Named entity recognition (NER) techniques implemented using the hidden Markov model (HMM) or conditional random field(s) (CRF) are aware of the sentence order and have been shown to outperform classification methods on a variety of tasks [27]. (v) Recently, word-representation features seemed to show the highest efficiency [28]. Word representation features are generated through unsupervised machine learning algorithms on unstructured texts. One of the advantages of using word representation is that this method takes into account the semantic information of the words. Currently, Word2VEC is one of the most widely used methods [29]. However, embedding for the whole sentence has achieved higher performance. In the case of query–answer scenarios, the distance between the embedded matrices of the question and the answers is used as an indication of the relevance between the question and the answer. However, the performances of sentence embedding-based systems in this context are limited [4].
In the current study, we developed a question–answer system based on AI, we call it Adera2.0. The input to the system is a user query (i.e., a question). The system is composed of several phases. In the first phase, the PubMed IDs relevant to the question posed by the user are retrieved from PubMed using a Python function. The downloaded PDFs are then parsed using Tika and stored in JSON format [30]. In the next phase, we utilized the universal autoencoder to embed each sentence in each parsed PDF into a 1 × 512 matrix. Following that, we use our relevance network to classify sentences within each PDF based on their relevance to the given question [4]. This network input is the 512 × 1 matrix embedding of the posed question. The output is a matrix embedding of a predicted answer based on an optimized loss regression function. Embedding matrices are sorted based on their distance from the predicted answer matrix. In the next phase, we built an extraction network. Its primary function is to extract drug names from relevant sentences identified in the previous phase. To achieve this output, each sentence is parsed into words and embedded using the universal encoder. After that, we used the extraction network to predict a generic compound embedding. Parsed words were sorted based on their distance from the predicted embedded matrix. We compared the performance of eight network designs to achieve the task of extracting compound names. We cross-validated Adera2.0’s performance against a gold standard dataset (Drug Repurposing Hub, Broad Institute dataset) and the current literature [31]. We investigated repurposing anti-oxidant compounds that could be used to regulate Th17 cells in depression. Previously, we found that Th17 infiltrates the blood–brain barrier through a paracellular route causing depression-like behavior [32,33]. The main objective of the in silico validation step was to find anti-oxidant drugs that could be used to reduce the effect of Th17 in depression and score at least four out of five on Lipinski’s score. Out of the top ten compounds generated by our software, two were shown to be true candidates.

2. Methods

2.1. Overview of the Workflow

The workflow consists of four main phases: (i) building a JSON database, (ii) sentence embedding, (iii) computation of relevance, and (iv) extraction of compound names.
The first phase of the workflow (phase I) covers the aim of building a database of the JSON format containing parsed PDFs (Figure 1). This phase consists of five steps. The first step’s objective is to fetch the PubMed IDs related to the search query. This is accomplished by using the PubMed fetcher function available through the Metapub python library. This step uses the input query to search for recent PubMed articles that match the query terms. After that, in the second step, the workflow fetches the abstracts and keywords of the retrieved PubMed IDs. This is achieved through the use of the python library Keybert. The third step in this phase involves downloading the identified PDFs; this is done using the fetch_PDFs library. The fourth step of this phase is the parsing of each PDF into separate sentences and storing all PDFs in a single database using the TIKA library [30]. The fifth step includes writing all parsed sentences into a single JSON format.
The second phase constitutes sentence embedding. This is done by employing the universal encoder. The main function of this network is to embed each parsed sentence in the JSON dataset into a 512 × 1 matrix. The universal encoder uses a deep averaging network as its main architecture. In this system, the word embedding is averaged. The output of this step is used as the input for a feedforward deep neural network to generate sentence embedding. The training data for this model include (i) unsupervised training datasets from Wikipedia, web news, and question and answer websites. (ii) Additionally, the neural network was trained on the supervised data of the Stanford natural language inference system [34]. It is important to note that the Pearson correlation was used to estimate the model’s accuracy. The model was used to predict the similarity between sentence pairs. The test was validated against human judgment. Overall, the model accuracy was estimated to be 0.76.
The objective of the third phase is to measure the relevance between the embedding of the parsed query (i.e., question) and the embedding of each parsed sentence in the JSON database. This phase is performed through a neural network. The input of the neural network is the embedding matrix of the query. The network uses a MSE loss function with an adaptive moment estimation (ADAM) optimizer. The network was trained on the entries sampled from the SciQ dataset. SciQ is a scientific answer dataset of 13,679 entries divided into 3 sub-datasets (training, test, and validation) [35]. The output of this network is a 512 × 1 matrix that represents the prediction of the answer embedding. Then, all the sentences in the JSON database are sorted based on the square root of the distance between the embedding matrix of each of them and the output of the second neural network (i.e., the predicted answer).
The fourth phase’s purpose is to extract the compound names from each answer resulting from the previous step. To achieve that, we designed a deep neural network. The inputs of this network are the sorted embedding matrices resulting from phase 3. The network was trained on an in-house-generated database [36]. The database is a collection of one thousand manually curated entries. The database consists of two columns representing the sentence–drug relationship. The first column includes sentences retrieved from; PubMed articles, Drug bank [14], and the n2c2 NLP Research data sets [37]. The second column represents the drug compounds, which were manually identified in each sentence. The dataset was divided into three smaller-size datasets as follows: (i) training dataset with 500 entries, (ii) test dataset with 250 entries, and (iii) validation dataset with 250 entries. It is important to note that our current trained network can be used for transfer learning when new categories of data or a larger set of validated drug entries are available. Similar to the relevance network, the extraction network uses an MSE loss function with an ADAM optimizer. We compared the performance of eight different architectures for this network (Figure 1, Table 1 and Table 2). Given an embedded sequence in 512 × 1 matrix format, the network predicts a word embedding in a 512 × 1 matrix format. Then the distance between the embedding of each word of each of the identified sentences and the predicted embedding matrix is calculated. The distance between the matrices is considered to correlate linearly to the similarity between the predicted answer and each embedding word. To measure the performance of our extraction network (E), we performed a sensitivity analysis (Equation (1)). Additionally, we compared the sensitivity values with random sampling.
sensitivity = Number   of   True   predictions   Number   of   True   predictions + Number   of   False   predictions ,

2.2. Cross Validation Using Known Drugs

We validated Adera2.0’s abilities by measuring its sensitivity. We compared the performances of our eight novel networks, focusing on their ability to extract drug names from sentences (phase 4). We utilized three strategies: (i) posing a question based on a disease. (ii) posing a question based on a specific type of drug, and (iii) posing a question pertaining to a specific pathway. Sensitivity (Equation (1)) was estimated for each model. After that K-means clustering was used to compare the sensitivity levels between various models investigated.

2.3. Case Study

The objective of the case study is to repurpose a compound that could target the inflammatory process regulated by Th17 infiltration of the brain during depression [32]. The constraints on the characteristics of the repurposed drug include its ability to inhibit Th17 differentiation, enhance depression prognosis and pass through the blood–brain barrier as well as being biologically safe [38]. The question used to query the system is “What are anti-oxidant drugs”. The output from our model was further filtered by examining the drug’s physicochemical properties. The number of violations of Lipinski’s rule was calculated using molinspiration server (www.molinspiration.com) accessed on 15 August 2022 [39]. The clogP, solubility, mol-weight, Tpsa, drug-likeness, drug toxicity, mutagenicity probability, tumorigenicity, irritant-ability, and reproductive effectiveness drug score, were determined by employing Osiris Property Explorer (OPE; http://www.organic-chemistry.org/prog/peo/) accessed on 15 August 2022. We utilized the BBB predictor (http://www.cbligand.org/BBB), accessed on 15 August 2022 to predict the ability of all the repurposed compounds to cross the blood–brain barrier.

2.4. Availability and Implementation

To facilitate the use of Adera2.0, we have developed a Windows©-compatible workflow that is quick and reliable. The application with user instruction pages could be found on the GitHub repository at https://github.com/michel-phylo/ADERA2.0.

3. Results

3.1. Workflow

Adera2.0 is a python-based neural network workflow specially designed to repurpose drugs pertaining to the field of immunology. Adera2.0 consists of three neural networks. Its main novelty is its third neural network structure. The first network is a universal autoencoder. The second structure is a relevance network based on a similar design used in our previous publication called Adera1.0 [4]. To that end, Adera2.0 receives a user query, performs a PubMed search, embeds the query, parses, and embeds each sentence in the retrieved PubMed PDFs. Following that, the distances between the matrices are calculated. The matrix with the shortest distance to the predicted embedding matrix is considered to represent the most relevant sentence. The novel-added network’s primary function is extracting drug compounds from relevant sentences resulting from the previous network (Figure 2). This is done by feeding the network the embedding matrix of the query. The network predicts an embedding representing a generic compound. After that, the distance between each word embedding and the predicted matrix is calculated. The word with the shortest distance between its matrix embedding and the predicted compound matrices is hypothesized to represent the compound.

3.2. Results of the Novel Neural Network

The third neural network extracts compound names from sentences. The design of the network layers played a vital role in determining the loss value. We compared the performance of eight different designs based on their ability to extract compound names from sentences (Table 1 and Table 2, Figure 3 and Figure 4). The core of networks A, EF, G, and H is an RNN network, while designs B, C, and D only possess dense layers with various activation functions (Table 1 and Figure 3). Networks A, E, F, G, and H achieved the lowest loss of 0.0013 and the lowest mean absolute error (0.02). Similar results were achieved on the test and the validation datasets. Conversely, networks B, C, and D achieved the highest loss (0.0017, 0.0019, 0.0017 respectively). Their poor performance was mirrored in the high mean absolute error on the validation network of 0.0377, 0.0348, and 0.0339 respectively (Table 2). It is noticeable that networks A, B, C, and G reached their maximum performance with lower than 20 epochs. The difference between training and validation network D seems to be increasing after 50 epochs, suggesting this particular network did not converge. Network F, H continued to learn even after more than 90 epochs, indicating that they perform better than the rest of the designs.
Table 1. Different structures used for network number (3): Extraction network.
Table 1. Different structures used for network number (3): Extraction network.
ArchitectureDescription
ADense(512, activation = ‘selu’)
Dense(512, activation = ‘sigmoid’)
Dense(512, activation = ‘relu’)
Dense(512, activation = ‘relu’)
Dense(512, activation = ‘selu’)
LeakyReLU()
tf.keras.layers.Activation(‘selu’)
SimpleRNN(512)
Dense(512, activation = ‘selu’)
Dense(512, activation = ‘relu’)
Dense(512, activation = ‘selu’)
Dense(512, activation = ‘relu’)
Dense(512, activation = ‘selu’)
BDense(512, activation = ‘selu’)
Dense(512, activation = ‘sigmoid’)
CDense(512, activation = ‘relu’)
Dense(512, activation = ‘selu’)
Dense(512, activation = ‘sigmoid’)
DDense(512, activation = ‘selu’)
Dense(512, activation = ‘Softmax’)
PReLU()
EDense(512, activation = ‘selu’)
SimpleRNN(512)
FActivation(‘selu’)
RNN(tf.keras.layers.LSTMCell(512))
LeakyReLU()
LeakyReLU()
GDense(512, activation = ‘selu’)
PReLU()
Dense(512, activation = ‘selu’)
SimpleRNN(512)
Activation(‘selu’)
LeakyReLU()
HActivation(‘selu’)
RNN(tf.keras.layers.LSTMCell(512))
PReLU()
LeakyReLU()
LeakyReLU()
Table 2. Loss and MAE comparison for Extraction network various architectures.
Table 2. Loss and MAE comparison for Extraction network various architectures.
Train DatasetTest DatasetValidate
ArchitectureLossMean Absolute_ErrorValidation LossValidation Mean Absolute ErrorLossMean Absolute_ErrorLossMean Absolute_Error
A0.00130.02900.00130.02860.00130.02880.00120.0283
B 0.0017 0.0333 0.0018 0.03380.00200.03540.00250.0377
C0.00190.03510.00180.03490.00180.03500.00180.0348
D0.00170.03420.00180.03420.00180.03410.00170.0339
E 0.0013 0.0295 0.0014 0.0306 0.00150.03090.00150.0310
F0.00130.02940.00130.02910.00130.02930.00130.0290
G0.00130.02910.00130.02870.00130.02930.00130.0290
H0.00130.02960.00130.02930.00130.02940.00130.0292

3.3. Sensitivity Analysis

We performed a sensitivity analysis to examine the performance of our proposed extraction networks. Our results indicate that the highest achieving architecture is network H, as it achieved a mean of 73.2% ± 11.5 with a p-value of 0.007 (Figure 4). In the case of a random sampling of data extraction, the sensitivity falls to 6.9 ± 3 % (p-value < 0.0001). Overall, our data indicate that compound extraction model H significantly outperforms random sampling of the data (p-value < 0.0007). Using the profiler function in TensorFlow we analyzed the relative time and memory efficiency values for the models used. We used the GoogleColab Nvidia K80/T4 (host machine) and NVIDIA GeForce GT 705 (local machine) to execute the training, testing, and validation process. Our results indicate that using RNN layers increases computation time, with networks AEFGH having a mean of 7.04 ± 1.87 ms, while networks BCD have a mean was 4.57 ± 0.20 ms. It could be noted that having multiple layers, as in the case of model A, resulted in having a slightly higher peak memory usage using the local device and on the Google Cloud host (Supplementary Materials).

3.4. In Silico Validation of Adera2.0 Ability to Mine Known Drugs

We validated Adera2.0’s abilities by measuring its sensitivity to extract known drug names. We compared the performances of our eight novel networks, focusing on their ability to extract drug names from sentences (Phase IV) (Figure 5A–C). To measure the sensitivity, we used three groups of sentences (each group consisting of five sentences). These groups were formed based on their types into (i) generic, (ii) specific to a drug family, and (iii) specific to a drug pathway. For the first category, we used the question “what are the drugs commonly used to treat cardiovascular diseases?”. For this category, Models G and H had the highest mean of 0.6 with a standard deviation of 0.15 (ii) In the case of the second category, we used the question; “What are the types of statins?”. All the models scored lower accuracy, with the highest being Model E (scoring 0.6). (iii) In the third category, we used the question, “Which drugs function as PPARγ agonists?”. In this category, Model F scored the highest with a value of 0.67. Overall, Model F was more accurate in extracting drug names from sentences related to biological pathways. Models G and H seemed to be more suitable for sentences acquired from more generic inquiries.

3.5. In Silico Validation Demonstrates Adera Performance

Adera2.0 is the new version of our previous publication (Adera1.0) [4]. The main differences between the two network architectures are highlighted in Table 3. In summary, we updated our workflow of Adera1.0 considerably to include multiple novel features. We added a neural network capable of extracting compound names from sentences. Network “H” was chosen to build the extraction network after comparing eight different neural network architectures. The neural network accepts the embedding matrix of the user query and outputs a 512x 1 matrix. The distance between the output matrix and the word embedding matrices of each relevant sentence is calculated to determine the shortest distance. Using this approach; a single compound name is extracted from each relevant sentence.
We demonstrated the ability of our workflow to repurpose drugs, by showing its ability to repurpose compounds based on their abilities to regulate the Th17 function in depression diseases. A query question was used to initiate the workflow. The query question was given as: “what are anti-oxidant drugs?”. The workflow produced ten compounds (Table 4). It could be noticed that the resulting compounds belong to different chemical and physical families. For example, coumarin is a phenylpropanoid used by plants to fend off animals [40]. It is mainly used as an anticoagulant. On the other hand, glycyrrhetinic acid is a triterpenoid derived from the root of licorice and could function as an anti-cancer and anti-inflammatory drug [41].
To examine the feasibility of our predictions to be used as Th17 inhibitors in depression, we subjected the compounds to 11 constraints. In the first set of constraints, pathogenicity was measured (Table 5). Several compounds were predicted not to have any pathogenicity, such as acebutolol, glycyrrhetinic acid, glycyrrhizin, porphyrin, bicalutamide, and fucoxanthin. Conversely, amygdalin, on the other hand, is predicted to be of high risk in three of the four categories, highlighting the limitations of our workflow in terms of being inherently unaware of toxicity. Moreover, our investigation of the physical and chemical properties of the repurposed compounds revealed that various compounds did not break any of Lipinski’s rules, such as acebutolol, beta-D-fructofuranose, coumarin, resveratrol, and bicalutamide, while glycyrrhizin and flavonoids broke three rules, deeming them weak on the level of oral absorbance. Furthermore, as our repurposing tasks are related to the ability of the compound to inhibit pathogenic Th17 function in the brain during depression, one of the main characteristics of the required drugs is blood–brain barrier infiltration ability. Our investigation revealed that coumarin, resveratrol, porphyrin, and bicalutamide were able to cross the BBB (Table 6 and Figure 6). We validated our findings by searching relevant literature and the Drug Repurposing Hub (Table 7). Notably, bicalutamide was inferred to decrease Th17 activity by inhibiting NF-κB signaling [42]. Paradoxically, it was reported to worsen depression prognosis. Conversely, the porphyrin derivative known as hematoporphyrin is regularly used as an antidepressant and antipsychotic. It was also used to inhibit the podoplanin–CLEC-2 interaction and decrease the incidence of metastasis and thrombosis [43]. Conversely, another derivative of porphyrin known as tin protoporphyrin IX was shown to inhibit Heme oxygenase-1, which may affect Th17 activity [44].
Bicalutamide scored 0.02 on the SVM-MACCCSFP BBB score indicating that it is highly likely to be able to cross the BBB.

4. Discussion

Our system offers a drug repurposing workflow. Thousands of literature documents are being added to PubMed and similar repositories each year. Mining drug databases is a viable option for drug repurposing. However, classical approaches use large dictionaries to extract the compound from mined sentences. Word2VEC approaches offer a solution by embedding text to numerical matrices. However, it lacks awareness of sentence context. Our updated workflow is both able to determine sentence relevance and also to extract compound names with a loss of less than 0.01 and a sensitivity of 67% (Table 2, Figure 4 and Figure 5).

4.1. Output of the Extraction Network Highlight the Importance of Architecture

Our approach, as well as others, demonstrates that neural network prediction results are highly dependable on the architecture and the loss function. The extraction network’s main task is, given a sentence embedding, to predict a word embedding representing a generic compound. To do this, we trained our network on an in-house dataset of 1000 entries. Our findings, in agreement with previous reports, suggest that RNN could be most suited for compound name extraction tasks. Notably, designs that lack the RNN structures perform significantly worse (Figure 3 and Figure 4). This observation was further validated by inspecting the ability of our eight models to extract known drug names. K-means clustering analysis (Figure 5C) showed clear clustering between Models A, E, F, G, and H (cluster I) and models B, C, and D (cluster II). RNNs are known to perform better than CNN-based models using a multi-classifier framework. CNN was reported not to be able to grasp the text context [60]. RNNs can predict time series and have been used successfully for text normalization, de-identification, and sequence labeling [61,62]. However, RNNs also suffer from low efficiency in multi-classifier settings [63]. In this report, we proposed a novel network design to extract drug compounds from sentences using a regression loss function (i.e., mean square root error (MSE)). Our model “H” which utilizes a simple RNN achieves a loss of less than 0.0013 as well as a sensitivity of an average of 67% (Figure 3, Figure 4 and Figure 5).

4.2. Case Study Results Show Our Workflow Predictions Accuracy

We demonstrated the ability of our workflow to correctly repurpose a compound that can target Th17 expansion in the brain. We have previously shown that Th17 is capable of infiltrating the brain during depression diseases, causing a localized inflammation of the hypothalamus. This inflammation is associated with depression-like behavior [64,65,66]. Currently, depression drugs focus on the role of neurotransmitters and overlook the inflammatory side [67]. Thus, repurposing drugs to target the inflammatory symptoms of depression can help ease the symptoms and ensure a better quality of life. We used Adera to mine PubMed PDFs for anti-oxidant drugs. We investigated a list of ten drugs for their ability to fulfill the needed objective (i.e., target Th17 expansion in the brain during depression). Additionally, we subjected the resulting drugs to a rigorous list of constraints that ensured that the repurposed drug would fulfill the objective while being biologically safe, orally available, and non-toxic. Our investigation proposes the use of porphyrin and bicalutamide. Our findings are supported by the literature, which shows that both drugs have interacted with the effect on Th17 and are regularly used to improve the prognosis of depression [31,38,44].

4.3. Limitations and Future Direction

Currently, Adera2.0 is unaware of the structure of the repurposed drug. Downstream analysis using molecular dynamics, molecular docking, and QM/MM programs are still essential to reduce the cost of the in vitro and in vivo validation of drugs that could be repurposed using our software. Future improvements to the software will include neural networks predicting protein–drug interactions to overcome this limitation.

5. Conclusions

Our workflow (Adera2.0) is capable of text mining highly specific PDFs to search for drugs that could be used in an immunological context. Our workflow is capable of automatically downloading and parsing PDFs from PubMed. The workflow predicts the relevance between each parsed sentence and the user query. The workflow extracts compound names from relevant sentences with a sensitivity in the 60–70% range. Overall, the use of Adera may reduce the time and costs required for R&D in drug discovery.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/molecules27196453/s1, Table S1: Relative time efficiency of the models used. Table S2: Relative memory efficiency of the models used (for the local memory). Table S3: Relative memory efficiency of the models used (on Google cloud).

Author Contributions

Conceptualization, M.E.M., J.O.H. and M.S.; data curation, M.L., K.D. and M.E.M.; formal analysis, M.L., J.P., O.A., K.D. and M.E.M.; investigation, M.L., O.A., M.E.M., J.P. and A.K.; methodology, M.L., O.A. and A.K.; supervision, J.O.H. and M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financed by the NLnet Foundation, grant number RfP 2021-10-039.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this manuscript is available through the following link: https://data.mendeley.com/datasets/whr7wrh42y/1.

Acknowledgments

We would like to thank Macrious Abraham and Meriam Joachim for their fruitful discussions and inspiration. We would like to thank the teams at NLnet and COST for their financial support. Additionally, we would like to acknowledge the authors of the open-source python libraries KeyBert and MetaPub.

Conflicts of Interest

The authors declare no conflict of interest.

Sample Availability

Not applicable.

References

  1. Scannell, J.W.; Blanckley, A.; Boldon, H.; Warrington, B. Diagnosing the Decline in Pharmaceutical R&D Efficiency. Nat. Rev. Drug Discov. 2012, 11, 191–200. [Google Scholar] [CrossRef] [PubMed]
  2. Prasad, V.; Mailankody, S. Research and Development Spending to Bring a Single Cancer Drug to Market and Revenues After Approval. JAMA Intern. Med. 2017, 177, 1569. [Google Scholar] [CrossRef] [PubMed]
  3. Simsek, M.; Meijer, B.; van Bodegraven, A.A.; de Boer, N.K.H.; Mulder, C.J.J. Finding Hidden Treasures in Old Drugs: The Challenges and Importance of Licensing Generics. Drug Discov. Today 2018, 23, 17–21. [Google Scholar] [CrossRef]
  4. Kubick, N.; Pajares, M.; Enache, I.; Manda, G.; Mickael, M.-E. Repurposing Zileuton as a Depression Drug Using an AI and In Vitro Approach. Molecules 2020, 25, 2155. [Google Scholar] [CrossRef] [PubMed]
  5. Anstey, A.; Lear, J.T. Azathioprine: Clinical Pharmacology and Current Indications in Autoimmune Disorders. BioDrugs 1998, 9, 33–47. [Google Scholar] [CrossRef] [PubMed]
  6. Watkins, J.; Dargie, H.J.; Bune, A.; Dollery, C.T. Reduction of Beta-Blocking Drugs in Hypertensive Patients Treated with Minoxidil. BMJ 1979, 1, 1400. [Google Scholar] [CrossRef]
  7. Mercurio, A.; Adriani, G.; Catalano, A.; Carocci, A.; Rao, L.; Lentini, G.; Cavalluzzi, M.M.; Franchini, C.; Vacca, A.; Corbo, F. A Mini-Review on Thalidomide: Chemistry, Mechanisms of Action, Therapeutic Potential and Anti-Angiogenic Properties in Multiple Myeloma. Curr. Med. Chem. 2017, 24, 2736–2744. [Google Scholar] [CrossRef]
  8. López-Muñoz, F.; Alamo, C.; Cuenca, E.; Shen, W.; Clervoy, P.; Rubio, G. History of the Discovery and Clinical Introduction of Chlorpromazine. Ann. Clin. Psychiatry 2005, 17, 113–135. [Google Scholar] [CrossRef]
  9. Nag, S.; Baidya, A.T.K.; Mandal, A.; Mathew, A.T.; Das, B.; Devi, B.; Kumar, R. Deep Learning Tools for Advancing Drug Discovery and Development. 3 Biotech 2022, 12, 110. [Google Scholar] [CrossRef]
  10. Cheng, D.; Knox, C.; Young, N.; Stothard, P.; Damaraju, S.; Wishart, D.S. PolySearch: A Web-Based Text Mining System for Extracting Relationships between Human Diseases, Genes, Mutations, Drugs and Metabolites. Nucleic Acids Res. 2008, 36, W399–W405. [Google Scholar] [CrossRef] [Green Version]
  11. Leaman, R.; Islamaj Dogan, R.; Lu, Z. DNorm: Disease Name Normalization with Pairwise Learning to Rank. Bioinformatics 2013, 29, 2909–2917. [Google Scholar] [CrossRef] [PubMed]
  12. Jarada, T.N.; Rokne, J.G.; Alhajj, R. A Review of Computational Drug Repositioning: Strategies, Approaches, Opportunities, Challenges, and Directions. J. Cheminform. 2020, 12, 46. [Google Scholar] [CrossRef] [PubMed]
  13. Davis, A.P.; Wiegers, T.C.; Rosenstein, M.C.; Mattingly, C.J. MEDIC: A Practical Disease Vocabulary Used at the Comparative Toxicogenomics Database. Database 2012, 2012, bar065. [Google Scholar] [CrossRef] [PubMed]
  14. Wishart, D.S. DrugBank: A Comprehensive Resource for in Silico Drug Discovery and Exploration. Nucleic Acids Res. 2006, 34, D668–D672. [Google Scholar] [CrossRef]
  15. Hansson, L.K.; Hansen, R.B.; Pletscher-Frankild, S.; Berzins, R.; Hansen, D.H.; Madsen, D.; Christensen, S.B.; Christiansen, M.R.; Boulund, U.; Wolf, X.A.; et al. Semantic Text Mining in Early Drug Discovery for Type 2 Diabetes. PLoS ONE 2020, 15, e0233956. [Google Scholar] [CrossRef]
  16. Henry, S.; Panahi, A.; Wijesinghe, D.S.; McInnes, B.T. A Literature Based Discovery Visualization System with Hierarchical Clustering and Linking Set Associations. AMIA Jt. Summits Transl. Sci. Proc. 2019, 2019, 582–591. [Google Scholar]
  17. Hu, G.; Agarwal, P. Human Disease-Drug Network Based on Genomic Expression Profiles. PLoS ONE 2009, 4, e6536. [Google Scholar] [CrossRef]
  18. Kanehisa, M.; Goto, S. KEGG: Kyoto Encyclopedia of Genes and Genomes. Nucleic Acids Res. 2000, 28, 27–30. [Google Scholar] [CrossRef]
  19. Wu, C.; Gudivada, R.C.; Aronow, B.J.; Jegga, A.G. Computational Drug Repositioning through Heterogeneous Network Clustering. BMC Syst. Biol. 2013, 7, S6. [Google Scholar] [CrossRef]
  20. Liu, S.; Tang, B.; Chen, Q.; Wang, X. Drug Name Recognition: Approaches and Resources. Information 2015, 6, 790–810. [Google Scholar] [CrossRef]
  21. Halgrim, S.; Xia, F.; Solti, I.; Cadag, E.; Uzuner, Ö. A Cascade of Classifiers for Extracting Medication Information from Discharge Summaries. J. Biomed. Semant. 2011, 2, S2. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Vamathevan, J.; Clark, D.; Czodrowski, P.; Dunham, I.; Ferran, E.; Lee, G.; Li, B.; Madabhushi, A.; Shah, P.; Spitzer, M.; et al. Applications of Machine Learning in Drug Discovery and Development. Nat. Rev. Drug Discov. 2019, 18, 463–477. [Google Scholar] [CrossRef] [PubMed]
  23. Bhatnagar, R.; Sardar, S.; Beheshti, M.; Podichetty, J.T. How Can Natural Language Processing Help Model Informed Drug Development?: A Review. JAMIA Open 2022, 5, ooac043. [Google Scholar] [CrossRef] [PubMed]
  24. Uzuner, Ö.; South, B.R.; Shen, S.; DuVall, S.L. 2010 I2b2/VA Challenge on Concepts, Assertions, and Relations in Clinical Text. J. Am. Med. Inform. Assoc. 2011, 18, 552–556. [Google Scholar] [CrossRef]
  25. Bell, D.; Hahn-Powell, G.; Valenzuela-Escárcega, M.A.; Surdeanu, M. An Investigation of Coreference Phenomena in the Biomedical Domain. In Proceedings of the 10th International Conference on Language Resources and Evaluation, Portorož, Slovenia, 23–28 May 2016. [Google Scholar]
  26. Wei, C.-H.; Peng, Y.; Leaman, R.; Davis, A.P.; Mattingly, C.J.; Li, J.; Wiegers, T.C.; Lu, Z. Overview of the BioCreative V Chemical Disease Relation (CDR) Task. Proc. Fifth BioCreative Chall. Eval. Work. 2015, 154–166. [Google Scholar]
  27. Ekbal, A.; Bandyopadhyay, S. A Hidden Markov Model Based Named Entity Recognition System: Bengali and Hindi as Case Studies. In Pattern Recognition and Machine Intelligence; Springer: Berlin/Heidelberg, Germany, 2007; pp. 545–552. ISBN 3540770453. [Google Scholar]
  28. Cer, D.; Yang, Y.; Kong, S.; Hua, N.; Limtiaco, N.; John, R.S.; Constant, N.; Guajardo-Cespedes, M.; Yuan, S.; Tar, C.; et al. Universal Sentence Encoder. arXiv 2018. [Google Scholar] [CrossRef]
  29. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  30. Mattmann, C.A.; Zitting, J.L. Tika in Action; Manning: New York, NY, USA, 2012; ISBN 9781935182856. [Google Scholar]
  31. Corsello, S.M.; Bittker, J.A.; Liu, Z.; Gould, J.; McCarren, P.; Hirschman, J.E.; Johnston, S.E.; Vrcic, A.; Wong, B.; Khan, M.; et al. The Drug Repurposing Hub: A next-Generation Drug Library and Information Resource. Nat. Med. 2017, 23, 405–408. [Google Scholar] [CrossRef]
  32. Mickael, M.E.; Bhaumik, S.; Chakraborti, A.; Umfress, A.A.; van Groen, T.; Macaluso, M.; Totenhagen, J.; Sorace, A.G.; Bibb, J.A.; Standaert, D.G.; et al. RORγt-Expressing Pathogenic CD4 + T Cells Cause Brain Inflammation during Chronic Colitis. J. Immunol. 2022, 208, 2054–2066. [Google Scholar] [CrossRef]
  33. Mickael, M.-E.; Kubick, N.; Klimovich, P.; Flournoy, P.H.; Bieńkowska, I.; Sacharczuk, M. Paracellular and Transcellular Leukocytes Diapedesis Are Divergent but Interconnected Evolutionary Events. Genes 2021, 12, 254. [Google Scholar] [CrossRef]
  34. Bowman, S.R.; Angeli, G.; Potts, C.; Manning, C.D. A Large Annotated Corpus for Learning Natural Language Inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 632–642. [Google Scholar] [CrossRef]
  35. Welbl, J.; Liu, N.F.; Gardner, M. Crowdsourcing Multiple Choice Science Questions. In Proceedings of the 3rd Workshop on Noisy User-Generated Text, Copenhagen, Denmark, 7 September 2017. [Google Scholar] [CrossRef]
  36. Mickael, M.E.K. Adera2.0 Text Mining Dataset for Training Neural Networks. 2022. Available online: https://data.mendeley.com/datasets/whr7wrh42y/1 (accessed on 8 August 2022).
  37. Henry, S.; Buchan, K.; Filannino, M.; Stubbs, A.; Uzuner, O. 2018 N2c2 Shared Task on Adverse Drug Events and Medication Extraction in Electronic Health Records. J. Am. Med. Inform. Assoc. 2020, 27, 3–12. [Google Scholar] [CrossRef] [PubMed]
  38. Bhaumik, S.; Basu, R. Cellular and Molecular Dynamics of Th17 Differentiation and Its Developmental Plasticity in the Intestinal Immune Response. Front. Immunol. 2017, 8, 254. [Google Scholar] [CrossRef] [PubMed]
  39. Chen, X.; Li, H.; Tian, L.; Li, Q.; Luo, J.; Zhang, Y. Analysis of the Physicochemical Properties of Acaricides Based on Lipinski’s Rule of Five. J. Comput. Biol. 2020, 27, 1397–1406. [Google Scholar] [CrossRef]
  40. Jacobowitz, J.R.; Weng, J.-K. Exploring Uncharted Territories of Plant Specialized Metabolism in the Postgenomic Era. Annu. Rev. Plant Biol. 2020, 71, 631–658. [Google Scholar] [CrossRef]
  41. Graebin, C.S. The Pharmacological Activities of Glycyrrhizinic Acid (“Glycyrrhizin”) and Glycyrrhetinic Acid. Sweeteners 2018, 245–261. [Google Scholar] [CrossRef]
  42. Zhong, S.; Huang, C.; Chen, Z.; Chen, Z.; Luo, J.-L. Targeting Inflammatory Signaling in Prostate Cancer Castration Resistance. J. Clin. Med. 2021, 10, 5000. [Google Scholar] [CrossRef]
  43. Quintanilla, M.; Montero, L.M.; Renart, J.; Villar, E.M. Podoplanin in Inflammation and Cancer. Int. J. Mol. Sci. 2019, 20, 707. [Google Scholar] [CrossRef]
  44. Zhang, Y.; Zhang, L.; Wu, J.; Di, C.; Xia, Z. Heme Oxygenase-1 Exerts a Protective Role in Ovalbumin-Induced Neutrophilic Airway Inflammation by Inhibiting Th17 Cell-Mediated Immune Response. J. Biol. Chem. 2013, 288, 34612–34626. [Google Scholar] [CrossRef]
  45. Krouf, D.; Bouchenak, M.; Mohammedi, B.; Cherrad, A.; Belleville, J.L.; Prost, J.L. Changes in Serum Lipids and Antioxidant Status in West Algerian Patients with Essential Hypertension Treated with Acebutolol Compared to Healthy Subjects. Med. Sci. Monit. 2003, 9, 109–115. [Google Scholar]
  46. Al-Khayri, J.M.; Sahana, G.R.; Nagella, P.; Joseph, B.V.; Alessa, F.M.; Al-Mssallem, M.Q. Flavonoids as Potential Anti-Inflammatory Molecules: A Review. Molecules 2022, 27, 2901. [Google Scholar] [CrossRef]
  47. Pannu, A.; Sharma, P.C.; Thakur, V.K.; Goyal, R.K. Emerging Role of Flavonoids as the Treatment of Depression. Biomolecules 2021, 11, 1825. [Google Scholar] [CrossRef] [PubMed]
  48. Kostova, I.; Bhatia, S.; Grigorov, P.; Balkansky, S.; Parmar, V.S.; Prasad, A.K.; Saso, L. Coumarins as Antioxidants. Curr. Med. Chem. 2011, 18, 3929–3951. [Google Scholar] [CrossRef] [PubMed]
  49. Yao, R.; Fu, Y.; Li, S.; Tu, L.; Zeng, X.; Kuang, N. Regulatory Effect of Daphnetin, a Coumarin Extracted from Daphne Odora, on the Balance of Treg and Th17 in Collagen-Induced Arthritis. Eur. J. Pharmacol. 2011, 670, 286–294. [Google Scholar] [CrossRef] [PubMed]
  50. Capra, J.C.; Cunha, M.P.; Machado, D.G.; Zomkowski, A.D.E.; Mendes, B.G.; Santos, A.R.S.; Pizzolatti, M.G.; Rodrigues, A.L.S. Antidepressant-like Effect of Scopoletin, a Coumarin Isolated from Polygala Sabulosa (Polygalaceae) in Mice: Evidence for the Involvement of Monoaminergic Systems. Eur. J. Pharmacol. 2010, 643, 232–238. [Google Scholar] [CrossRef]
  51. Xia, N.; Daiber, A.; Förstermann, U.; Li, H. Antioxidant Effects of Resveratrol in the Cardiovascular System. Br. J. Pharmacol. 2017, 174, 1633–1646. [Google Scholar] [CrossRef]
  52. Guo, N.-H.; Fu, X.; Zi, F.-M.; Song, Y.; Wang, S.; Cheng, J. The Potential Therapeutic Benefit of Resveratrol on Th17/Treg Imbalance in Immune Thrombocytopenic Purpura. Int. Immunopharmacol. 2019, 73, 181–192. [Google Scholar] [CrossRef]
  53. Moore, A.; Beidler, J.; Hong, M. Resveratrol and Depression in Animal Models: A Systematic Review of the Biological Mechanisms. Molecules 2018, 23, 2197. [Google Scholar] [CrossRef]
  54. Li, X.-L.; Zhou, A.-G.; Zhang, L.; Chen, W.-J. Antioxidant Status and Immune Activity of Glycyrrhizin in Allergic Rhinitis Mice. Int. J. Mol. Sci. 2011, 12, 905–916. [Google Scholar] [CrossRef]
  55. Chen, X.; Fang, D.; Li, L.; Chen, L.; Li, Q.; Gong, F.; Fang, M. Glycyrrhizin Ameliorates Experimental Colitis through Attenuating Interleukin-17-Producing T Cell Responses via Regulating Antigen-Presenting Cells. Immunol. Res. 2017, 65, 666–680. [Google Scholar] [CrossRef]
  56. Murck, H.; Lehr, L.; Hahn, J.; Braunisch, M.C.; Jezova, D.; Zavorotnyy, M. Adjunct Therapy with Glycyrrhiza Glabra Rapidly Improves Outcome in Depression—A Pilot Study to Support 11-Beta-Hydroxysteroid Dehydrogenase Type 2 Inhibition as a New Target. Front. Psychiatry 2020, 11, 605949. [Google Scholar] [CrossRef]
  57. Albogami, S.; Hassan, A.; Ahmed, N.; Alnefaie, A.; Alattas, A.; Alquthami, L.; Alharbi, A. Evaluation of the Effective Dose of Amygdalin for the Improvement of Antioxidant Gene Expression and Suppression of Oxidative Damage in Mice. PeerJ 2020, 8, e9232. [Google Scholar] [CrossRef] [PubMed]
  58. Gago-López, N.; Lagunas Arnal, C.; Perez, J.J.; Wagner, E.F. Topical Application of an Amygdalin Analogue Reduces Inflammation and Keratinocyte Proliferation in a Psoriasis Mouse Model. Exp. Dermatol. 2021, 30, 1662–1674. [Google Scholar] [CrossRef] [PubMed]
  59. Chen, K.-C.; Chen, C.-R.; Chen, C.-Y.; Tzou, K.-Y.; Peng, C.-C.; Peng, R.Y. Bicalutamide Elicits Renal Damage by Causing Mitochondrial Dysfunction via ROS Damage and Upregulation of HIF-1. Int. J. Mol. Sci. 2020, 21, 3400. [Google Scholar] [CrossRef] [PubMed]
  60. Sun, L.; Cao, B.; Wang, J.; Srisa-an, W.; Yu, P.S.; Leow, A.D.; Checkoway, S. Kollector: Detecting Fraudulent Activities on Mobile Devices Using Deep Learning. IEEE Trans. Mob. Comput. 2021, 20, 1465–1476. [Google Scholar] [CrossRef]
  61. Jagannatha, A.N.; Yu, H. Structured Prediction Models for RNN Based Sequence Labeling in Clinical Text. Proc. Conf. Empir. Methods Nat. Lang. 2016, 2016, 856–865. [Google Scholar] [CrossRef]
  62. Leevy, J.L.; Khoshgoftaar, T.M.; Villanustre, F. Survey on RNN and CRF Models for De-Identification of Medical Free Text. J. Big Data 2020, 7, 73. [Google Scholar] [CrossRef]
  63. Zhu, H.; He, C.; Fang, Y.; Ge, B.; Xing, M.; Xiao, W. Patent Automatic Classification Based on Symmetric Hierarchical Convolution Neural Network. Symmetry 2020, 12, 186. [Google Scholar] [CrossRef]
  64. Beurel, E.; Lowell, J.A. Th17 Cells in Depression. Brain. Behav. Immun. 2018, 69, 28–34. [Google Scholar] [CrossRef]
  65. Edwar-Mickael, M.; Kubick, N. CD4+Tregs May Be Essential for Solving Astrocyte Glial Scar Deadlock. Neural Regen. Res. 2021, 16, 2563. [Google Scholar] [CrossRef]
  66. Kubick, N.; Henckell Flournoy, P.C.; Klimovich, P.; Manda, G.; Mickael, M.E. What Has Single-Cell RNA Sequencing Revealed about Microglial Neuroimmunology? Immun. Inflamm. Dis. 2020, 8, 825–839. [Google Scholar] [CrossRef]
  67. Kubick, N.; Flournoy, P.C.H.; Enciu, A.-M.; Manda, G.; Mickael, M.-E. Drugs Modulating CD4+ T Cells Blood–Brain Barrier Interaction in Alzheimer’s Disease. Pharmaceutics 2020, 12, 880. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A detailed workflow of our proposal. There are four phases in the workflow. The user first enters a disease or a primary pathway to be investigated. The search step is done to select relevant PubMed articles. Then each accepted article is sorted and parsed. The first network converts each sentence in each PDF to a single embedding matrix. The second neural network sorts each sentence in each PDF based on its relevance to the query. The third neural network extracts drug names from each relevant sentence.
Figure 1. A detailed workflow of our proposal. There are four phases in the workflow. The user first enters a disease or a primary pathway to be investigated. The search step is done to select relevant PubMed articles. Then each accepted article is sorted and parsed. The first network converts each sentence in each PDF to a single embedding matrix. The second neural network sorts each sentence in each PDF based on its relevance to the query. The third neural network extracts drug names from each relevant sentence.
Molecules 27 06453 g001
Figure 2. Workflow description. Phase (I&II); After parsing PDFs into single sentences, each sentence is embedded into A 512 × 1 vector (here reshaped into 32 × 16) using the embedding (E) network. Phase (III); the relevance (R) network predicts the answer to the embedded question. The form of the predicted answer is a 512 vector. The distance tween the predicted answer and the embedding of each sentence is calculated. The most relevant sentences are estimated to possess the shortest distance from the embedded question matrix. After that, in phase (IV), the extraction (ER) network predicts the compound embedding for each relevant sentence. The distance between each word embedding and the predicted answer embedding is calculated. The word with the shortest distance between its matrix embedding and the predicted compound embedding is estimated to represent the compound name.
Figure 2. Workflow description. Phase (I&II); After parsing PDFs into single sentences, each sentence is embedded into A 512 × 1 vector (here reshaped into 32 × 16) using the embedding (E) network. Phase (III); the relevance (R) network predicts the answer to the embedded question. The form of the predicted answer is a 512 vector. The distance tween the predicted answer and the embedding of each sentence is calculated. The most relevant sentences are estimated to possess the shortest distance from the embedded question matrix. After that, in phase (IV), the extraction (ER) network predicts the compound embedding for each relevant sentence. The distance between each word embedding and the predicted answer embedding is calculated. The word with the shortest distance between its matrix embedding and the predicted compound embedding is estimated to represent the compound name.
Molecules 27 06453 g002
Figure 3. Loss and mean absolute error functions of the extraction network in Adera2.0 workflow. (A) Models A, E, F, G, and H achieve the lowest loss. Models B, C, D and H show the highest loss. Green is utilized for the loss function, while red is used for the validation loss calculations. (B) Model A achieves the lowest MAE value. The mean absolute error is in blue, while the validation of the MAE curve is in green. It is worth noting that the MAE validation values are lower than their respective MAE training values. The MAE is calculated over any given dataset. The dataset is split randomly into training and validation datasets with a ratio of 0.61:0.39. Thus the size of the validation dataset is smaller, resulting in slightly lower MAE values compared to the MAE values resulting from the training dataset.
Figure 3. Loss and mean absolute error functions of the extraction network in Adera2.0 workflow. (A) Models A, E, F, G, and H achieve the lowest loss. Models B, C, D and H show the highest loss. Green is utilized for the loss function, while red is used for the validation loss calculations. (B) Model A achieves the lowest MAE value. The mean absolute error is in blue, while the validation of the MAE curve is in green. It is worth noting that the MAE validation values are lower than their respective MAE training values. The MAE is calculated over any given dataset. The dataset is split randomly into training and validation datasets with a ratio of 0.61:0.39. Thus the size of the validation dataset is smaller, resulting in slightly lower MAE values compared to the MAE values resulting from the training dataset.
Molecules 27 06453 g003
Figure 4. Sensitivity analysis for Adera2.0 extraction networks. (A) Normalized sensitivity for different architectures of the extraction network (B) Boxplot of True and negative prediction for various architectures of the extraction network. ** represents p-value of <0.001, *** represents p-value of <0.0001, while ns stand for non-significant.
Figure 4. Sensitivity analysis for Adera2.0 extraction networks. (A) Normalized sensitivity for different architectures of the extraction network (B) Boxplot of True and negative prediction for various architectures of the extraction network. ** represents p-value of <0.001, *** represents p-value of <0.0001, while ns stand for non-significant.
Molecules 27 06453 g004
Figure 5. Sensitivity analysis of the investigated models based on their ability to identify known drugs. To validate the performance of our models, we compared their performance using three types of questions. (A,B) We found that models G and H achieve the highest sensitivity on generic questions. (A,B) Model F could be more suitable for the pathway question–answer category. (C) Our K-means clustering analysis indicates that the models separate based on their sensitivity into a cluster containing the B, C, and D models and another cluster containing the models’; A, E, F, H, and G.
Figure 5. Sensitivity analysis of the investigated models based on their ability to identify known drugs. To validate the performance of our models, we compared their performance using three types of questions. (A,B) We found that models G and H achieve the highest sensitivity on generic questions. (A,B) Model F could be more suitable for the pathway question–answer category. (C) Our K-means clustering analysis indicates that the models separate based on their sensitivity into a cluster containing the B, C, and D models and another cluster containing the models’; A, E, F, H, and G.
Molecules 27 06453 g005
Figure 6. BBB infiltration predictions of bicalutamide.
Figure 6. BBB infiltration predictions of bicalutamide.
Molecules 27 06453 g006
Table 3. Comparison between Adera1.0 and Adera2.0.
Table 3. Comparison between Adera1.0 and Adera2.0.
DescriptionAdera1.0Adera2.0
Function Question–answer system (QA) to find the nearest answer to a given query
Input to the systemUser input (query question)
Overall number of networks23
Networks usedEmbedding and relevance Embedding, relevance, and extraction
Embedding network functionGenerate embedding for query question and each sentence in the relevant PDFs
Embedding network architectureDAN (deep averaging network) followed by feedforward autoencoder
Training of the networkWikipedia, web news, and question and answer websites
Pearson correlation performance0.76
Output of the embedding network512 × 1 embedding matrix
Relevance network functionDetermine the relevance of the answers
Relevance network architecture Convolution network
Training of the networkSmall in-house datasetSciQ
Best loss function performance0.0018–0.002
Output of the relevant network Embedding matrices sorted based on relevance to the answer
Extraction of compounds from sentences functionFunctionality does not existTo extract compound names from the relevant sentence
Relevance network architectureeight different architectures
Best loss function performance0.0013
Input Sorted sentences
Training of the networkDataset Extraction (1.0)
OutputList of compounds
Table 4. Summary of the compounds identified using our software.
Table 4. Summary of the compounds identified using our software.
CompoundMolecule StructureBiological Activity
Acebutolol (O=C(N)Cc1ccc(cc1)OCC(O)CNC(C)C)Acebutolol is a selective 1-receptor antagonist that lowers blood pressure and heart rate.
beta-D-fructofuranoseC(C1C(C(C(O1)(CO)O)O)O)Obeta-D-fructofuranose plays a key role in glycolysis. However, its exact function is still unknown. Currently, it is in a clinical trial.
(clinicaltrials.gov Identifier: NCT05207488)
Elatin (flavonoid)CC1C(C(C(C(O1)C2=C3C(=C(C(=C2O)C4C(C(C(C(O4)CO)O)O)O)O)C(=O)C=C(O3)C5=CC(=C(C=C5)O)O)O)O)O Elatin was reported to be antioxidant, anti-cancer, anti-microbial, neuroprotective, and anti-inflammation.
CoumarinO=C1C=Cc2ccccc2O1Coumarin is an anticoagulant used to treat deep vein thrombosis and pulmonary embolism.
ResveratrolOc1ccc(cc1)C=Cc1cc(O)cc(c1)OResveratrol was reported to reduce markers of inflammation.
Glycyrrhetinic acidCC1(C2CCC3(C(C2(CCC1O)C)C(=O)C=C4C3(CCC5(C4CC(CC5)(C)C(=O)O)C)C)C)CGlycyrrhetinic acid was shown to inhibit 11β-hydroxysteroid dehydrogenase, thus inhibiting the conversion of cortisol.
GlycyrrhizinCC1(C2CCC3(C(C2(CCC1OC4C(C(C(C(O4)C(=O)O)O)O)OC5C(C(C(C(O5)C(=O)O)O)O)O)C)C(=O)C=C6C3(CCC7(C6CC(CC7)(C)C(=O)O)C)C)C)C.N.N. N Glycyrrhizin was approved for use as a flavor and aroma in manufactured foods.
AmygdalinC1=CC=C(C=C1)C(C#N)OC2C(C(C(C(O2)COC3C(C(C(C(O3)CO)O)O)O)O)O)OAmygdalin was reported to suppress oxidative damage.
PorphyrinC1=CC2=CC3=CC=C(N3)C=C4C=CC(=N4)C=C5C=CC(=N5)C=C1N2 Porphyrin constitutes a part of the heme in the hemoglobin and myoglobin.
BicalutamideCC(CS(=O)(=O)C1=CC=C(C=C1)F)(C(=O)NC2=CC(=C(C=C2)C#N)C(F)(F)F)OThe antiandrogen drug bicalutamide is primarily used to treat prostate cancer.
Table 5. Pathogenicity constraints predication for the identified compounds.
Table 5. Pathogenicity constraints predication for the identified compounds.
CompoundMutagenicityTumorgenicityIrritantReproductive Effective
AcebutololNoNoNoNo
beta-D-fructofuranoseNoNoNoYes
Elatin (flavonoid)YesYesNoNo
CoumarinYesYesNoYes
resveratrolNoNoNoYes
Glycyrrhetinic acidNoNoNoNo
glycyrrhizinNoNoNoNo
AmygdalinYesNoYesYes
PorphyrinNoNoNoNo
BicalutamideNoNoNoNo
Table 6. Comparison of repurposed drugs based on their physical and chemical characteristics.
Table 6. Comparison of repurposed drugs based on their physical and chemical characteristics.
CompoundClogSolubilityMol WeightTpsaDrug-LikenessDrug ScoreBBBViolations
Acebutolol1.7−3.5336.087.664.910.83BBB-0
beta-D-fructofuranose−2.70.38180110−2.560.32BBB-0
Elatin (flavonoid)−1.91−1.89594267.20.420.14BBB-3
Coumarin1.5−2.3714626−1.830.12BBB+0
resveratrol2.83−2.86228.060.69−3.250.27BBB+0
Glycyrrhetinic acid5.36−5.7847074.6−2.360.2BBB-1
Glycyrrhizin0.39−5.14822267.0−4.290.19BBB-3
Amygdalin−3.08−1.12457202.3−8.70.09BBB-2
Porphyrin2.05−4.34310.052.540.970.67BBB+1
Bicalutamide2.14−5.08430115−110.24BBB+0
Table 7. Repurposed components arranged based on their capabilities.
Table 7. Repurposed components arranged based on their capabilities.
Repurposed ComponentsAntioxidantsTh17 InteractionKnown Effects on Depression
AcebutololYes [45]UnknownUnknown
beta-D-fructofuranoseUnknownUnknownUnknown
Elatin (flavonoid)Yes [46]Yes [46]Yes [47]
CoumarinYes [48]Yes [49]Yes [50]
ResveratrolYes [51]Yes [52]Yes [53]
Glycyrrhetinic acidYes [54]Yes [55]Yes [56]
GlycyrrhizinYes [54]Yes [55]Yes [56]
AmygdalinYes [57]Yes [58]unknown
PorphyrinYes [31]Yes [44]Yes [31] (enhance prognosis)
BicalutamideNo [59]Yes [42]Yes (worsen prognosis)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lazarczyk, M.; Duda, K.; Mickael, M.E.; AK, O.; Paszkiewicz, J.; Kowalczyk, A.; Horbańczuk, J.O.; Sacharczuk, M. Adera2.0: A Drug Repurposing Workflow for Neuroimmunological Investigations Using Neural Networks. Molecules 2022, 27, 6453. https://doi.org/10.3390/molecules27196453

AMA Style

Lazarczyk M, Duda K, Mickael ME, AK O, Paszkiewicz J, Kowalczyk A, Horbańczuk JO, Sacharczuk M. Adera2.0: A Drug Repurposing Workflow for Neuroimmunological Investigations Using Neural Networks. Molecules. 2022; 27(19):6453. https://doi.org/10.3390/molecules27196453

Chicago/Turabian Style

Lazarczyk, Marzena, Kamila Duda, Michel Edwar Mickael, Onurhan AK, Justyna Paszkiewicz, Agnieszka Kowalczyk, Jarosław Olav Horbańczuk, and Mariusz Sacharczuk. 2022. "Adera2.0: A Drug Repurposing Workflow for Neuroimmunological Investigations Using Neural Networks" Molecules 27, no. 19: 6453. https://doi.org/10.3390/molecules27196453

APA Style

Lazarczyk, M., Duda, K., Mickael, M. E., AK, O., Paszkiewicz, J., Kowalczyk, A., Horbańczuk, J. O., & Sacharczuk, M. (2022). Adera2.0: A Drug Repurposing Workflow for Neuroimmunological Investigations Using Neural Networks. Molecules, 27(19), 6453. https://doi.org/10.3390/molecules27196453

Article Metrics

Back to TopTop