Next Article in Journal
Deposition of Lead Phosphate by Lead-Tolerant Bacteria Isolated from Fresh Water near an Abandoned Mine
Next Article in Special Issue
Phenotypic and Transcriptional Changes of Pulmonary Immune Responses in Dogs Following Canine Distemper Virus Infection
Previous Article in Journal
Further Evidence of the Melatonin Calmodulin Interaction: Effect on CaMKII Activity
Previous Article in Special Issue
Time-Dependent Molecular Motifs of Pulmonary Fibrogenesis in COVID-19
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Augmentation of Transcriptomic Data for Improved Classification of Patients with Respiratory Diseases of Viral Origin

1
Institute for Animal Breeding and Genetics, University of Veterinary Medicine Hannover, Buenteweg 17p, 30559 Hannover, Germany
2
Department of Pathology, University of Veterinary Medicine Hannover, Buenteweg 17, 30559 Hannover, Germany
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2022, 23(5), 2481; https://doi.org/10.3390/ijms23052481
Submission received: 25 November 2021 / Revised: 17 February 2022 / Accepted: 21 February 2022 / Published: 24 February 2022

Abstract

:
To better understand the molecular basis of respiratory diseases of viral origin, high-throughput gene-expression data are frequently taken by means of DNA microarray or RNA-seq technology. Such data can also be useful to classify infected individuals by molecular signatures in the form of machine-learning models with genes as predictor variables. Early diagnosis of patients by molecular signatures could also contribute to better treatments. An approach that has rarely been considered for machine-learning models in the context of transcriptomics is data augmentation. For other data types it has been shown that augmentation can improve classification accuracy and prevent overfitting. Here, we compare three strategies for data augmentation of DNA microarray and RNA-seq data from two selected studies on respiratory diseases of viral origin. The first study involves samples of patients with either viral or bacterial origin of the respiratory disease, the second study involves patients with either SARS-CoV-2 or another respiratory virus as disease origin. Specifically, we reanalyze these public datasets to study whether patient classification by transcriptomic signatures can be improved when adding artificial data for training of the machine-learning models. Our comparison reveals that augmentation of transcriptomic data can improve the classification accuracy and that fewer genes are necessary as explanatory variables in the final models. We also report genes from our signatures that overlap with signatures presented in the original publications of our example data. Due to strict selection criteria, the molecular role of these genes in the context of respiratory infectious diseases is underlined.

1. Introduction

Viral diseases of the respiratory system are frequently studied on the molecular level of the transcriptome [1,2,3]. This can help to identify genes that are involved in the immune response or pathogenesis [4] but also to use these genes as predictors in machine-learning (ML) models to classify patients based on molecular data. Finally, ML models can help to optimize treatment strategies for patients. Facilitated by the technological progress of DNA microarrays and high-throughput RNA-sequencing (RNA-seq), molecular signatures in high-dimensional gene-expression data have been studied for over 20 years, employing classical ML techniques such as support vector machines (SVM) [5], linear discriminant analysis (LDA) [6], random forest (RF) [7], artificial neural networks (ANN) [8], or more specialized methods [9]. The term ‘molecular signature’ usually refers to a set of genes that constitute the predictor variables in an ML model. It has been shown that molecular signatures from ML models can provide ‘valuable insights into cell biology and mechanisms of human disease’ [10]. Recently, di Julio et al. [11] also studied ’transfer transcriptomic signatures’ which could have predictive value for infectious diseases across species, based on the fact that transcriptomic changes are among the early responses to infection. Such transfer signatures could therefore also be valuable to better understand the molecular processes across species. A prior step before fitting an ML model with transcriptomic data is to select a set of differentially expressed genes between two study groups (e.g., infected versus uninfected)—typically by means of statistical testing—before investigating this ‘molecular signature’ for its diagnostic or predictive power in ML models. Although differential expression analysis is usually performed on a single dataset, diagnostic or predictive signatures in an ML model are usually evaluated in independent datasets or at least by means of cross-validation. One can therefore argue that the high performance of a molecular signature in an ML model underlines the role of these genes in the molecular processes of the studied disease. Molecular processes involved in viral respiratory diseases include virus–host cell interactions, signaling pathways of the innate and adaptive immune response, as well as viral evasion strategies and tissue-repair mechanisms, which are orchestrated by the expression of specific genes. A better understanding of these pathways and the role of the respective genes is essential for the discovery of new therapeutic targets or potent prevention strategies [12,13,14].
In this work, we reanalyze two transcriptomic datasets from study contexts of respiratory, viral-caused diseases to demonstrate how data augmentation can help to improve the performance of ANNs based on high-dimensional gene-expression data. Gene-expression data for the two selected studies are freely available from the gene-expression omnibus (GEO) archive [15]. One of these two studies involves gene-expression data taken from human blood samples by means of DNA microarrays. All patients suffered from an acute respiratory illness (ARI), but with either viral or bacterial origin [16]. In the second study, gene expression was taken by means of RNA-seq and involved human nasopharyngeal swab samples from a diagnosed SARS-CoV-2 group and another group of patients with respiratory symptoms caused by different viruses such as influenza A and B, adenovirus (B, E), parainfluenza 1–4, respiratory syncytial virus A and B, human metapneumovirus, human rhinovirus, coronavirus (229E, OC43, NL63, HKU1), coxsackie/echo virus and bocavirus [8]. The authors of the original publications of the two datasets also fitted ML models to assess the diagnostic performance of selected molecular signatures, but without a step of data augmentation. Given that ML models trained with augmented versions of the original data show higher classification accuracies than without data augmentation, there would be more evidence for the important molecular role of the predictor genes involved in these models. Furthermore, ML models based on augmented data tend to be less overfitted. When being applied to draw treatment decisions, such models would result in fewer misclassifications and therefore in an improved treatment of patients.
To obtain robust ML models, a high quantity of annotated training data is usually required [17]. However, it is often difficult and expensive to obtain enough labeled samples [18], especially in the context of transcriptomics. High-dimensional gene-expression data represent expression levels for thousands of genes in a rather small number of biological samples. Thus, the dimension d of the data is much larger than the sample size n, i.e., d > > n . Furthermore, with a small sample size, it is difficult to achieve good generalization performance [19], i.e., the training data are learned well, but the trained ML model cannot easily be transferred to new independent data. In the past ten years, as ANNs have become more frequently employed for biomedical classification problems, it was found that augmented datasets can improve the classification accuracy of the trained models [20], and can also help to reduce the risk of overfitting [21]. Data augmentation means to artificially increase the sample size by modifying instances of the original data. Here, we compare three different strategies for data augmentation of high-dimensional gene-expression data. In the past few years, many algorithms based on ANNs have been developed, including the generative adversarial network (GAN), which represents a new approach to data augmentation [22]. Recent work has shown that GANs can be used successfully, among image or text generation, to add synthetic samples to insufficient training data and thus improve classification accuracy. However, it mainly addresses the problem of class imbalance [19]. Studies on enlarging relatively small datasets for classification (such as small-sized transcriptomics data) are rare [19,23]. For fitting the ANN models on the data of the two respiratory studies, we compare two variants of GAN algorithms as well as a less complex approach of mixing observations for data augmentation.
In the next section, we detail the results of the comparison of the three data augmentation techniques in the two respiratory datasets. We also compare the sets of genes in the molecular signatures from our ANN models with the signatures from the original publications. The results section is followed by a discussion about the computational and biological findings. The datasets used and the methods for data augmentation and classification are described in the last chapter.

2. Results

Using the two gene-expression datasets on respiratory infections, we investigated the applicability of data augmentation methods to both continuous microarray data and discrete RNA-seq count data. More specifically, we compared three different augmentation approaches: (a) a simple approach that is based on mixing observations, hereafter referred to as the mixed weighted observations (MWO) approach, as described in Section 4.6, (b) a standard GAN as described in Section 4.7, and (c) a Wasserstein GAN with gradient penalty (WGAN-GP) as described in Section 4.8. With these methods, the training datasets were expanded in such a way that initially twice as many artificial samples were added for each class as original samples were available. Evaluating the classification performance using these augmentation methods, we wanted to answer the following questions: (i) Does data augmentation increase the classification accuracy? (ii) Comparing the aforementioned methods (MWO, GAN and WGAN-GP), which one performs best for augmenting high-dimensional expression data? (iii) Can we identify genes found in the original studies, whose molecular importance is emphasized by our approach?

2.1. Signature Selection and Classifier Performance without Data Augmentation

We first describe the incremental signature selection for both datasets via ML without data augmentation. By increasing the number of genes, their influence on the performance of the classifier was examined. As mentioned in the introduction ANNs were used as classification model, more specifically we fitted multilayer perceptrons (MLP) with two hidden layers with 150 and 50 neurons each and rectified linear unit (ReLU) activation. The number of input neurons corresponded to the number of genes that were used as input signature. The output layer was one single neuron with linear activation. The ANN architecture is described in more detail in the methods Section 4.4.
A 10-fold cross-validation was used to estimate the performance of each classification model, i.e., the original dataset D was split into K = 10 subsets, and for each k ϵ { 1 , 2 , , K } , the ANN was trained on the training subset D \ D k and afterwards used to predict the labels of the test subset D k .
To identify the minimum number of genes as input features required for optimal ANN classification performance, an incremental feature selection was carried out for each cross-validation training subset D \ D k . This subset was used to test the genes for differential expression and genes were ranked according to their p-values in ascending order, i.e., the gene with the lowest p-value being the first important input feature, and so on. This approach assumes that the genes with the most evidence for differential expression are also important for the discrimination of phenotypes. As a result, including an increasing number of the top ranked genes, i.e., the first 50, 100, 150, …, 750 most significant genes, as input features for training and testing the classification ANN, led to different classification performances. Figure 1 shows the distribution of the Brier score for models fit on the two respiratory datasets as obtained from the cross-validation and for different numbers of input genes. Including an increasing number of genes as input features tended to result in better performances of the ANNs, i.e., the median Brier score tended to decrease when increasing the number of top genes as input features from 50 to 750 (from 0.16 to 0.10 for the microarray data and from 0.24 to 0.22 for the RNA-seq experiment, respectively).
Performances of the ANN models for both datasets when using data augmentation are provided in the following subsections.

2.2. Classification Performance of Molecular Signatures Selected in the Microarray Ari Study

Here, we describe the performance of ANN models trained with genes selected in the microarray study that aimed to distinguish between viral and bacterial origin of ARIs. Figure 2 shows the accuracies (left plot) and Brier scores (right plot) for the ANNs trained on different augmented datasets compared to ANNs trained without data augmentation. Numbers are also provided in Table 1 and Table 2 as well as in the Supplementary Materials. Again, the number of top genes selected as input signature was varied. Maximum accuracy, sensitivity, specificity and minimum Brier score are highlighted in each table, so that one can see the minimum number of top genes required for an optimized classification results. Looking at the Brier scores, the ANNs trained with augmented data performed clearly better than those trained without data augmentation, e.g., the lowest Brier score is obtained when using the WGAN-GP augmentation method and 450 genes. Regarding the accuracies, the WGAN-GP approach for augmenting the original training data showed the most accurate classification results by achieving the maximum mean accuracy of 0.91 when using 700 genes as input signature, followed by the MWO approach. Additionally, in terms of the Brier score, Figure 2 suggests that an increasing number of genes did not affect classification performance as much when the training data were augmented as it did with the non-augmented training data. Concerning the obtained accuracies, it seemed that the standard GAN was not performing better than the no-augmentation approach, whereas the Brier scores for the GAN approach were consistently lower than for classification using only the not-augmented original training data.
Since the above detailed differences between augmented and non-augmented scenarios describe tendencies, significant difference can also be detected using the duality between statistical test and confidence interval, as we have shown 95%—confidence intervals (CI) in all figures and tables. According to the duality between statistical test and CI, a significant difference can be inferred with a significance level of 5% when the 95%—CI for two scenarios do not overlap. Using this principle, one can see in the figures and tables clearly that data augmentation yields significantly better results in some of our simulated scenarios. Regarding the different influence of the number of genes being included in the model, from Figure 2 one can infer a significant difference between the Brier score when comparing the performances of the no-augmentation approach with either 50 or 150 top genes, whereas the CIs for the classifiers trained with augmented data do overlap.
As the previously described results of the 2-fold augmentation already led to promising results, we tested further augmentation factors, namely 1-fold, 3-fold and 5-fold augmentation. Regarding the WGAN-GP approach, it could be shown that the accuracy and the Brier score improved with increasing augmentation; an improvement in the Brier score could be observed for the MWO approach. In contrast, when increasing the number of GAN-generated samples for training, the trained classification network performed worse on the test data. The detailed results are part of the Supplementary Materials.
Besides studying the use of data augmentation for assessing the value of a molecular signature, we were also interested how well each augmentation method maintained the correlation structure among genes. When performing data augmentation, it is important that the distribution of the artificial data closely meets that of the original data. That applies, for example, to the correlation structure between genes, which is represented, for example, by the covariance matrix of the expression matrices. Since a covariance matrix between thousands of genes cannot be estimated unbiased, we randomly selected small subsets of genes for which we studied the covariance matrices of original and artificial data. Figure 3 shows graphical representations of covariance matrices for two sets of 50 randomly selected genes from the ARI study. Comparing the covariance structures of the original data and those of the different synthetic datasets suggests that the covariance structure is maintained by each of the artificially generated samples, whereby the covariance structure seems to be most precisely maintained when generating artificial data with the MWO approach or the WGAN-GP approach. The covariance matrices of the artificial samples generated with a GAN look more different from that calculated from the original data.
Although Figure 3 illustrates the results for two selected subsets of genes, Figure 4 shows the distribution of normalized root mean square errors and mean absolute differences for the different augmentation methods when comparing their covariance structure with the original covariance structure when 50 genes were randomly drawn, repeatedly 500 times. Comparing the three augmentation strategies, the covariance structures are best maintained when generating artificial samples with the MWO approach. Regarding the normalized root mean square error, the WGAN-GP approach shows less deviation than the GAN approach, but in terms of the mean absolute difference of eigenvalues, the WGAN-GP approach performs a bit worse than the GAN approach.
Although the training data in this example had a rather large sample size of 171, many microarray studies are much smaller. Therefore, we randomly selected a subset of 30 samples as training data, and compared again classifier performance with and without data augmentation. In this analysis, from a total of 190 samples, the 160 remaining samples were used as test samples. The whole procedure was repeated ten times to obtain confidence intervals for accuracy and the other performance measures. We compared the classification results for 2-fold and 5-fold augmentation with the classification results obtained without augmentation. Both 2-fold and 5-fold augmentation, i.e., having 30 original and 60 (150) artificial samples as training data, led to a significant improvement of the classification performance in terms of the Brier score, whereas considerable improvements of the accuracy could be observed only when including 50–200 genes into the classifier. The results for the different augmentation methods and different numbers of genes for the 2-fold and 5-fold augmentation of randomly chosen subsets are shown in the Supplementary Materials.

2.3. Classification Performance of Molecular Signatures Selected in the RNA-Seq SARS-CoV-2 Study

Here, we describe the performance of classification models trained with genes selected in the RNA-seq study that aimed to distinguish between SARS-CoV-2 and other viral origins of respiratory infections. Figure 5 shows the classification performances for the different augmentation strategies for the RNA-seq data. Numbers are also provided in Table 3 and Table 4 as well as in the Supplementary Materials. The accuracies (left plot) and Brier scores (right plot) for the different data augmentation methods were compared to classification without augmentation and for various numbers of genes being included as input features. Again, maximum accuracy, sensitivity, specificity and minimum Brier score are highlighted in each table, so that one can see the minimum number of top genes required for optimized classification results. E.g., in the RNA-seq example, when augmenting the original sample size by a factor of 2 using the WGAN-GP approach, the maximum accuracy of 0.77 is reached with 300 genes. However, when looking at the Brier score, 400 genes would be necessary to reach the minimum of 0.17. Regarding the Brier scores, the classification with data augmentation using the MWO or WGAN-GP approach always leads to better results than without augmentation of the training data, regardless of the number of input genes. Including the accuracy values, the WGAN-GP approach for augmenting the original training data shows the most accurate classification results, followed by the MWO approach. Regarding the obtained accuracies, it seems that the standard GAN is not performing better than the no-augmentation approach, whereas the Brier scores for the GAN approach are consistently lower than for classification using only not-augmented training data.
Again, the duality-principle between statistical test and CIs can be used to show that augmentation yields a significantly better performance of a classifier than training only with original data or between a particular number of top genes included into a classifier. One can for example infer a significant difference between the Brier score in Figure 5, when comparing the performance of the WGAN approach with either 50 or 150 of the top genes. In particular, there is no overlap between CIs of the Brier scores for classifiers trained with augmented and non-augmented data for most numbers of top genes.
As for the microarray study, also 1-fold, 3-fold and 5-fold augmentation scenarios were tested to investigate the quality of the augmented samples. With further augmentation, the results improved further for the MWO and the WGAN approach, whereas the results for the ANNs trained with GAN augmented training data were not consistently better and showed an increasing variety in the classification of the test data. Detailed results can be found in the Supplementary Materials.
Figure 6 shows again heatmaps illustrating the covariance matrices of original and artificially generated data for two sets of 50 randomly selected genes from the SARS-CoV-2 study. Comparing the covariance structures of the original data and the different synthetic data suggests that the covariance structure is mostly maintained when generating artificial data with the MWO approach. The heatmaps for the covariance matrices of the data generated with the WGAN-GP approach implies that the covariance structure is partially preserved. With the GAN approach, the covariance structure is obviously hardly preserved.
Figure 7 shows the distribution of log-transformed normalized root mean square errors and log-transformed mean absolute differences for the different augmentation methods when comparing their covariance structure with the original covariance structure when 50 genes were randomly drawn, repeatedly 500 times. In this comparison of the three augmentation strategies, the distance between the covariances of the original data and those of the artificial data generated with the MWO approach is the smallest. The corresponding deviations between the original and WGAN-GP-generated samples, on the other hand, are somewhat larger and the distance to the artificial samples that were generated with a standard GAN is the greatest.
We also studied in this RNA-seq example whether data augmentation would also be helpful with a smaller set of training data. From the 258 samples, we selected randomly a subset of 30 samples as training data, and compared again classifier performance with and without data augmentation. In this analysis, all remaining 228 samples were taken as test data. The whole procedure was repeated ten times to obtain confidence intervals for accuracy and the other performance measures. To augment the training data, we considered 2-fold augmentation and 5-fold augmentation of the small training subset. The results of the analysis support the previously shown results that augmentation improves classification performance, since here too the accuracy increased, and the Brier score decreased with increasing augmentation; for the WGAN-GP and the MWO approach more than for the GAN approach. Detailed results are part of the Supplementary Materials.

2.4. Molecular Signatures of Microarray and RNA-Seq Study, and Overlap with Originally Published Signatures

To select genes as molecular signatures from the microarray and RNA-seq studies, we counted in how many folds of the cross-validation a gene was selected. In each fold, the selection criterion for a gene was that the FDR-adjusted p-value was <0.05 and the absolute log fold change was >2. Using the number of cross-validation folds is a sharper criterion than statistical testing on the whole dataset, because multiple reduced datasets (i.e., with reduced power) are used and genes are validated in the left-out dataset.
In the microarray ARI study, 44 genes were selected in each of the 10 cross-validation runs. In the RNA-seq SARS-CoV-2 study, 88 genes were selected, accordingly. Of these 88 genes, 41 coded for small nuclear RNAs.
We compared the selected genes with those presented as signatures in the original publications of the two datasets ([8,16]). It must be said that the different ML models and different subsamples and groupings of the data were used in these publications. Hence, classifiers and selected signatures are not directly comparable. However, genes that are jointly selected in different models have more evidence of being involved in the molecular processes of the studied respiratory diseases. For the microarray study, the overlap of our signatures with the signatures presented in the originally published results was rather small, i.e., the overlap were three genes: LY6E, OASL, IFIT1. For the RNA-seq study, the overlap were 21 genes: ACAN, BATF3, BHLHA15, CACNA1F, CCL2, CCL8, CMKLR1, CST8, CXCL11, FCN1, GUK1, IFNL1, IRG1, OPRL1, OR2W5, SERPINB12, SLAMF7, TFEC, TNIP3, TRPV5, ZNF81.
Most of these genes encode for molecules that are potentially involved in the pathogenesis of viral respiratory disease, such as antiviral and interferon-associated proteins of the innate immune response (LY6E, OASL, IFIT1, IRG1, CMKLR1, FCN1), transcription factors regulating cell differentiation (BATF3), chemokines (CCL2, CCL8, CXCL11), cytokines (IFNL1), molecules of membrane receptor-ligand systems (OPRL1, OR2W5), and molecules involved in calcium pathways (BHLHA15, CACNA1F) [24,25,26,27,28,29].
Besides looking at individual genes, we also performed gene ontology (GO) enrichment analysis [30], again based on selected genes from the 10 cross-validation runs. GO terms related to biological processes (BP), molecular functions (MF) and cellular components (CC) were studied separately. Among the significant GO terms in the microarray example, only a few were related to infection or immune response, e.g., antigen processing (GO:0002483, GO:0019885). In contrast, the significant GO terms in the RNA-seq example included those related to immune response (GO:0006955), T-cell differentiation (GO:0002292), and virus or immune receptor activity (GO:0001618, GO:0140375). The detailed results of GO enrichment analysis are provided as Supplementary Tables. Further significant GO terms, which are related to infection are listed in the following:
  • Microarray example, biological processes: antigen processing and presentation of endogenous peptide antigen (GO:0002483); antigen processing and presentation of endogenous peptide antigen via MHC class I (GO:0019885); antigen processing and presentation (GO:0019882); antigen processing and presentation of peptide antigen via MHC class I (GO:0002474); regulation of dendritic cell differentiation (GO:2001198); dendritic cell differentiation (GO:0097028); macrophage cytokine production (GO:0010934); antigen processing and presentation of endogenous antigen (GO:0019883); phagocytosis, engulfment (GO:0006911); antigen processing and presentation of peptide antigen (GO:0048002)
  • Microarray example, cellular component: phagocytic vesicle (GO:0045335)
  • Microarray example, molecular function: MHC protein binding (GO:0042287); MHC class I protein binding (GO:0042288); MHC protein complex binding (GO:0023023)
  • RNA-seq example, biological processes: macrophage activation (GO:0042116); T-cell differentiation involved in immune response (GO:0002292); leukocyte activation involved in immune response (GO:0002366); regulation of macrophage activation (GO:0043030); CD4-positive, alpha-beta T-cell differentiation involved in immune response (GO:0002294); cell activation involved in immune response (GO:0002263); alpha-beta T-cell activation involved in immune response (GO:0002287); alpha-beta T-cell differentiation involved in immune response (GO:0002293); regulation of interleukin-1 production (GO:0032652); regulation of interleukin-1 beta production (GO:0032651); myeloid leukocyte activation (GO:0002274); regulation of cytokine production
    (GO:0001817); regulation of interleukin-17 production (GO:0032660); immune response (GO:0006955); T-cell activation involved in immune response (GO:0002286); response to molecule of bacterial origin (GO:0002237); cytokine production (GO:0001816); response to lipopolysaccharide (GO:0032496); cellular response to oxidative stress (GO:0034599); NIK/NF-kappaB signaling (GO:0038061); regulation of microglial cell activation (GO:1903978); T-helper cell differentiation (GO:0042093); microglial cell activation (GO:0001774); leukocyte activation involved in inflammatory response (GO:0002269); interleukin-1 production (GO:0032612); reactive oxygen species metabolic process (GO:0072593); leukocyte activation (GO:0045321); interleukin-1 beta production (GO:0032611); plasma membrane bounded cell projection assembly (GO:0120031); positive regulation of cytokine production (GO:0001819); interleukin-17 production (GO:0032620); mitochondrial translation (GO:0032543); cytokine-mediated signaling pathway (GO:0019221); CD4-positive, alpha-beta T-cell differentiation (GO:0043367); lymphocyte activation involved in immune response (GO:0002285); cell projection assembly (GO:0030031)
  • RNA-seq example, molecular function: antigen binding (GO:0003823); virus receptor activity (GO:0001618); immune receptor activity (GO:0140375); cytokine receptor binding (GO:0005126); double-stranded RNA binding (GO:0003725); cytokine receptor activity (GO:0004896); chemokine binding (GO:0019956)

3. Discussion

Early diagnosis of patients with respiratory diseases caused by viral infections is important to provide a good treatment. Since changes in the transcriptome are among the early responses upon infection, classifiers based on high-throughput gene-expression data would provide a reasonable means for diagnosis. Early diagnosis is important to allow rapid therapeutical intervention, preventing progression of tissue damage and respiratory distress, resulting in a better prognosis. Additionally, early diagnosis is essential for the reduction of viral shedding and the initiation of early disease control strategies in the environment of affected individuals, which limits further virus spread.
Several studies have been performed to study the differentiability of patients with respiratory diseases of viral origin, usually using standard ML models fitted to transcriptomics data. We selected two of such studies to demonstrate that ANNs in combination with a step of data augmentation, can provide improved classification accuracy. Although data augmentation has a long tradition in classifying samples based on other data types (e.g., images), it has rarely been considered for transcriptomic data. We selected two different kinds of transcriptomic data, i.e., one study based on DNA microarray data and one based on RNA-seq data. Although high-dimensional gene-expression data are distributed on a continuous scale as fluorescence values from DNA microarrays [31], discrete count data are produced when using RNA-sequencing (RNA-seq) instead [32]. In summary, our comparative analyses yielded that data augmentation can indeed improve the classification performance. The best performance was achieved by an advanced generative adversarial network (WGAN-GP approach) which is, however, computationally time-consuming. The second-best method was to simply mix individual observations (MWO approach) which is computationally much less demanding.
We mainly focused on accuracy and Brier score when presenting the results. A common metric for evaluating medical diagnoses is the F1-score, which is the harmonic mean of sensitivity (recall) and positive predictive value (precision) [33]. Looking at the F1-scores supported the previous results as it behaved similar to the findings for the accuracies. i.e., for the microarray dataset no significant improvement could be observed when using all the training data for training the classifier whereas when using only a smaller subset for training at least for the first 200 most differentially expressed genes as input signature the classifiers trained with augmented data outperformed that without augmentation. For the RNA-seq data, further augmentation, i.e., increase of the augmentation factor, improved the classification both in the case of using all training data for training and the case where only a smaller subset of the data was used for training. F1-scores are also added to the Supplementary Tables and Figures.
Furthermore, our simulations have shown that the amount of augmentation has an impact on the classifier performance. In principle, the larger the amount of augmentation, the larger was the observed performance. In some cases, already with a 1-fold augmentation, a significant improvement was observed. However, a significant improvement might not necessarily mean a clinical relevant improvement. With a 5-fold augmentation in the RNA-seq example, one could obtain a significant improvement of the accuracy by 20%. This would mean a clearly larger number of correct classifications, and would be important if such a classification had an impact on therapy decisions. Therefore, it is not reasonable here to make a general statement regarding the minimum required amount of augmentation for classifiers with high-dimensional gene-expression data. Instead, we recommend checking the performance individually when other scientist use augmentation for such data.
In this manuscript, we have evaluated the use of data augmentation in the context of ANNs. Moreover, we did some additional evaluations of transcriptomic data augmentation with other classifier methods, specifically SVMs, LDA and RF, but could not see the same amount of improvement. Regarding LDA, it might be that additional observations generated by the studied approaches do not much change the estimates for group means and covariance matrices. For SVMs, additional observations generated by the studied approaches may not change much the shape of the separating hyperplane. It may be that other augmentation methods could work with SVMs, LDA and RF. Therefore, more research will be necessary to understand how augmentation of transcriptomics data can be performed in combination with different classifier models.
Independent from the question, whether augmented data can improve the classifier performance, we have compared covariance matrices of randomly selected gene sets to see, whether artificial data would also be useful for computer simulations, where researchers want to have data that reflects the correlation between genes. We found that the difference between the covariance matrix of original and artificial data was smallest for the MWO approach but larger for the GAN and WGAN-GP approaches. Although we observed these significant differences, the general shape of the covariance structure seems to be maintained as can be seen in Figure 3 and Figure 6. A reason for the observed differences with the GAN and WGAN approaches may be that with these approaches random data vectors are generated as new observations without taking the correlation structure between genes into account. For the MWO, the covariance structure is changed through the different weights with which the original columns in the gene-expression matrix are mixed. In summary, the artificial data may be useful for simulation purposes since the general shape of the covariance structure remains. Furthermore, if the covariance structure of artificial and original data was the same, the artificial data might have no added value for the classifier training. It should be added that the differences for the RNA-seq data appear much larger than the differences observed for the microarray data. This is due to the different scale of both data types.
Our classification results are not comparable with that of the ML models presented in the original publications of the two example datasets. First, we did use different groupings of samples, and second, we used other classifier models. In general, it is difficult to reimplement exactly ML models from other publications. However, our basic goal was to show that data augmentation in combination with ANNs can improve classification performance in contrast to ANNs fitted on the original data only, and that is what we have demonstrated here.
Knowledge about changes in the transcriptome during a disease can significantly contribute to a better understanding of the molecular processes of the pathogenesis and therapy response. Thus, it can be helpful to know which genes are significantly up- or down-regulated, and which GO terms are involved. We found that some of the significantly enriched GO terms have been mentioned in the context of other infectious diseases. In the microarray example, we selected the biological process GO:0002483 which was also found enriched in a study on juvenile dermatomyositis [34], which is supposed to be caused by viruses and can involve inflammation of the lung. The cellular component of phagocytic vesicles (GO:0045335) was also mentioned in the context of SARS-CoV-2 infection and pulmonary fibrosis [35]. Additionally, the enriched molecular function GO:0042287, responsible for MHC protein binding was found in a recent SARS-CoV-2 study [36], based on an enrichment analysis in 76 selected neighbor proteins of the SARS-CoV-2 spike protein. In the EBI QuickGO database the term GO:0042287 is further described as ‘Binding to a major histocompatibility complex molecule; a set of molecules displayed on cell surfaces that are responsible for lymphocyte recognition and antigen presentation’. In the RNA-seq example, the selected biological process ‘macrophage activation’ (GO:0042116) was also mentioned in the context of Asthma [37]. Furthermore, the molecular function ‘antigen binding’ (GO:00003823) was found to be enriched in a study on early-stage lung carcinoma [38].
Typically, differentially expressed genes are detected by statistical tests to judge significance and by the log fold change to quantify the strength of the effect. We have argued in the introduction that there is more evidence for the molecular importance of a set of differentially expressed genes, if these genes show additionally a high performance in a ML model. To estimate this performance more robust, we used the concept of data augmentation. In the two datasets, we have shown that the selected signatures have also a higher diagnostic power when using data augmentation for fitting the ANN models. Following our above argumentation, we think that there is good evidence regarding the molecular role of the selected set of differentially expressed genes and GO terms.

4. Materials and Methods

We developed a standardized pipeline with varying data augmentation methods followed by training a classifier using an increasing number of preselected features and testing the classifier for unseen test data. The pipeline was performed for two real world datasets to assess the applicability of those augmentation methods and their influence on the robustness of a classification model. For the latter, we chose an artificial neural network with two hidden layers. The number of input features was successively increased, with the genes being sorted based on how strongly they are differentially expressed, i.e., incremental feature selection [39]. Two main approaches for data augmentation were taken into account: Generative adversarial networks and a less complex method as a benchmark. The number of synthetic samples generated with the different augmentation methods was two times the number of samples in each class, i.e., the class proportions have been retained. The generic pipeline is shown in Figure 8.

4.1. Example 1: Microarray Data from Infection Study

Data were taken from the GEO archive (https://www.ncbi.nlm.gov/geo, accessed on 15 April 2020) under the accession number GSE63990, containing microarray data of patients with respiratory infection. In the original study [16], RNA from 317 subjects with clear clinical phenotypes, namely bacterial ARI, viral ARI, non-infectious illness and healthy controls was analyzed. Total RNA was extracted from peripheral blood. The authors developed a classifier to diagnose bacterial ARI and viral ARI based on gene-expression data, finding that using patients with non-infectious illness as the control group is more appropriate than healthy individuals to identify gene-expression classifiers. For our study, we only took sequence data from individuals with viral ARI and bacterial ARI, i.e., 190 samples. No preprocessing was conducted, because the available data were already normalized based on robust multi-array average (RMA) [40] and log2 transformed.

4.2. Example 2: RNA-Seq Data from SARS-CoV-2 Study

Samples were taken from GEO archive (https://www.ncbi.nlm.gov/geo, accessed on 23 February 2021) under the accession number GSE163151 and contained RNA-sequencing data from nasopharyngeal swabs of humans with either SARS-CoV-2 infection, other viral ARI or non-viral ARI. Originally, the data also included whole blood samples and were used to investigate the differential host responses to SARS-CoV-2 infection [8]. For our study, we only took count data from nasopharyngeal (NP) swabs of patients with either SARS-CoV-2 infection or viral ARI, i.e., 258 samples. We conducted pre-filtering of the count data keeping only genes where at least ten samples showed at least ten counts. The filtered count data were normalized using the R package DESeq2’ [41].

4.3. K-Folds Cross-Validation

K-folds cross-validation was used to evaluate the different augmentation methods. K-folds cross-validation is a common approach for estimating the accuracy of a classifier, and for model selection [42]. Therefore, the data D is randomly split into K non-overlapping subsets (folds) D 1 , D 2 , …, D K of approximately equal size. The algorithm is therefore trained and validated K times: for each k ϵ { 1 , 2 , , K } , it is trained on D \ D k and validated on D k [43]. In our study, we split the respective training data into K = 10 subsets preserving the percentage of samples per class (i.e., stratified K-folds cross-validation [42]) by a single split of the data before starting the training. Beginning with the first fold as test set, feature selection, data augmentation and training of the classifier is conducted for the respective training subset. The trained network is then used to predict the labels (outcome) of the so-far unseen test data. Performance metrics including accuracy and Brier score are calculated for each fold and stored for later comparisons. Although the accuracy is the proportion of correctly classified samples of all samples, the Brier score measures the mean squared difference between the probability of belonging to class 1 and the actual outcome of the sample (0 or 1):
Let y i ϵ { 0 , 1 } be the outcome of the event for sample i. Transforming the output of the classification ANN as described in Section 4.4 by means of the logistic sigmoid function, i.e., S ( x i ) = e x i e x i + 1 with x i being the output value of the classification ANN for sample i, the latter is mapped to a value between zero and one. This can be interpreted as the estimated probability that sample i has the outcome 1, denoted by p ^ ( y i = 1 ) . The predicted outcome is assigned as follows:
y i ^ = 0 , if p ^ ( y i = 1 ) 0.5 1 , if p ^ ( y i = 1 ) > 0.5
The proportion of correctly classified samples, i.e., y i ^ = = y i , of all samples, is defined as the accuracy.
In contrast, the Brier score measures the mean squared difference between the probability of belonging to class 1 and the actual outcome of the sample [44]:
B r i e r s c o r e = 1 N i = 1 N ( y i p ^ ( y i = 1 ) ) 2
Thus, the Brier score provides a more detailed measure for the performance of a classifier than the accuracy.

4.4. Classification with Artificial Neural Networks

For classification, a deep neural network with two hidden layers with 150 and 50 neurons each was used. The number of input neurons corresponded to the number of input features (genes). In the hidden layers, rectified linear unit activation (ReLU) function was used to transform the input values, i.e., the output of the respective previous layer. ReLU function is a common type of non-linear activation used in hidden layers that maps negative values to zero and for positive values the function returns the input value [45]. The output layer was one single neuron with linear activation. As the criterion for backward propagation, binary crossentropy with logits loss, which is a combination of a sigmoid layer and binary cross entropy loss in one single class, was used. Stochastic gradient descent algorithm was used to optimize the height of the weights, i.e., to minimize the classification loss. Dropout regularization, i.e., random selection of neurons that are set to zero, was used to reduce overfitting [46]. The neural network was built using the open-source program library PyTorch ([47]) and the PyTorch research framework PyTorch Lightning [48].

4.5. Incremental Feature Selection

The number of input features that were passed to the classification ANN was successively increased, with those features being sorted accordingly to their importance, i.e., the most differentially expressed genes. Therefore, a feature selection was carried out for each cross-validation fold and accordingly for each training subset, and the classification network was trained and tested with different numbers of those preselected genes. For both microarray data and RNA-seq data, the selection of genes followed a test for differential expression which was conducted using the R packages ‘Limma’ [49] and ‘DESeq2’ [41], respectively. p-values were obtained for gene ranking and the resulting lists of ranked genes were used to extract the top n genes.

4.6. Augmentation by Weighted Mixed Observations

The approach to augment data by weighted mixed observations is mainly used for image data augmentation and is described by Shorten et al. [21] as mixing images. Similarly, we generated new samples by matrix multiplication. The original data, represented by a p × n matrix, with p being the number of features and n being the number of samples is multiplied by a (n × 1) vector of randomly chosen small weights. Repeating this calculation n * times, each time with a randomly drawn weights vector and appending the vectors generated in this way, results in a n * × p matrix, with n * being the required number of artificial samples. Z-transforming that matrix, we obtain a synthetic dataset consisting of the new generated arrays that lie within the range of the original data.

4.7. Augmentation by Gan

A generative adversarial network (GAN) is a type of generative model based on deep learning [50] which was introduced to the machine-learning community in 2014 by Ian J. Goodfellow [22]. Unlike the supervised deep learning method, we use for classification, GAN models belong to unsupervised learning, i.e., no labels are given to train the model [51]. GANs have been widely used for image datasets and are continually developing. However, to our knowledge, only a few attempts have been made to use GANs for augmentation of gene-expression datasets [23,52,53].
A conventional GAN consists of two deep neural networks which are embedded in a competitive process. The discriminator works as a classifier and discriminates between real and fake samples. The generator generates synthetic samples from a latent vector z to degrade the classification performance of the discriminator, i.e., to maximize the cost of the discriminator [17]. As suggested by Goodfellow et al. [22], the discriminator is trained for c epochs to minimize its loss before the generator is trained for one epoch. Through this competition, the model theoretically learns a generator that creates realistic data, i.e., that samples from the complex, high-dimensional training distribution [22]. Viñas et al. [52] point to a problem with the augmentation of gene-expression data, namely the difficulty of recognizing whether the gene-expression data generated by the GAN is realistic, which is a subordinate problem in other domains such as image generation, where it can be determined by empirical investigation. Essentially, the objective of the GAN is to minimize the so-called Jensen–Shannon Divergence (JSD), which itself is a metric for the similarity between two probability distributions [22].
In our model, the GAN consisted of a generator, which in turn was a MLP with three hidden layers with 250, 500 and 1000 neurons each with ReLU activation, and a discriminator, which was also a MLP with three hidden layers with 1000, 500 and 250 neurons each with leaky ReLU activation. The generator takes a latent vector z, consisting of 100 randomly selected values between 0 and 1 from the normal distribution, processes it through the hidden layers and maps it onto a high-dimensional vector that describes an artificial sample. Thus, the generator has as many output neurons as the number of genes that make up an original sample, each with hyperbolic tangent (Tanh) activation. The discriminator takes such a vector as input vector, i.e., its number of input features corresponds to the number of output neurons of the generator, processes it through the hidden layers and maps it to a single output neuron with linear activation. Binary crossentropy with logits loss was used as the criterion for backward propagation for both discriminator and generator. Adam optimization algorithm [54] with a learning rate of 0.0001 was used to optimize the height of the weights of the discriminator ANN and the generator ANN. Training the GAN started with training of the discriminator network for one time, followed by training of the generator once as well. After training the GAN accordingly for 2000 epochs, a n * × 100 matrix, consisting of n * 100-dimensional noise vectors each containing values that are randomly sampled from gaussian distribution, was passed to the trained generator to generate n * artificial samples. The GAN was built using PyTorch [47].

4.8. Augmentation by Wasserstein Gan with Gradient Penalty

As one problem with GANs is their training instability [55], much of the recent work on GANs is on suggesting variants of the classical GAN architecture to stabilize the training process. One alternative is the Wasserstein GAN (WGAN) proposed by Arjovski et al. [55], which uses the Earth-Mover (Wasserstein) distance instead of Jensen–Shannon divergence (JSD) to measure the closeness of the modelled probability distribution and the real probability distribution. Earth-Mover distance has its roots in transport theory and is, in simple terms, the ‘cost of the optimal transport plan’ [55] to transform one distribution into another. Compared to JSD, the Wasserstein Distance has the advantage that it is always continuous and has an exploitable gradient almost everywhere, which makes it a more sensible cost function if the real data distribution and the parametric probability distribution do not share the support or do not overlap, i.e., when the distributions are disjoint [55,56]. However, the Wasserstein GAN which is based on the Kantorovich-Rubinstein duality of Wasserstein distance, brings an additional Lipschitz requirement, i.e., the discriminator must lie within the space of 1-Lipschitz functions [55]. In the standard WGAN [55], this constraint is implemented using weight clipping. Instead of the latter, Gulrajani et al. [57] propose the use of Wasserstein GAN with gradient penalty, i.e., constraining the gradient norm of the critic with respect to its input, which showed better performance than standard GAN or standard WGAN.
In our study, we employed a Wasserstein GAN with gradient penalty (WGAN-GP). For this we used the same architecture for the generator ANN and the discriminator ANN as with the GAN, but changed the loss function by adding a gradient penalty. For implementing the gradient penalty term, we modified code from Persson [58], who wrote a PyTorch script for a WGAN-GP based on the suggestions of Arjovski et al. [55] and Guljarani et al. [57].

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/ijms23052481/s1.

Author Contributions

Conceptualization, M.K. and K.J.; methodology, M.K. and K.J.; software: M.K., J.K. and B.S.; investigation, M.K.; data curation, M.K.; writing—original draft preparation, M.K. and K.J.; writing—review and editing, M.K., E.C., K.J., B.S., A.B. and J.K.; visualization, M.K.; supervision, K.J.; validation, K.J., E.C. and A.B.; funding acquisition, K.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) grant number 398066876/GRK 2485/1.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in Gene-Expression Omnibus Archive (https://www.ncbi.nlm.gov/geo, accessed on 15 April 2020) under the accession numbers GSE63990 and GSE163151. No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial neural network
ARIAcute respiratory illness
CIConfidence interval
GANGenerative adversarial network
GEOGene-expression omnibus
GOGene ontology
MLMachine learning
MWOMixed weighted observations
ReLURectified linear unit
WGANWasserstein generative adversarial network
WGAN-GPWasserstein generative adversarial network with gradient penalty

References

  1. Bhattacharya, S.; Rosenberg, A.F.; Peterson, D.R.; Grzesik, K.; Baran, A.M.; Ashton, J.M.; Gill, S.R.; Corbett, A.M.; Holden-Wiltse, J.; Topham, D.J.; et al. Transcriptomic biomarkers to discriminate bacterial from nonbacterial infection in adults hospitalized with respiratory illness. Sci. Rep. 2017, 7, 6548. [Google Scholar] [CrossRef] [PubMed]
  2. de Lamballerie, C.N.; Pizzorno, A.; Dubois, J.; Julien, T.; Padey, B.; Bouveret, M.; Traversier, A.; Legras-Lachuer, C.; Lina, B.; Boivin, G.; et al. Characterization of cellular transcriptomic signatures induced by different respiratory viruses in human reconstituted airway epithelia. Sci. Rep. 2019, 9, 11493. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Forno, E.; Celedón, J.C. Epigenomics and transcriptomics in the prediction and diagnosis of childhood asthma: Are we there yet? Front. Pediatr. 2019, 7, 115. [Google Scholar] [CrossRef] [PubMed]
  4. Mejias, A.; Dimo, B.; Suarez, N.M.; Garcia, C.; Suarez-Arrabal, M.C.; Jartti, T.; Blankenship, D.; Jordan-Villegas, A.; Ardura, M.I.; Xu, Z.; et al. Whole blood gene expression profiles to assess pathogenesis and disease severity in infants with respiratory syncytial virus infection. PLoS Med. 2013, 10, e1001549. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Furey, T.S.; Cristianini, N.; Duffy, N.; Bednarski, D.W.; Schummer, M.; Haussler, D. Support vector machine classification and validation of cancer tissue samples using microarray expression data. Bioinformatics 2000, 16, 906–914. [Google Scholar] [CrossRef]
  6. Guo, Y.; Hastie, T.; Tibshirani, R. Regularized linear discriminant analysis and its application in microarrays. Biostatistics 2007, 8, 86–100. [Google Scholar] [CrossRef] [Green Version]
  7. Díaz-Uriarte, R.; De Andres, S.A. Gene selection and classification of microarray data using random forest. BMC Bioinform. 2006, 7, 3. [Google Scholar] [CrossRef] [Green Version]
  8. Ng, D.L.; Granados, A.C.; Santos, Y.A.; Servellita, V.; Goldgof, G.M.; Meydan, C.; Sotomayor-Gonzalez, A.; Levine, A.G.; Balcerek, J.; Han, L.M.; et al. A diagnostic host response biosignature for COVID-19 from RNA profiling of nasal swabs and blood. Sci. Adv. 2021, 7, eabe5984. [Google Scholar] [CrossRef]
  9. Tibshirani, R.; Hastie, T.; Narasimhan, B.; Chu, G. Class prediction by nearest shrunken centroids, with applications to DNA microarrays. Stat. Sci. 2003, 104–117. [Google Scholar] [CrossRef]
  10. Nilsson, R.; Björkegren, J.; Tegnér, J. On reliable discovery of molecular signatures. BMC Bioinform. 2009, 10, 38. [Google Scholar] [CrossRef] [Green Version]
  11. di Iulio, J.; Bartha, I.; Spreafico, R.; Virgin, H.W.; Telenti, A. Transfer transcriptomic signatures for infectious diseases. Proc. Natl. Acad. Sci. USA 2021, 118, e2022486118. [Google Scholar] [CrossRef] [PubMed]
  12. Oshansky, C.M.; Zhang, W.; Moore, E.; Tripp, R.A. The host response and molecular pathogenesis associated with respiratory syncytial virus infection. Future Microbiol. 2009, 4, 279–297. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Zhou, R.; Liu, L.; Wang, Y. Viral proteins recognized by different TLRs. J. Med Virol. 2021, 93, 6116–6123. [Google Scholar] [CrossRef] [PubMed]
  14. Gralinski, L.E.; Baric, R.S. Molecular pathology of emerging coronavirus infections. J. Pathol. 2015, 235, 185–195. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Barrett, T.; Troup, D.B.; Wilhite, S.E.; Ledoux, P.; Rudnev, D.; Evangelista, C.; Kim, I.F.; Soboleva, A.; Tomashevsky, M.; Edgar, R. NCBI GEO: Mining tens of millions of expression profiles—Database and tools update. Nucleic Acids Res. 2007, 35, D760–D765. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Tsalik, E.L.; Henao, R.; Nichols, M.; Burke, T.; Ko, E.R.; McClain, M.T.; Hudson, L.L.; Mazur, A.; Freeman, D.H.; Veldman, T.; et al. Host gene expression classifiers diagnose acute respiratory illness etiology. Sci. Transl. Med. 2016, 8, 322ra11. [Google Scholar] [CrossRef] [Green Version]
  17. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  18. Min, S.; Lee, B.; Yoon, S. Deep learning in bioinformatics. Briefings Bioinform. 2017, 18, 851–869. [Google Scholar] [CrossRef] [Green Version]
  19. Liu, Y.; Zhou, Y.; Liu, X.; Dong, F.; Wang, C.; Wang, Z. Wasserstein GAN-based small-sample augmentation for new-generation artificial intelligence: A case study of cancer-staging data in biology. Engineering 2019, 5, 156–163. [Google Scholar] [CrossRef]
  20. Taylor, L.; Nitschke, G. Improving deep learning using generic data augmentation. arXiv 2017, arXiv:1708.06020. [Google Scholar]
  21. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  22. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  23. Chaudhari, P.; Agrawal, H.; Kotecha, K. Data augmentation using MG-GAN for improved cancer classification on gene expression data. Soft Comput. 2019, 24, 11381–11391. [Google Scholar] [CrossRef]
  24. Home—Gene—NCBI. National Center for Biotechnology Information. Available online: https://www.ncbi.nlm.nih.gov/gene (accessed on 15 April 2020).
  25. Pfaender, S.; Mar, K.B.; Michailidis, E.; Kratzel, A.; Boys, I.N.; V’kovski, P.; Fan, W.; Kelly, J.N.; Hirt, D.; Ebert, N.; et al. LY6E impairs coronavirus fusion and confers immune control of viral disease. Nat. Microbiol. 2020, 5, 1330–1339. [Google Scholar] [CrossRef] [PubMed]
  26. Schoggins, J.W.; Wilson, S.J.; Panis, M.; Murphy, M.Y.; Jones, C.T.; Bieniasz, P.; Rice, C.M. A diverse range of gene products are effectors of the type I interferon antiviral response. Nature 2011, 472, 481–485. [Google Scholar] [CrossRef] [PubMed]
  27. Zhu, J.; Ghosh, A.; Sarkar, S.N. OASL—A new player in controlling antiviral innate immunity. Curr. Opin. Virol. 2015, 12, 15–19. [Google Scholar] [CrossRef] [Green Version]
  28. Murphy, T.L.; Tussiwand, R.; Murphy, K.M. Specificity through cooperation: BATF–IRF interactions control immune-regulatory networks. Nat. Rev. Immunol. 2013, 13, 499–509. [Google Scholar] [CrossRef]
  29. Rose, C.E., Jr.; Sung, S.S.J.; Fu, S.M. Significant involvement of CCL2 (MCP-1) in inflammatory disorders of the lung. Microcirculation 2003, 10, 273–288. [Google Scholar] [CrossRef]
  30. Alexa, A.; Rahnenführer, J.; Lengauer, T. Improved scoring of functional groups from gene expression data by decorrelating GO graph structure. Bioinformatics 2006, 22, 1600–1607. [Google Scholar] [CrossRef] [Green Version]
  31. Giles, P.J.; Kipling, D. Normality of oligonucleotide microarray data and implications for parametric statistical analyses. Bioinformatics 2003, 19, 2254–2262. [Google Scholar] [CrossRef] [Green Version]
  32. Kruppa, J.; Kramer, F.; Beißbarth, T.; Jung, K. A simulation framework for correlated count data of features subsets in high-throughput sequencing or proteomics experiments. Stat. Appl. Genet. Mol. Biol. 2016, 15, 401–414. [Google Scholar] [CrossRef]
  33. Goutte, C.; Gaussier, E. A probabilistic interpretation of precision, recall and F-score, with implication for evaluation. In Proceedings of the European Conference on Information Retrieval, Santiago de Compostela, Spain, 21–23 March 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 345–359. [Google Scholar]
  34. Roberson, E.; Mesa, R.A.; Morgan, G.A.; Cao, L.; Marin, W.; Pachman, L.M. Transcriptomes of peripheral blood mononuclear cells from juvenile dermatomyositis patients show elevated inflammation even when clinically 2 inactive. Sci. Rep. 2022, 12, 275. [Google Scholar] [CrossRef] [PubMed]
  35. Mahmud, S.H.; Al-Mustanjid, M.; Akter, F.; Rahman, M.S.; Ahmed, K.; Rahman, M.H.; Chen, W.; Moni, M.A. Bioinformatics and system biology approach to identify the influences of SARS-CoV-2 infections to idiopathic pulmonary fibrosis and chronic obstructive pulmonary disease patients. Briefings Bioinform. 2021, 22, bbab115. [Google Scholar] [CrossRef] [PubMed]
  36. Gollapalli, P.; B.S, S.; Rimac, H.; Patil, P.; Nalilu, S.K.; Kandagalla, S.; Shetty, P. Pathway enrichment analysis of virus-host interactome and prioritization of novel compounds targeting the spike glycoprotein receptor binding domain–human angiotensin-converting enzyme 2 interface to combat SARS-CoV-2. J. Biomol. Struct. Dyn. 2020; ePub ahead of print. [Google Scholar] [CrossRef] [PubMed]
  37. Yin, H.; Fan, Y.; Mu, D.; Song, F.; Tian, F.; Yin, Q. Transcriptomic Analysis Exploring the Molecular Mechanisms of Hanchuan Zupa Granules in Alleviating Asthma in Rat. Evid.-Based Complement. Altern. Med. 2021, 2021, 5584099. [Google Scholar] [CrossRef]
  38. Liu, J.; Yang, X.; Zhang, L.; Yang, B.; Rao, W.; Li, M.; Dai, N.; Yang, Y.; Qian, C.; Zhang, L.; et al. Microarray analysis of the expression profile of immune-related gene in rapid recurrence early-stage lung adenocarcinoma. J. Cancer Res. Clin. Oncol. 2020, 146, 2299–2310. [Google Scholar] [CrossRef]
  39. Liu, H.; Setiono, R. Incremental feature selection. Appl. Intell. 1998, 9, 217–230. [Google Scholar] [CrossRef]
  40. Irizarry, R.A.; Hobbs, B.; Collin, F.; Beazer-Barclay, Y.D.; Antonellis, K.J.; Scherf, U.; Speed, T.P. Exploration, normalization, and summaries of high density oligonucleotide array probe level data. Biostatistics 2003, 4, 249–264. [Google Scholar] [CrossRef] [Green Version]
  41. Love, M.; Anders, S.; Huber, W. Differential analysis of count data—The DESeq2 package. Genome Biol. 2014, 15, 10–1186. [Google Scholar]
  42. Zeng, X.; Martinez, T.R. Distribution-balanced stratified cross-validation for accuracy estimation. J. Exp. Theor. Artif. Intell. 2000, 12, 1–12. [Google Scholar] [CrossRef]
  43. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. In Proceedings of the 14th International Joint Conference on Artificial intelligence, Montreal, QC, Canada, 20–25 August 1995; Volume 14, pp. 1137–1145. [Google Scholar]
  44. Brier, G.W. Verification of forecasts expressed in terms of probability. Mon. Weather. Rev. 1950, 78, 1–3. [Google Scholar] [CrossRef]
  45. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  46. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  47. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  48. Falcon, W.A. PyTorch Lightning. GitHub. 2019, Volume 3. Available online: https://github.com/PyTorchLightning/pytorch-lightning (accessed on 15 February 2021).
  49. Smyth, G.K. Limma: Linear models for microarray data. In Bioinformatics and Computational Biology Solutions using R and Bioconductor; Springer: Cham, Switzerland, 2005; pp. 397–420. [Google Scholar]
  50. Creswell, A.; White, T.; Dumoulin, V.; Arulkumaran, K.; Sengupta, B.; Bharath, A.A. Generative adversarial networks: An overview. IEEE Signal Process. Mag. 2018, 35, 53–65. [Google Scholar] [CrossRef] [Green Version]
  51. Sathya, R.; Abraham, A. Comparison of supervised and unsupervised learning algorithms for pattern classification. Int. J. Adv. Res. Artif. Intell. 2013, 2, 34–38. [Google Scholar] [CrossRef] [Green Version]
  52. Viñas, R.; Andrés-Terré, H.; Liò, P.; Bryson, K. Adversarial generation of gene expression data. bioRxiv 2019, 38, 836254. [Google Scholar] [CrossRef]
  53. Marouf, M.; Machart, P.; Bansal, V.; Kilian, C.; Magruder, D.S.; Krebs, C.F.; Bonn, S. Realistic in silico generation and augmentation of single-cell RNA-seq data using generative adversarial networks. Nat. Commun. 2020, 11, 166. [Google Scholar] [CrossRef] [Green Version]
  54. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  55. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  56. Weng, L. From gan to wgan. arXiv 2019, arXiv:1904.08994. [Google Scholar]
  57. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved training of wasserstein gans. arXiv 2017, arXiv:1704.00028. [Google Scholar]
  58. Persson, A. WGAN-GP. GitHub. 2021. Available online: https://github.com/aladdinpersson/Machine-Learning-Collection/tree/master/ML/Pytorch\/GANs/4.%20WGAN-GP (accessed on 5 April 2021).
Figure 1. Brier scores from 10-fold cross-validation for the original microarray dataset (orange) and the original RNA-seq dataset (blue) when including an increasing number of top differentially expressed genes as input features for the ANN. The median Brier score tended to decrease when increasing the number of genes as predictors from 50 to 750.
Figure 1. Brier scores from 10-fold cross-validation for the original microarray dataset (orange) and the original RNA-seq dataset (blue) when including an increasing number of top differentially expressed genes as input features for the ANN. The median Brier score tended to decrease when increasing the number of genes as predictors from 50 to 750.
Ijms 23 02481 g001
Figure 2. Results from 10-fold cross-validation in terms of accuracies (A) and Brier scores (B) for the microarray data when using different data augmentation methods and various numbers of top genes as input signatures. Error bars represent the 95% confidence interval of the mean.
Figure 2. Results from 10-fold cross-validation in terms of accuracies (A) and Brier scores (B) for the microarray data when using different data augmentation methods and various numbers of top genes as input signatures. Error bars represent the 95% confidence interval of the mean.
Ijms 23 02481 g002
Figure 3. Heatmaps representing covariance matrices for selected genes from the microarray study. Covariance structures of original and artificial samples are compared for two sets of 50 randomly chosen genes. The covariance structures are best maintained when generating artificial samples with the MWO or WGAN-GP approach. The details are showed in (A,B).
Figure 3. Heatmaps representing covariance matrices for selected genes from the microarray study. Covariance structures of original and artificial samples are compared for two sets of 50 randomly chosen genes. The covariance structures are best maintained when generating artificial samples with the MWO or WGAN-GP approach. The details are showed in (A,B).
Ijms 23 02481 g003
Figure 4. Distribution of normalized mean square errors (top) and mean absolute difference of eigenvalues (bottom) for microarray data when comparing the covariance matrices of artificial and original data. The boxplots represent data for 500 runs, in each of which 50 genes were randomly drawn and the correlation matrices were created. The symbol **** denotes p-value < 0.0001.
Figure 4. Distribution of normalized mean square errors (top) and mean absolute difference of eigenvalues (bottom) for microarray data when comparing the covariance matrices of artificial and original data. The boxplots represent data for 500 runs, in each of which 50 genes were randomly drawn and the correlation matrices were created. The symbol **** denotes p-value < 0.0001.
Ijms 23 02481 g004
Figure 5. Cross validation classification accuracies (A) and Brier scores (B) for the RNA-seq experiment for different augmentation methods and various numbers of input features. Error bars represent the 95% confidence interval of the mean.
Figure 5. Cross validation classification accuracies (A) and Brier scores (B) for the RNA-seq experiment for different augmentation methods and various numbers of input features. Error bars represent the 95% confidence interval of the mean.
Ijms 23 02481 g005
Figure 6. Heatmaps illustrating covariance matrices for the RNA-seq data study. Comparison of covariance matrices of original and artificial samples for two sets of 50 randomly chosen genes. The details are showed in (A,B).
Figure 6. Heatmaps illustrating covariance matrices for the RNA-seq data study. Comparison of covariance matrices of original and artificial samples for two sets of 50 randomly chosen genes. The details are showed in (A,B).
Ijms 23 02481 g006
Figure 7. Boxplots of log-transformed normalized mean square errors (top) and log-transformed mean absolute differences of eigenvalues (bottom) for the RNA-seq experiment when comparing the covariance matrices of artificial and original data. The boxplots contain results for 500 runs, in each of which 50 genes were drawn at random and the correlation matrices were created. The symbol **** denotes p-value < 0.0001.
Figure 7. Boxplots of log-transformed normalized mean square errors (top) and log-transformed mean absolute differences of eigenvalues (bottom) for the RNA-seq experiment when comparing the covariance matrices of artificial and original data. The boxplots contain results for 500 runs, in each of which 50 genes were drawn at random and the correlation matrices were created. The symbol **** denotes p-value < 0.0001.
Ijms 23 02481 g007
Figure 8. Generic pipeline: Starting at the top left the whole dataset is split into a training subset and a test subset. The training subset is then augmented using various augmentation methods. Following a feature selection, the augmented dataset with a reduced number of genes is used to train a classification network which itself is used to classify the so-far unseen test data.
Figure 8. Generic pipeline: Starting at the top left the whole dataset is split into a training subset and a test subset. The training subset is then augmented using various augmentation methods. Following a feature selection, the augmented dataset with a reduced number of genes is used to train a classification network which itself is used to classify the so-far unseen test data.
Ijms 23 02481 g008
Table 1. Classification performance measures for microarray data (test subsets) after training with original training subset only. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
Table 1. Classification performance measures for microarray data (test subsets) after training with original training subset only. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
MethodNumber of Top GenesAccuracyBrier ScoreSensitivitySpecificity
No500.86 ± 0.04400.16 ± 0.01430.92 ± 0.05150.76 ± 0.0998
Augmentation1000.86 ± 0.04130.15 ± 0.01400.92 ± 0.04870.78 ± 0.1090
1500.87 ± 0.04660.13 ± 0.01210.92 ± 0.05140.79 ± 0.1090
2000.88 ± 0.04880.13 ± 0.01280.93 ± 0.05350.79 ± 0.1090
2500.88 ± 0.04360.12 ± 0.01240.93 ± 0.04110.79 ± 0.1037
3000.88 ± 0.04290.12 ± 0.01290.92 ± 0.05140.82 ± 0.0887
3500.88 ± 0.04630.11 ± 0.01440.92 ± 0.05150.81 ± 0.0939
4000.87 ± 0.04660.11 ± 0.01680.92 ± 0.04870.81 ± 0.1028
4500.87 ± 0.05150.10 ± 0.01840.92 ± 0.05150.79 ± 0.1118
5000.88 ± 0.04630.10 ± 0.01600.92 ± 0.05140.81 ± 0.0998
5500.88 ± 0.04010.10 ± 0.01450.92 ± 0.04870.84 ± 0.0748
6000.88 ± 0.04880.10 ± 0.01640.92 ± 0.05150.81 ± 0.0939
6500.88 ± 0.04810.10 ± 0.01840.92 ± 0.05150.82 ± 0.0919
7000.88 ± 0.04290.10 ± 0.01680.92 ± 0.05150.82 ± 0.0783
7500.89 ± 0.04470.10 ± 0.01660.93 ± 0.04110.82 ± 0.0919
Table 2. Classification performance measures for microarray data (test subsets) after training with augmented training data either augmented with mixed weighted observations (‘MWO’), a standard generative adversarial network (‘GAN’) or the advanced Wasserstein GAN with gradient penalty (‘WGAN-GP’) with the number of synthetic samples being two times the number of original samples in each class, respectively. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
Table 2. Classification performance measures for microarray data (test subsets) after training with augmented training data either augmented with mixed weighted observations (‘MWO’), a standard generative adversarial network (‘GAN’) or the advanced Wasserstein GAN with gradient penalty (‘WGAN-GP’) with the number of synthetic samples being two times the number of original samples in each class, respectively. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
MethodNumber of Top GenesAccuracyBrier ScoreSensitivitySpecificity
MWO500.87 ± 0.03840.09 ± 0.01810.91 ± 0.04510.81 ± 0.0805
1000.88 ± 0.03460.09 ± 0.02160.91 ± 0.04510.83 ± 0.0798
1500.88 ± 0.03780.09 ± 0.02280.92 ± 0.04870.82 ± 0.0783
2000.88 ± 0.04560.09 ± 0.02350.92 ± 0.04870.84 ± 0.0953
2500.88 ± 0.04080.09 ± 0.02450.91 ± 0.04510.84 ± 0.0953
3000.88 ± 0.04080.09 ± 0.02430.91 ± 0.04510.84 ± 0.0953
3500.88 ± 0.04560.09 ± 0.02360.92 ± 0.04870.84 ± 0.0953
4000.89 ± 0.04610.09 ± 0.02390.92 ± 0.04870.86 ± 0.0760
4500.88 ± 0.04010.09 ± 0.02470.92 ± 0.04870.84 ± 0.0748
5000.88 ± 0.04010.09 ± 0.02420.91 ± 0.04510.85 ± 0.0815
5500.89 ± 0.04610.09 ± 0.02400.92 ± 0.04870.86 ± 0.0760
6000.89 ± 0.04610.09 ± 0.02350.92 ± 0.04870.86 ± 0.0760
6500.89 ± 0.03910.09 ± 0.02380.91 ± 0.04510.86 ± 0.0760
7000.89 ± 0.04470.09 ± 0.02400.92 ± 0.04870.85 ± 0.0815
7500.89 ± 0.04200.09 ± 0.02370.91 ± 0.04510.86 ± 0.0760
GAN500.87 ± 0.03830.09 ± 0.02480.9 ± 0.04840.84 ± 0.0748
1000.87 ± 0.03830.10 ± 0.02610.90 ± 0.04840.84 ± 0.0748
1500.87 ± 0.04420.09 ± 0.02750.90 ± 0.04840.82 ± 0.0919
2000.87 ± 0.03840.10 ± 0.03070.89 ± 0.04970.84 ± 0.0748
2500.87 ± 0.03840.10 ± 0.03080.89 ± 0.04970.84 ± 0.0748
3000.86 ± 0.04360.10 ± 0.02830.88 ± 0.05040.82 ± 0.0919
3500.86 ± 0.04660.10 ± 0.02860.88 ± 0.05040.84 ± 0.0889
4000.86 ± 0.04660.10 ± 0.02970.88 ± 0.05040.84 ± 0.0889
4500.86 ± 0.04660.10 ± 0.02800.87 ± 0.05610.85 ± 0.0848
5000.87 ± 0.04660.10 ± 0.02910.88 ± 0.05040.87 ± 0.0897
5500.87 ± 0.04910.10 ± 0.02980.89 ± 0.04970.85 ± 0.0945
6000.86 ± 0.04660.10 ± 0.02970.88 ± 0.05040.84 ± 0.0889
6500.87 ± 0.04680.10 ± 0.02750.88 ± 0.05040.85 ± 0.0848
7000.88 ± 0.04880.10 ± 0.02700.89 ± 0.04970.87 ± 0.0897
7500.87 ± 0.04130.09 ± 0.02770.88 ± 0.05600.86 ± 0.0635
WGAN-GP500.87 ± 0.03840.09 ± 0.0210.92 ± 0.04870.79 ± 0.0916
1000.88 ± 0.03460.09 ± 0.02350.91 ± 0.04510.83 ± 0.0798
1500.88 ± 0.04010.09 ± 0.02490.92 ± 0.04870.84 ± 0.0748
2000.87 ± 0.03840.09 ± 0.02640.91 ± 0.04510.81 ± 0.0906
2500.88 ± 0.04010.09 ± 0.02540.91 ± 0.04510.85 ± 0.0815
3000.88 ± 0.04290.09 ± 0.02400.92 ± 0.04870.84 ± 0.0953
3500.88 ± 0.04290.09 ± 0.02600.92 ± 0.04870.84 ± 0.0953
4000.88 ± 0.04080.09 ± 0.02500.91 ± 0.04510.84 ± 0.0953
4500.90 ± 0.04470.08 ± 0.02310.93 ± 0.05350.85 ± 0.0862
5000.89 ± 0.04200.09 ± 0.02450.92 ± 0.04870.85 ± 0.0815
5500.90 ± 0.04200.08 ± 0.02410.92 ± 0.04870.88 ± 0.0688
6000.89 ± 0.03910.09 ± 0.02350.92 ± 0.04870.85 ± 0.0862
6500.88 ± 0.04010.09 ± 0.02480.91 ± 0.04510.85 ± 0.0916
7000.91 ± 0.04010.09 ± 0.02470.92 ± 0.05140.88 ± 0.0618
7500.90 ± 0.04200.09 ± 0.02310.92 ± 0.05140.86 ± 0.0696
Table 3. Classification performance measures for RNA-seq data (test subsets) after training with original training data only. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
Table 3. Classification performance measures for RNA-seq data (test subsets) after training with original training data only. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
MethodNumber of Top GenesAccuracyBrier ScoreSensitivitySpecificity
No500.59 ±0.05570.24 ±0.00270.24 ±0.11000.98 ±0.0218
Augmentation1000.64 ± 0.05640.24 ± 0.00500.52 ± 0.18060.79 ± 0.2591
1500.63 ± 0.05630.24 ± 0.00510.43 ± 0.15010.86 ± 0.1902
2000.66 ± 0.05270.24 ± 0.00730.43 ± 0.09760.92 ± 0.0544
2500.66 ± 0.05240.24 ± 0.00740.54 ± 0.12110.79 ± 0.1842
3000.64 ± 0.05120.24 ± 0.00510.54 ± 0.13320.76 ± 0.1813
3500.63 ± 0.05500.24 ± 0.00560.50 ± 0.14050.78 ± 0.1825
4000.67 ± 0.04640.23 ± 0.00720.51 ± 0.07200.86 ± 0.049
4500.67 ± 0.05180.23 ± 0.01090.57 ± 0.11600.78 ± 0.1652
5000.65 ± 0.06570.23 ± 0.01170.46 ± 0.11850.86 ± 0.0845
5500.67 ± 0.04110.23 ± 0.00860.54 ± 0.06300.82 ± 0.0722
6000.67 ± 0.04690.23 ± 0.01100.54 ± 0.06630.82 ± 0.0514
6500.66 ± 0.04250.23 ± 0.00690.53 ± 0.05160.80 ± 0.0697
7000.64 ± 0.05700.23 ± 0.00780.52 ± 0.04820.78 ± 0.1222
7500.67 ± 0.03550.22 ± 0.00940.54 ± 0.05470.81 ± 0.0490
Table 4. Classification performance measures for RNA-seq data (test subsets) after training with augmented training data either augmented with mixed weighted observations (‘MWO’), a standard generative adversarial network (‘GAN’) or the advanced Wasserstein GAN with gradient penalty (‘WGAN-GP’) with the number of synthetic samples being two times the number of original samples in each class, respectively. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
Table 4. Classification performance measures for RNA-seq data (test subsets) after training with augmented training data either augmented with mixed weighted observations (‘MWO’), a standard generative adversarial network (‘GAN’) or the advanced Wasserstein GAN with gradient penalty (‘WGAN-GP’) with the number of synthetic samples being two times the number of original samples in each class, respectively. The table shows the 95% confidence intervals (mean ± 1.96 standard errors) for accuracy, Brier score, sensitivity and specificity, respectively. Values in bold are the best scores for the respective performance metric.
MethodNumber of Top GenesAccuracyBrier ScoreSensitivitySpecificity
MWO500.68 ± 0.05330.22 ± 0.01110.43 ± 0.09330.96 ± 0.0365
1000.71 ± 0.04500.22 ± 0.01250.52 ± 0.06540.92 ± 0.0544
1500.70 ± 0.04480.21 ± 0.01090.53 ± 0.07290.89 ± 0.0691
2000.71 ± 0.04470.21 ± 0.01260.56 ± 0.06510.89 ± 0.0547
2500.70 ± 0.05110.21 ± 0.01110.56 ± 0.05950.86 ± 0.0772
3000.72 ± 0.03380.20 ± 0.01120.58 ± 0.04810.88 ± 0.0609
3500.72 ± 0.03740.19 ± 0.01260.59 ± 0.05050.87 ± 0.0606
4000.71 ± 0.03410.19 ± 0.01360.60 ± 0.04660.82 ± 0.0618
4500.71 ± 0.04050.19 ± 0.01750.62 ± 0.04800.81 ± 0.0732
5000.71 ± 0.03380.19 ± 0.01390.62 ± 0.05370.82 ± 0.0618
5500.71 ± 0.03380.19 ± 0.01310.62 ± 0.04870.81 ± 0.0425
6000.71 ± 0.03380.20 ± 0.01940.64 ± 0.04060.79 ± 0.0609
6500.71 ± 0.03730.19 ± 0.01600.63 ± 0.04750.81 ± 0.0547
7000.71 ± 0.03180.19 ± 0.01510.63 ± 0.05180.80 ± 0.0499
7500.71 ± 0.03730.19 ± 0.01650.63 ± 0.04750.81 ± 0.0547
GAN500.63 ± 0.07920.24 ± 0.01020.76 ± 0.17490.48 ± 0.3068
1000.67 ± 0.08280.23 ± 0.01130.78 ± 0.12920.53 ± 0.2768
1500.66 ± 0.07940.23 ± 0.01310.80 ± 0.12470.50 ± 0.2700
2000.70 ± 0.08360.22 ± 0.01570.86 ± 0.09340.52 ± 0.2481
2500.70 ± 0.08050.22 ± 0.01830.87 ± 0.07760.50 ± 0.2435
3000.70 ± 0.08730.21 ± 0.01780.87 ± 0.09740.50 ± 0.2447
3500.70 ± 0.08230.21 ± 0.02070.89 ± 0.08400.49 ± 0.2354
4000.70 ± 0.07980.20 ± 0.02020.89 ± 0.08400.48 ± 0.2320
4500.68 ± 0.07470.20 ± 0.02730.89 ± 0.07600.43 ± 0.2294
5000.68 ± 0.07240.20 ± 0.02220.89 ± 0.07600.44 ± 0.2259
5500.67 ± 0.07120.20 ± 0.02250.88 ± 0.08620.43 ± 0.2242
6000.68 ± 0.05500.21 ± 0.02250.89 ± 0.08140.45 ± 0.1936
6500.71 ± 0.05910.20 ± 0.02280.89 ± 0.07600.49 ± 0.1909
7000.71 ± 0.06530.20 ± 0.02620.91 ± 0.06280.49 ± 0.2029
7500.71 ± 0.05970.20 ± 0.02440.90 ± 0.06330.48 ± 0.1802
WGAN-GP500.70 ± 0.04180.22 ± 0.01140.48 ± 0.08010.95 ± 0.0436
1000.72 ± 0.03330.20 ± 0.01110.56 ± 0.03670.92 ± 0.0544
1500.72 ± 0.03340.20 ± 0.00880.59 ± 0.03980.88 ± 0.0609
2000.74 ± 0.04740.18 ± 0.01430.65 ± 0.05800.85 ± 0.0722
2500.76 ± 0.04620.18 ± 0.01370.68 ± 0.06050.86 ± 0.0646
3000.77 ± 0.04480.18 ± 0.01200.70 ± 0.05720.86 ± 0.0691
3500.76 ± 0.03790.18 ± 0.01430.70 ± 0.05330.83 ± 0.0568
4000.74 ± 0.04690.17 ± 0.01460.70 ± 0.06440.80 ± 0.0606
4500.74 ± 0.04690.17 ± 0.01650.71 ± 0.05840.78 ± 0.0697
5000.75 ± 0.04040.17 ± 0.01650.72 ± 0.04820.78 ± 0.0653
5500.74 ± 0.04400.18 ± 0.01600.72 ± 0.05000.75 ± 0.0689
6000.72 ± 0.03850.18 ± 0.01770.70 ± 0.04900.75 ± 0.0596
6500.72 ± 0.04020.18 ± 0.01870.70 ± 0.05610.75 ± 0.0596
7000.74 ± 0.03730.18 ± 0.02000.75 ± 0.06680.74 ± 0.0665
7500.74 ± 0.03440.18 ± 0.01730.74 ± 0.04720.74 ± 0.0618
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kircher, M.; Chludzinski, E.; Krepel, J.; Saremi, B.; Beineke, A.; Jung, K. Augmentation of Transcriptomic Data for Improved Classification of Patients with Respiratory Diseases of Viral Origin. Int. J. Mol. Sci. 2022, 23, 2481. https://doi.org/10.3390/ijms23052481

AMA Style

Kircher M, Chludzinski E, Krepel J, Saremi B, Beineke A, Jung K. Augmentation of Transcriptomic Data for Improved Classification of Patients with Respiratory Diseases of Viral Origin. International Journal of Molecular Sciences. 2022; 23(5):2481. https://doi.org/10.3390/ijms23052481

Chicago/Turabian Style

Kircher, Magdalena, Elisa Chludzinski, Jessica Krepel, Babak Saremi, Andreas Beineke, and Klaus Jung. 2022. "Augmentation of Transcriptomic Data for Improved Classification of Patients with Respiratory Diseases of Viral Origin" International Journal of Molecular Sciences 23, no. 5: 2481. https://doi.org/10.3390/ijms23052481

APA Style

Kircher, M., Chludzinski, E., Krepel, J., Saremi, B., Beineke, A., & Jung, K. (2022). Augmentation of Transcriptomic Data for Improved Classification of Patients with Respiratory Diseases of Viral Origin. International Journal of Molecular Sciences, 23(5), 2481. https://doi.org/10.3390/ijms23052481

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop