Next Article in Journal
Well-Preserved Urinary Bladder Anatomy in Rats After Minimally Invasive Surgery
Previous Article in Journal
Non-Invasive Retinal Biomarkers for Early Diagnosis of Alzheimer’s Disease
Previous Article in Special Issue
Fluorescence Angiography with Dual Fluorescence for the Early Detection and Longitudinal Quantitation of Vascular Leakage in Retinopathy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Future Picture: A Review of Current Generative Adversarial Neural Networks in Vitreoretinal Pathologies and Their Future Potentials

by
Raheem Remtulla
1,
Adam Samet
1,
Merve Kulbay
1,2,
Arjin Akdag
3,
Adam Hocini
4,
Anton Volniansky
5,
Shigufa Kahn Ali
2,6 and
Cynthia X. Qian
2,6,*
1
Department of Ophthalmology & Visual Sciences, McGill University, Montreal, QC H4A 3SE, Canada
2
Centre de Recherche de l’Hôpital Maisonneuve-Rosemont, Université de Montréal, Montreal, QC H1T 2M4, Canada
3
Faculty of Medicine and Health Sciences, McGill University, Montreal, QC H3G 2M1, Canada
4
Faculty of Medicine, Université de Montréal, Montreal, QC H3T 1J4, Canada
5
Department of Psychiatry, Université Laval, Quebec City, QC G1V 0A6, Canada
6
Department of Ophthalmology, Centre Universitaire d’Ophtalmologie (CUO), Hôpital Maisonneuve-Rosemont, University of Montreal, Montreal, QC H1T 2M4, Canada
*
Author to whom correspondence should be addressed.
Biomedicines 2025, 13(2), 284; https://doi.org/10.3390/biomedicines13020284
Submission received: 9 December 2024 / Revised: 6 January 2025 / Accepted: 14 January 2025 / Published: 24 January 2025
(This article belongs to the Special Issue Retinal Diseases: Imaging and Treatment)

Abstract

:
Machine learning has transformed ophthalmology, particularly in predictive and discriminatory models for vitreoretinal pathologies. However, generative modeling, especially generative adversarial networks (GANs), remains underexplored. GANs consist of two neural networks—the generator and discriminator—that work in opposition to synthesize highly realistic images. These synthetic images can enhance diagnostic accuracy, expand the capabilities of imaging technologies, and predict treatment responses. GANs have already been applied to fundus imaging, optical coherence tomography (OCT), and fluorescein autofluorescence (FA). Despite their potential, GANs face challenges in reliability and accuracy. This review explores GAN architecture, their advantages over other deep learning models, and their clinical applications in retinal disease diagnosis and treatment monitoring. Furthermore, we discuss the limitations of current GAN models and propose novel applications combining GANs with OCT, OCT-angiography, fluorescein angiography, fundus imaging, electroretinograms, visual fields, and indocyanine green angiography.

1. Introduction

The potential applications of artificial intelligence (AI) in medicine have significantly increased in recent years. Novel AI-derived models have been developed with increasing computational power over the years. Advances in image classification (e.g., improvement of top-five accuracy on ImageNet from ~84% (AlexNet, 2012) to over 99% (EfficientNet, 2022)) and reinforcement learning (e.g., the surge in Atari games performance from ~200% of the human score (DQN, 2015) to ~800% of the human score (2023)) have been reported [1,2]. Initially, the academic interest was directed towards discriminatory models, which allow for discrimination between variables and perform classifications. These discriminatory models include deep convolutional neural networks (DCNNs) [3], convolutional neural networks (CNNs) [4,5], artificial neural networks (ANNs) [6,7], random forest [8], and decision tree [9]. Several of these models have already been integrated into clinical practice within the field of ophthalmology [10,11,12]. In 2018, the Food and Drug administration (FDA) approved for clinical practice the first-ever neural network, an algorithm capable of early identification of diabetic retinopathy (DR) with a sensitivity ranging from 92% to 93% and specificity ranging from 89% to 94% [13]. The incorporation of AI in health care also occurred in other countries, such as India, where similar networks allow screening for the early identification and treatment of patients with DR [13,14,15]. Further machine learning models have been used in screening patients with retinopathy of prematurity (ROP) [16], age-related macular degeneration (AMD) [17], and vein occlusions [18,19,20]. The extent of machine learning algorithms is not restricted to diagnosis; algorithms can also perform predictions regarding therapy response. Indeed, machine learning approaches have demonstrated the ability to predict response to anti-vascular endothelial growth factor (VEGF) in patients with AMD [21,22,23,24].
In recent years, generative models have also gained worldwide interest. Generative models learn patterns from a dataset and generate new but similar data [25,26,27,28,29,30]. Despite the large interest in the generative data science field, such as building models with text and language inputs for generative models [31], image generation has not gathered as much attention in the field of medicine compared to conventional discriminatory models. Image generation has a multitude of clinical applications, particularly in image-based specialties such as radiology and ophthalmology [32,33]. Generative adversarial networks (GANs) are the most common models used to generate realistic synthetic images. These models have proven their high degree of realism when generating high-resolution images [34,35]. In radiology, GANs have been demonstrated to improve the quality of diagnostic imaging by reducing noise and artifacts [36,37,38]. They have even been used to convert images between computational tomography (CT) and magnetic resonance imaging (MRI) acquisitions [39]. In ophthalmology, GANs have been implicated in image segmentation to identify structures such as optic nerve cupping [40], retinal vessels [41], and the meibomian glands [42]. GANs have also demonstrated the ability to generate fluorescein angiograms from fundus photos [43], as well as predicting response to treatments such as anti-VEGF therapy. Given that the applications of GANs in ophthalmology are likely to expand with increasing databases and computing power (Figure 1), it is essential to provide an overview that explains how these networks function and addresses their common pitfalls within the context of ophthalmology. The aim of this review is to serve as a primer for vitreoretinal surgeons by focusing on three key objectives: (1) to provide a foundational understanding of GAN architecture and functionality, (2) to explore clinical applications in retinal imaging, and (3) to highlight future directions and limitations of GANs in ophthalmology.

2. Fundamental Concepts on Generative Adversarial Networks

2.1. Basic Terminology of GANs and Neural Networks

To help navigate through the technical sections involving GANs, Table 1 summarizes common terminology encountered. Most terms will also be defined individually within the text when presented for the first time. We suggest referring to the references provided in Table 1 for more in-depth explanations of GAN architecture.

2.2. How Do GANs Work? Explanation with the Counterfeit’s Analogy

GANs were developed in 2014 in Montreal (Quebec, Canada) by Ian Goodfellow, a PhD student, under the supervision of Yoshua Bengio and his team [53]. GANs function on an adversarial game theory model between two CNNs: a generator and a discriminator (Figure 2) [33,36,53,54,55,56,57]. The analogy often employed to illustrate this concept involves a counterfeit representing the generator and a police officer representing the discriminator. The purpose of the counterfeit is to produce an authentic copy of an object, while the purpose of the police officer is to distinguish the replica from the original. As the counterfeit begins to learn his art, his first attempts are easily distinguishable from the original objects, and the police officer identifies the replica. However, the feedback from the police officer about the counterfeit serves as a quality control step and forces the counterfeiter to improve his replica. Thus, after several attempts, the counterfeit gains experience and eventually masters his art. The competition between these two antagonists is the origin of the term adversarial in GANs and allows the counterfeit (i.e., generator) to fulfill its function. The analogy aside, in imaging, the purpose of the generator is to generate synthetic images based on a provided input (e.g., another image, text, speech, or video files) [33,36,53,54,55,56,57]. The generator model uses a distribution of data to perform internal comparisons and provide a candidate image close to a real counterpart [33,36,53,54,55,56,57]. The purpose of a discriminator is to differentiate between a real and a synthetic image by calculating the probability that the provided image is real [33,36,53,54,55,56,57]. Feedback from the discriminator to the generator allows model improvement with each classification iteration. Mathematically, the generator can be defined as a function G(z; θg), where z is the input noise vector and θg represents the model parameters [58,59,60]. Similarly, the discriminator is a function D(x; θd), where x represents the input image and θd represents the discriminator parameters. The optimization goal for GANs is a minimax game defined as:
m i n G m a x D E x ~ P d a t a log D x + E z ~ P z log 1 D G z
Here, Pdata represents the distribution of real images, whereas Pz is the prior distribution of the noise vector. The generator aims to minimize the loss while the discriminator maximizes it [33,36,53,54,55,56,57]. Improvement of the generator is usually performed by altering the weights in the nodes of the hidden layers of the model based on provided feedback [33,36,53,54,55,56,57]. This backward adjustment of the weights of the model generator is called backpropagation [33,36,53,54,55,56,57]. The mathematic optimization technique used for weight update is the partial derivatives of the loss function to each weight, which is called gradient descent [33,36,53,54,55,56,57]. The loss function is the function that evaluates the algorithm’s performance when comparing its output to the ground truth [33,36,53,54,55,56,57]. For ophthalmologic applications, the loss function can take various forms, such as pixel-wise mean squared error (MSE) for reconstruction tasks or perceptual loss to enhance the realism of generated OCT images. Specific loss functions tailored to clinical applications, like the Structural Similarity Index (SSIM) loss, have also been employed for improved clinical interpretability [33,36,53,54,55,56,57]. Typical GANs are based on generator and discriminator CNNs; however, DCNNs can also be employed [33,36,53,54,55,56,57]. CNNs are used for image classification and feature extraction, while DCNNs are a deeper version of CNNs with more layers, enabling them to learn more complex and abstract features for tasks like image generation [33,36,53,54,55,56,57].
Translation of GANs to ophthalmology can take various forms. For example, in the case of a GAN built to predict the macular OCT findings in a patient with DR from fundus photos, the role of the generator is to compare provided fundus photos (inputs) to the available training set and learn to generate a new, realistic, and corresponding OCT image [33,36,53,54,55,56,57]. The role of the discriminator is to provide feedback to the generator during training regarding the appropriateness of the generated OCT images (i.e., look realistic) and accuracy (i.e., if they correspond to appropriate fundus photos) [33,36,53,54,55,56,57].

3. Generative Adversarial Network Model Training

GAN training encompasses two main steps: discriminator training and generator training [33,36,53,54,55,56,57]. Usually, GAN generators and discriminators are trained separately, but simultaneous is also possible. This is known as alternate training [33,36,53,54,55,56,57]. In this section, we will delve into the fundamental principles underlying each process.

3.1. Discriminator Training

Training a discriminator is the first step in training GANs. The generator model functionality is kept constant while the functionality of the discriminator is allowed to fluctuate to create an “ideal” discriminator [33,36,53,54,55,56,57]. Input data for discriminator training comes from both real examples (i.e., inputs to classify as positive examples) and from the generator (i.e., inputs to classify as negative examples) [33,36,53,54,55,56,57]. Both data are then classified by the discriminator; their predictions are compared to the ground truth to establish total loss [33,36,53,54,55,56,57]. Through backpropagation, the discriminator is penalized for errors to which the weights in decision-making are altered [33,36,53,54,55,56,57]. Each cycle during which the complete training dataset is fed to the discriminator for training is called an epoch [33,36,53,54,55,56,57]. Through the training process, multiple epochs occur, and with each cycle, the performance of the discriminator improves [33,36,53,54,55,56,57]. Training stops when additional epochs add little value to performance or when the maximum allowed epochs is reached [33,36,53,54,55,56,57]. Batch size, a critical parameter determining the number of samples processed at a time, significantly affects the model’s performance and training stability. Optimizing batch size is non-trivial and requires empirical testing to balance computational efficiency with model accuracy. A computer scientist is best equipped to fine-tune this parameter [33,36,53,54,55,56,57]. In summary, discriminator training involves iterative optimization using real and generated data to improve its ability to classify inputs accurately, with performance stabilizing after sufficient epochs.

3.2. Generator Training

In the training process, the generative model is provided with an input and attempts to produce a plausible image output [33,36,53,54,55,56,57]. As mentioned previously, this output acts as a training example for the discriminator model, which then attempts to distinguish the generator’s synthetic output from a real image [33,36,53,54,55,56,57]. The discriminator penalizes the generator for being recognized [33,36,53,54,55,56,57]. A similar process for the discriminator happens with the generator: weight adjustments through backpropagation allow improvement of the generator, and training stops when each additional epoch adds little value to the performance outcome or when the maximum allowed epochs are reached [33,36,53,54,55,56,57]. Changes to the discriminator through training would make reaching an appropriate endpoint difficult for the generator. Therefore, the discriminator model functionality is kept constant while the generator functionality is altered [33,36,53,54,55,56,57]. In summary, generator training focuses on improving the model’s ability to produce realistic outputs by minimizing loss through iterative updates, aiming for stable convergence (i.e., the point where the loss function is minimized and training parameters reach a stable state, allowing accurate generations).

3.3. Alternate Training

The alternative training approach is now a standard of practice for modern GANs and consists of training the generator and discriminator simultaneously [33,57,61]. In this process, the discriminator may be trained for a set number of epochs, and then the same process is undertaken to train the generator [33,57,61]. This process is repeated a multitude of times until there is minimal improvement in the functionality of the generator [33,57,61]. The advantage of this method is that it theoretically improves the functionality of the end point generator, as both the discriminator and generator are faced with more challenging tasks as training progresses [33,57,61]. However, this can make convergence more challenging [33,57,61]. Learning rate selection, another key parameter, directly influences the convergence speed and stability of training [33,57,61]. A high learning rate may cause instability, while a low rate could lead to slow convergence [33,57,61]. Learning rate schedules or adaptive methods, such as Adam optimizers, are preferred for GAN training. Determining the optimal configuration often requires expertise in machine learning [33,57,61]. As the performance of the discriminator decreases, the generator function improves [33,57,61]. Through the training process, the generator function continues to improve by learning from feedback provided by the discriminator [33,57,61]. As a result, the generator can produce images that closely mimic real images, making them nearly indiscernible to the discriminator and, potentially, to human observers [33,57,61]. At this point, even the most ideal discriminator performance would be no better than chance or 50% [33,57,61]. If the network continues to train, then the generator model will adapt to feedback from the discriminator, which will provide inaccurate feedback [33,57,61]. This phenomenon decreases the overall functionality of the generator model and illustrates an example of failure to converge [33,57,61]. While this method enhances both models, it can complicate convergence due to the adversarial dynamics. Variants of gradient descent, such as RMSProp or AdaGrad, are often employed to stabilize training. The choice of optimization algorithm is highly dependent on the application and data characteristics. This task is best left to experts in machine learning [33,57,61]. In summary, alternate training of GANs improves both models’ performance through simultaneous optimization, but adversarial dynamics can complicate convergence.

3.4. Categories of Input

Inputs serve as a starting point for the generator to produce synthetic images [62]. In the first models, the inputs of GANs were noise images [62]. Through training, the GANs learn to develop a coherent image from that noise [62]. In the field of ophthalmology, such GANs have limited value since noise base inputs are unlikely to provide valuable clinical value. More practical applications of GAN inputs typically include text, image, or video files. During training, these data inputs are paired with target image outputs, with the goal of the GAN being to generate the target image [62]. As with all machine learning programs, there is a significant reliance on the underlying data.

4. Evaluating GANs

4.1. Evaluative Measures

Two terms have been used to describe the “realness” of outputs from GANs: fidelity and diversity [33,56,57,63]. Fidelity often refers to the overall quality of the produced image [33,56,57,63]. This can depend on a variety of features, such as, but not limited to, texture, structure, and detail [56,62,64]. Diversity refers to the variety of images that GANs can produce. GANs that produce only one type of image, or duplicates of a similar image, have limited practical value and thus are considered to have low diversity [56,62,64]. Functional GANs must be able to produce images with both great fidelity and diversity. In machine learning discriminatory problems, total accuracy, sensitivity, and specificity are commonly used to measure the performance of the neural network [56,62,64]. In generative neural networks, the evaluation task is more challenging [56,62,64]. The fundamental issue is that a “real image” is difficult to define [56,62,64]. Therefore, unlike classification tasks, where an algorithm can clearly distinguish between two categories, evaluating whether an image is ’real’ or ’not real’ involves subjective judgment (i.e., human input) and assessment of multiple complex characteristics of the image [56,62,64].

4.2. Qualitative vs. Quantitative Methods (Pixel-Wise Loss)

There are two categories of methods for evaluating GANs: qualitative methods and quantitative methods [56,62,64]. Qualitative means usually include human raters and encompass, but are not limited to, methods that use masked graders to evaluate outputs from GANs [56,62,64]. It is common practice to compare outputs from GANs to their ground truth images. However, such methods are human labor intensive and are subject to bias [56,62,64]. Conversely, quantitative methods do not involve human judgment but can be difficult to implement [56,62,64]. Quantitative methods encompass tools that can compare pixel variations of GAN outputs to their corresponding ground truth counterparts, which needs to be available for comparison [56,62,64]. This pixel-by-pixel evaluation of deviation is called pixel-wise loss. Pixel-wise loss allows for a more objective evaluation of a GAN’s performance. However, this approach is not a flawless metric of GAN performance since simple deviations in pixel values from the ground truth do not directly correlate to realism [33,56,57,63]. A synthetic image may have limited pixel variation from its ground truth counterpart, but incoherency in a small area can question the realism of the entire image [33,56,57,63]. For example, if a GAN generates an image of a human hand with six fingers instead of five, the realism of the entire picture would be questioned. Similarly, if a synthetic image has a large pixel variation but no incoherence in it, the pixel-wise loss will underestimate the realism [33,56,57,63].

4.3. Other Objective Methods

Other objective measures include using an SSIM, the Inception Score (IS), or the Frechet Inception Distance (FID). SSI compares GAN outputs to their ground truth counter parts by looking at features such as luminance, contrast, and structure between the two images [65]. Although this does allow for a more reliable metric than simple pixel-to-pixel variation, this method still does not measure the realness of the image [65]. The IS uses a classification model pre-trained on a large annotated dataset of images (e.g., ImageNET for Inception v3) to evaluate the generated outputs [66,67,68]. The Inception v3 model has become popular since this approach has been highly correlated with human evaluators [66,67,68]. However, an important limitation of the clinical translation of Inception V3 is the highly dependent nature of the model on the database used in pretraining [66,67,68]. Since ImageNET has limited medical images, the model is, unfortunately, sometimes of limited value in medical applications [66,67,68].
FID has also become a popular outcome measure for GANs [33,69]. In this methodology, an Inception v3 model is employed to extract features from both the real and generated images and subsequently generate a distribution [33,69]. The difference between these distributions is known as the FID [33,69]. One advantage of FID is that it can identify mode collapse, unlike IS [33,69].
Overfitting refers to an undesirable behavior of an algorithm that fails to generalize and fits too closely to its training data, resulting in it being unable to correctly perform accurate prediction [57,62,63,64,70]. A helpful analogy to understand the concept is an engineering student who memorized all the physics exercises in his chapter instead of understanding the physics itself. Currently, there is no simple measure to evaluate overfitting in GANs, and all metrics have their own core flaws and no clear gold standard [57,62,63,64,70]. The ideal metric for evaluating the functionality of a GAN should be determined by a multitude of factors, including the network architecture, the nature of the presented data, and the task performed [57,62,63,64,70]. Unfortunately, specific threshold values for retinal imaging cannot be universally standardized due to the inherent variability in imaging modalities and the diverse range of disease entities encountered in clinical practice. Each imaging technique has its own set of parameters and sensitivities that influence threshold determination. Moreover, the threshold values often need to be tailored to the specific pathology under investigation, as different retinal diseases may require distinct diagnostic criteria. Given the complexity of optimizing these thresholds across various modalities and diseases, input from an expert in computer science is essential to develop and fine-tune algorithms that can dynamically adjust and apply these parameters for improved diagnostic precision.

5. Loss Function

Loss functions evaluate models (generator or discriminator) through training [33,56,57,63]. The loss functions not only allow the user to evaluate the performance of a generator but also provide the necessary feedback to the generator model for optimization through backpropagation and gradient descent [33,56,57,63]. These loss functions represent the difference between the distribution of produced data and the distribution of real data [33,56,57,63]. The objective of the training is to build a discriminator whose performance is maximized and who is unable to distinguish between real vs fake data provided by the generator and whose performance is also maximized [33,56,57,63]. There are three primary loss functions that have been employed that we will discuss in the ensuing sections [33,56,57,63].

5.1. Minimax Loss

The first loss function that was described in the paper that introduced GANs by Ian Goodfellow is called minimax loss—we suggest referring to Section 2.1 for its mathematical explanation [53,71]. However, this loss function had a flaw: occasionally, at the beginning of the training, the discriminator performed his task of discriminating synthetic images so well that the generator failed to catch up (i.e., produce images), thus stopping prematurely to training [53,71]. To alleviate the previous issue, modified minimax loss and Wasserstein loss functions were developed [53,71].

5.2. Modified Minimax Loss and Wasserstein Loss

Modified minimax loss attempts to counteract early termination of training by determining the generator’s loss from the probability that a discriminator believes a synthetic image is real rather than comparing the direct distribution of real data points to synthetic ones [57,71,72]. Mathematically, this approach is defined as LG = −Ez∼Pz[log(D(G(z)))] [57,71,72]. Wasserstein loss, often referenced as Wasserstein GAN or wGAN since it is an extension of a classic GAN, attempts to counteract the previous issue with another approach [57,71,72]. The output of a Wasserstein GAN discriminator is different from the output of a classic GAN discriminator [57,71,72]. The first provides an output score of being “real” (higher score = more realistic image), while a classic GAN provides a probability, an odd between 0 and 1, of the image being real [57,71,72]. Mathematically, Wasserstein loss is defined as LG = −Ez∼Pz[D(G(z))]. Wasserstein loss has been demonstrated to decrease the probability of early termination of training [57,71,72]. Thresholds can be created to define whether an image is real or synthetic. The goal of the Wasserstein GAN’s discriminator is to maximize its score accuracy [57,71,72].

5.3. Other Loss Functions

Other variations in loss functions include the use of mean squared loss [57]. For standard GANs, log loss is used in the log function through training [57]. However, there is evidence suggesting that employing mean squared loss increases the probability that synthetic images will approximate their ground truth counterparts [57]. All loss functions can be useful in evaluating the functionality of a GAN; however, the primary function of loss functions is to guide the training process [57].

6. Overview of GANs Applications in Ophthalmology

6.1. Image Quality Improvement

Super-resolution and progressive GANs are specific subtypes of GANs aimed at increasing the resolution of images. However, the structure of these GANs is more dynamic, where resolution is improved progressively through the training process [73,74]. In the field of medical imaging, these types of GANs have applications for improving image quality and decreasing artifacts [62,73,74]. In ophthalmology, image quality improvement can enhance the resolution of images (Figure 3), such as optic disc photography that can be limited by patient cooperation, small pupils, or media opacities [75]. GANs can also improve retinal images taken by a handheld fundus camera in limited imaging conditions [76]. Similarly, in OCT images with low resolution and speckle noise, due to media opacities, which can limit accurate diagnosis, super resolution GANs can output high resolution OCT images allowing better classification [77].

6.2. Inpainting

GANs designed for inpainting are those used to fill-in images with missing pixels or artifacts (Figure 4) [78]. Furthermore, GANs can be applied to generate intermediate synthetic slices in between two imaged-derived slices [78]. This technology could be used to generate thinner slices from an OCT from a thicker imaging protocol or to generate phases of fluorescein angiography that were not captured [78].

6.3. Conditional GANs

cGANs function similarly to standard GANs; however, they differ in that they allow the user to define conditioning variables to alter training [79,80,81]. This means that the GAN can be trained with more information than just the data provided [79,80,81]. Theoretically, a condition can be applied to any variable in training; however, research evaluating the impact of a single status variable, images of multiple domains, masked images, and heat-map-guided images is currently being conducted [79,80,81]. The ability to provide additional information to the GAN allows for more control over the training process and the desired characteristics of the final model [79,80,81].

6.4. Multimodal GANs

Recent applications of GANs have taken advantage of the multimodal inputs. GANs can be used for translation between text, images, speech, and video [62,64,70]. In the medical field, speech-to-text GANs have been implicated in generating clinical notes by recording audio files from clinical interactions [82]. In the field of vision science, video-to-speech GANs have been used to help patients with visual impairment navigate their environment [83].

6.5. Translational GANs

Translational GANs allow for the translation of one input to another output with a defined property [84,85]. The majority of the research body for medical applications of GANs has centered around these types of GANs, with a focus on image-to-image models [84,85]. Image-to-image GANs have been used to convert ultrasound images and CT scans into MRI images [84,85]. In the field of ophthalmology, studies have demonstrated that fundus photos can be translated into corresponding fluorescein autofluorescence images [84,85]. One of the most well-studied translational GANs is CycleGAN due to its simplicity and high performance in generating plausible outputs [84,85]. CycleGAN functions to translate images from a data distribution into images that belong to another data distribution of a differing characteristic [84,85]. Unlike other GANs, CycleGAN has two generator and two discriminator functions [84,85]. In this setup, the output of the first generator is used as the input for the second generator [84,85]. For this reason, CycleGAN does not require labels or pairwise correspondence, as the cycle between the generators allows the model to learn the characteristics of the two separate image domains [84,85]. This makes producing datasets for CycleGAN simpler and allows for the development of networks for denoising without paired datasets [84,85]. For translational image-to-image GANs, loss can be evaluated by comparing the pixels of the synthetic images to their ground truth counterparts [84,85].

7. Current Generative Adversarial Networks Applications with Different Imaging and Functional Testing Modalities in Ophthalmology and Retina

In recent years, numerous studies have sought to develop GANs to generate various imaging results based on the available modalities in ophthalmology. Given the enormous body of literature on the topic, in this section, we outline the most recent and/or most accurate and powerful GANs being developed for potential use in clinical practice based on the various imaging modalities. However, one should be aware that GANs in ophthalmology currently being developed can be applied to various retinal pathologies, mainly AMD, DR, vascular occlusive retinal diseases, and inherited retinal diseases (IRD). The aim of this section is to provide an overview of the clinical potential of GANs in ophthalmology.

7.1. Fundus Autofluorescence

Fundus autofluorescence (FAF) is a non-invasive imaging technique that reduces the risk of potential complications and adverse effects when compared to the FA imaging technique. Limitations of FAF imaging include lower signal strength than FA and greater susceptibility to artifacts from anterior structures [86]. This imaging technique uses the fluorescent proprieties of lipofuscin in the retinal pigment epithelium (RPE) to create an image. The image abnormalities found in FAF can be further subdivided into two groups based on their appearance in autofluorescence imaging: hyper-autofluorescence and hypo-autofluorescence [86].
Wu and coworkers developed a conditional GAN, RA-cGAN, using Adam optimization to generate synthetic FAF images from en-face OCT 200 × 200 × 1024 voxels images for automatic geographic atrophy (GA) segmentation [87]. The synthetic FAF and en-face OCT images were then processed in a DCNN to enhance GA segmentation accuracy. By training the RA-cGAN to translate features from en-face OCT to synthetic FAF images while preserving lesion details, the method was shown to improve GA detection using a spatial fuzzy c-means algorithm, leading to superior performance over previous models. Conversely, Su and coworkers generated synthetic FAF images from color fundus images using a Pix2PixHD GAN method, harnessing both imaging modalities to enhance accurate classification and screening of AMD patients [88]. Furthermore, Veturi and colleagues used a GAN-based framework to address the inherent class imbalance in datasets for rare diseases like IRD, erroneously resulting in deep learning models favoring the more prevalent diseases in 512 × 512 resolution images [89]. The authors generated artificial FAF data for nine different classes of IRDs using a StyleGAN2-ADA model. They further proposed the synthetic dataset-trained deep learning-based IRD classifier as a proxy for real data-trained models.

7.2. Fluorescein Angiography

Fluorescein angiography (FA) is considered the gold standard for retinal and choroidal circulation assessment. To obtain FA images, a fluorescein dye is injected intravenously before capturing images through a fundus camera equipped with a barrier and an excitation filter. The images are captured in multiple phases: choroidal, arterial, arteriovenous, venous, and recirculation [90]. Depending on the fluorescein patterns, abnormalities can be classified into hyperfluorescent and hypofluorescent images [91]. However, in addition to being invasive, this technique has several limitations. Fluorescein is classified as a category C drug and has teratogenous effects, especially in the first trimester [92]. The risk of allergic reaction from mild skin rash up to anaphylactic shock exists [93]. Therefore, the possibility of generating FA images is highly promising and could potentially leverage the challenges and limitations associated with FA in clinical practice.
Many exploratory investigations have shown the potential of utilizing GANs for translating color fundus photographs to synthetic FA images, notably by using LA-GAN [94], LrGAN and HrGANS [95], Sequence GANs [96], a cGAN with a hierarchal setup [97], Fundus2Angio [98], Attention2AngioGAN [99], and VTGAN [100]. Recently, Kamran and coworkers designed the FA4SANS-GAN architecture for color fundus photography to FA translation for the detection of Spaceflight-Associated Neuro-Ocular Syndrome (SANS) [101]. Their deep learning technology significantly outperformed three other state-of-the-art GAN-based models. With the goal of improving DR screening and providing an alternative when fundus FA imaging is not feasible, Shi et al. used a PixtoPixHD to translate color fundus images into venous and late phase FA images [102]. The same team applied this model on FA and color fundus photography paired images to conduct retinal-vessel segmentation in a finetuned single shot manner using a cross-modality soft AV label pretraining method [103]. Recently, Ge and colleagues explored the translation of ultra-wide field scanning laser ophthalmoscopy (UWF-SLO) images into synthetic ultra-wide field FA (UWF-FA) generated by a novel UWAFA-GAN network. Their methodology captures microvascular pathological changes in superior resolution [104]. Conversely, Wang and colleagues proposed a unified cycleGAN and CNN framework for automated grading of DR. The model differentiated non-proliferative (NPDR) and proliferative DR (PDR) based on the ischemic and leakage indexes and assessed severity with UWF-FA, providing accurate categorization [105]. This method was also shown to yield accurate classification in simulated 7-SF, although not as precise as UWF-FA. Finally, Abdelmotaal and colleagues reported the Pix2Pix GAN network’s capacity to synthesize clinically relevant OCT color-coded macular thickness maps from a modest-sized original FA dataset and the reverse process. The network translation implementation aimed to provide clinically useful alternatives to imaging methods for diabetic macular edema patients (NCT05105620) [106]. Overall, the current advances regarding GAN models for FA image generation are a valuable tool for clinical application. Patients with DR constitute a large proportion of patients undergoing FA testing in clinics. These models can limit the associated risks and psychosocial burden associated with conducting an FA assessment in the clinic.

7.3. Indocyanine Green Angiography

Indocyanine green angiography (ICGA) is a widely recognized method for identifying chorioretinal diseases. One key feature of indocyanine is its ability to be visualized through the RPE, lipid exudates, and fluid, making it a better imaging technique to look at the choroid when compared to fluorescein. Indocyanine is also a type of dye with limited leakage from the vessel walls in comparison to fluorescein [107].
Chen and coworkers developed a deep learning algorithm that uses generative adversarial networks for cross-modal data translation and augmentation and assessed its accuracy in classifying AMD [108]. By employing Pix2PixHD GAN, they successfully synthesized high-visual fidelity ICGA early, medium, and late phase images from color fundus images. Their findings indicate that color fundus photography to ICGA translation effectively boosts the accuracy of AMD classification in deep learning studies, dynamically predicts choroidal lesions, and enhances population-based AMD screening. Jiang and colleagues proposed a cGAN-based method to automate linear lesion segmentation in ICGA images [44]. Linear lesions are key indicators of myopic macular degeneration and crucial markers for the progression of high myopia [109]. Additionally, their proposed model outperforms previously studied non-adversarial deep learning networks in lesion segmentation.

7.4. Optical Coherence Tomography

OCT is a non-invasive imaging method that allows for the acquisition of a cross-sectional map of the retina within seconds to allow better characterization of lesions and abnormalities within the different layers of the retina [26]. This imaging technique utilizes a concept known as interferometry to provide images with a resolution of 1–15 microns [110]. Time domain, spectral domain, and swept source are three different types of OCT imaging modalities, which differ according to their scanning speed, among other differences [111]. Depending on the location, pattern, and distribution of the hypo- or hyper-reflective changes in the image, ophthalmological diagnoses can be reached.
In recent years, GANs have demonstrated significant potential in enhancing OCT imaging applications across various domains. Image segmentation is studied to pinpoint pathologic regions within identified areas. Wang et al. introduced a CycleGAN-based approach for lesion segmentation in full-width OCT images, achieving an AUC of 96.94% and a DICE coefficient of 0.83239, with a notable image generation time of just 0.039 s, highlighting the method’s efficiency and accuracy in real-time clinical settings [112]. Menten et al. utilized a counterfactual GAN to generate a synthetic longitudinal time series of retinal layers, providing high-resolution and high-fidelity images for studying retinal aging [113]. With the goal of feature extraction and classification, Sun et al. showed that deep learning models trained on a synthetic balanced dataset generated by StyleGAN2-ADA outperformed those trained on unbalanced datasets for retinal condition classification [114].
Additionally, GANs have been extensively studied for synthesizing realistic ophthalmic scans for enhanced data augmentation. In an attempt to expand the training dataset for OCT image classification and drusen body identification, He et al. used a Least Squares GAN (LSGAN) model, concluding that synthetic unlabeled images could improve CNN classifier performance on limited datasets [115]. Kugelman et al. successfully applied cGANs to generate synthetic images, showing improved OCT chorioretinal boundary segmentation [116]. GANs are also used to eliminate noise that can possibly obscure pathologic features in the images, improving on other deep learning technologies that are limited in their ability to produce high-quality denoised images. Wu and coworkers innovated a GAN-trained strategy named ground-truth OCT (tGT-OCT) for speckle-free imaging [117]. In the realm of image enhancement, based on previously researched SiameseGAN, Mehdizadeh and coworkers developed several GAN-based networks to denoise OCT images, preserving visual texture akin to unaltered OCT images while removing noise that could obscure pathological features [118]. Comprehensive evaluations composed of clinician analysis and qualitative and quantitative assessments concluded that the UNet-PatchGAN/WGAN-MSE network outperformed the previous SiameseGAN. Liang et al. contributed to image enhancement by utilizing a CGAN architecture for resolution enhancement in a micro-OCT system [119]. Similarly, Cheong et al. introduced a DeshadowGAN model to remove shadows created by blood vessels in the optic nerve head in OCT images [120]. Halupka and coworkers’ work on speckle reduction using a WGAN further illustrates the improvements in OCT image quality achievable through GANs [121]. Previously, Ren et al. proposed a CycleGAN-based OCT harmonization method that significantly improved image quality in terms of fidelity, sharpness, and contrast, demonstrating superior performance compared to baseline methods [43].
GANs have also been employed for domain transfer and generalization between imaging modalities. Wu et al. demonstrated the harmonization of a GANSeg system across OCT devices without labeled data, achieving segmentation performance comparable to human graders [122]. Similarly, Chen and coworkers introduced a two-stage CycleGAN-based network to standardize retinal OCT images from different devices, significantly boosting segmentation accuracy without the need for manual labeling [123]. Lazaridis and colleagues showed that cyclical GANs could enhance older TD-OCT images to match the signal quality of modern SD-OCT images, improving the reliability of glaucoma progression predictions and improving the statistical power of the UK Glaucoma Treatment Study [124]. Similarly, Romo-Buchelli and colleagues' work on CycleGANs indicated their effectiveness in addressing image variability across different OCT domains, enhancing the generalizability of segmentation models [125].
In an attempt to find a superior method for the diagnosis of macular edema, Tripathi et al. compared multiple GAN models, identifying the top-performing model and super-optimizing said StyleGAN2 using Particle Swarm Optimization (FID 18.84) [126]. The automated model, in pinpointing biomarkers, allows for standardized treatment planning as it forecasts disease status and progression. Another study highlighted the use of RegGAN against six state-of-the-art GAN models for its prognostic accuracy in predicting retina structural changes post-anti-VEGF treatment for diabetic macular edema (DME), assisting clinicians in treatment planning [127]. Post-intervention prediction was also studied by Lee and coworkers using a cGAN network trained on a dataset of baseline OCT-B scans in addition to FA and ICGA images of neovascular AMD [128]. The addition of the beforementioned images to the OCT-B training set resulted in improved post-treatment OCT images of nAMD. Synthetic post-therapeutic OCT imaging for short-term prediction of the same disease was also generated by Liu and coworkers using a Pix2PixHD method [129]. Furthermore, the Pix2PixHD algorithm was shown to accurately predict short-term response to anti-VEGF therapy in patients with retinal vein occlusion by generating synthetic post-therapeutic OCT images [130].

7.5. Optical Coherence Tomography-Angiography

OCT-A is also a non-invasive imaging modality utilized to image the microvasculature of the retina and choroid. The technique relies on performing multiple OCT scans one after the other, using the backscattering of light to generate high-quality images. The difference in motion-contrast between each OCT scan is the key point necessary for creating three-dimensional (3D) images. OCT-A images allow visualization of blood flow in the retinal microvasculature, enabling the examination of the choroidal, retinal, and optic nerve blood vessels without requiring dye injections [131]. Other advantages of OCT-A are its reproducibility, as it does not require a skilled operator, and imaging speed since it can produce an image in approximately six seconds [132]. Despite the many advantages of OCTA, this imaging modality has a few limitations. These include its small field of view, the impossibility of accurately displaying leakage, and the possibility of image artifacts from patient movement [133].
Badhon and colleagues leveraged the relationship between OCT and OCT-A imaging to produce translated OCTAs (TR-OCTA) exclusively from OCT data [134]. This 3D-GAN-based framework is proposed to facilitate more accessible characterization of retinal features, which are traditionally limited to the use of costly OCTA devices. Coronado et al. introduced a novel approach to extract detailed retinal perfusion data using fundus images alone by utilizing a cGAN to synthesize en-face 45-degree OCT-A images from local patches of paired fundus and OCT-A images [135]. GANs can also be applied to OCT-angiography to combat the common stripe artifacts and low contrast, limiting the imaging modalities’ diagnostic precision. Cao and colleagues developed a promising enhancement framework constituting a Perceptual Structure Generative Adversarial Network (PS-GAN) to re-enhance synthesized OCTA images previously de-striped by a Stripe Removal Net (SR-NET), preserving vascular integrity [136]. Jiang et al. compared deep learning models, including Pix2PixGAN, for OCTA reconstruction [137].

7.6. Electroretinogram

ERG is a non-invasive test used to diagnose various retinal pathologies by measuring the electrical activity of the retina in response to different light stimuli. Multiple parameters are used to interpret an ERG; the waveform components are separated into four components: the a-wave reflects the outer retinal function (corresponding to the early hyperpolarization of the rod and cone photoreceptors), while the b-wave reflects the phototransduction activity corresponding to the positive deflection following the a-wave. Oscillatory potentials reflect the inner synaptic retinal feedback circuits as well as vascular function, and photopic negative response corresponds to the response of the retinal ganglion cells. Those waveforms are then analyzed based on amplitude, implicit time, and latency; then, the b-wave to a-wave ratio is calculated [138]. The full-field ERG reflects the overall functionality of the retina in comparison to a more recent electrophysiologic test called the multi-focal ERG that detects distinct areas of outer retinal damage [139]. One of the main limitations of full-field ERG is the possibility of missing focal retinal disease. This limitation can generally be overcome by using focal ERG or multifocal ERG [138].
In their proof-of-concept study, Kulyabin and colleagues employed cGANs to generate synthetic ERG data from ground truth ERG waveform signals, focusing on increasing the sample size for underrepresented classes to balance and augment datasets [140]. They established that the cGAN architecture could artificially create waveforms of different shapes with characteristics closely mimicking those found in authentic temporal data, including an a-wave, OPs, and b-wave. By increasing the sample size of minority classes, the study showed improvements in classification accuracy, supporting the development of robust models for diverse clinical conditions. Applying machine learning and signal analysis to ERGs may prove particularly beneficial for improving the statistical accuracy of the classification of rare and complex disorders like IRDs and in neurodegenerative and neurodevelopmental conditions that present with variable clinical phenotypes.

7.7. Visual Fields

Although standard automated perimetry is the gold standard for assessing visual function, the modality proves challenging with variable results for higher-stage disease or the presence of ophthalmic comorbidities [141]. Although not yet applied to vitreoretinal pathologies, GANs’ accuracy for the modeling of visual field (VF) results has been demonstrated in glaucoma. Hussain and coworkers developed a novel approach to model expected glaucoma progression and VF loss by synthesizing future OCT images using a Pix2PixGAN-based model conditioned on baseline images to predict changes in VF mean deviation 12 months after the initial patient visit. This AI-based approach provides an alternative to glaucoma diagnosis without static perimetry, reducing examination time. Numerous vitreoretinal conditions can lead to VF defects, such as retinal vascular occlusions, retinal detachments, certain retinitis, and macular lesions. Therefore, the possibility of modeling tools to predict the impact of a vitreoretinal disease on VF is highly sought after. For example, projected VF progression in conditions involving retinal vascular occlusions as well as inherited retinal diseases may be helpful in predicting patient outcomes and the impact of medical therapy on visual potential.

8. Future Perspectives

With their introduction to ophthalmology, GANs have the potential to change and improve multimodal imaging for vitreoretinal surgeons. With reliable GANs, numerous uses can be extrapolated in clinical practice to generate higher quality images, generate complementing investigations in centers with limited resources, or for teaching purposes. GANs have also been described as great learning tools for trainees. Given the importance of pattern recognition in ophthalmology and especially retinal subspecialties, the use of high-fidelity generative models to generate a vast number of realistic images of various pathologies has the potential to improve the quality and quantity of learning opportunities. This concept could also be expanded not only to fundus imaging but to all associated elements of multimodal imaging and functional testing, including OCT, OCTA, FA, and VF. Progressive GANs are designed to improve image resolution progressively during training. They start by generating low-resolution images and systematically refine details in subsequent training phases. These models have been successfully used to enhance the quality of retinal photographs taken with handheld cameras. Inpainting-specialized GANs are customized to complete images by filling in missing pixels, effectively restoring photos by replacing obscured regions. Furthermore, imaging applications can generate synthetic intermediary slices between two captured slices. This capability could produce thinner slices from OCT scans originally obtained with thicker imaging protocols or generate phases of fluorescein angiography not initially captured. These advancements have the potential to improve diagnostic accuracy and comprehensiveness in medical imaging practices. Although this cannot replace an actual good quality picture, the ability of generative models to optimize imaging modalities allows for mitigation for media opacities affecting their quality (surface disease, cataract, vitreous opacities, etc. GANs have also been shown to be effective for multimodal imaging translation as well as lesion-region segmentation, which could allow for centers with limited resources to make up for outdated devices, low-quality scans, and lack of data points, reduce expenses, and optimize segmentations. Flash ERG and multifocal ERG represent crucial functional imaging modalities for rare and complex retinal pathologies by measuring the electrical response of the retina to light stimulus. However, their use remains limited by the rarity of the disease requiring such investigations, as well as the limited centers and technicians able to perform them. A future application of GAN models would allow for functional imaging from structural imaging translation. This would allow for the generation of ERGs from otherwise readily available imaging modalities, including OCT or fundus photographs.

9. Limitations

GANs encounter limitations in requiring substantial data input for the generator to produce convincing images that can deceive the discriminator effectively. Data scarcity poses a significant challenge to implementing GANs in medical imaging, which is critical for modern retinal specialists. Class imbalance also presents a notable issue affecting the accuracy of various neural network architectures. When one class in the training dataset contains significantly more data points than others, the network tends to prioritize learning from the majority class. This can lead to strong performance in majority-class instances but poor performance in minority-class instances. This imbalance is particularly problematic in classification tasks and extends to GANs as well. In non-GAN models, it is advisable to evaluate metrics beyond overall accuracy to ensure that the network’s performance is not adversely affected by class imbalance, thus providing a more comprehensive assessment of model effectiveness. At the heart of machine learning applications is the process of creating neural networks by feeding them data. This concept applies to GANs as well. However, unlike basic DCNNs and CNNs, these networks are highly dependent on their data source. If there is an imbalance between the number of class datapoints, a GAN can become highly functional in that particular class but of limited clinical value for cases outside the presented class. An example of this would be developing a GAN trained on patients for NPDR. Such a GAN would have limited functionality in cases of NPDR. In addition, compared to traditional programming, where distinct rules are written by developers with a transparent understanding of the underlying code, the inner workings of neural networks generated by machine learning approaches are often unclear. In these models, the computational reasoning by the neural network is often a mystery. This is often called the “black box” effect. Attempts have been made to open the black box with heat mapping in DCNNs and feature importance in ANNs. However, minimal successful attempts have been made to open the black box in GANs. Furthermore, the structure of a GAN is that of a competitive network where a discriminator function and a generator function compete to outperform one another. This typically results in improving the performance of both networks. However, if the generator model begins to outperform the discriminator, then the discriminator may provide poor feedback to the generator, decreasing its performance. Furthermore, if a generator learns to produce images from one class very well but fails with other classes, then it will consistently obtain positive feedback from the discriminator for producing one image class and negative feedback for producing images for other classes. As a result, the generative model will only produce images for one image class. For example, if a GAN designed to generate FAs is given a dataset of fundus photos for both NPDR and PDR consistently fails at producing appropriate FAs for PDR and succeeds at generating images for NPDR, then the generator model may choose not to produce PDR FAs even when appropriate. Fundamentally, the competitive design of these networks allows for improving performance through training cycles; however, such a structure can still result in error.
Importantly, GANs are inherently challenging to evaluate. Unlike classification problems where specificity, sensitivity, and total accuracy can be measured and compared easily, the output of GANs is an image judged based on realism. Such a term is poorly defined, and many have struggled to develop a standard agreed-upon metric for evaluating GANs. With the advent of AI, we have seen means of determining patients’ sex, age, and blood pressure with a simple picture of their respective fundus. As such, this could, in turn, limit the usability of numerous scans as the data would then need to be encrypted or protected. With the increased interest in AI’s implementation in medical imaging, there are rising concerns for patient data privacy protection.

10. Tackling the Challenges Associated with Generative Adversarial Networks

10.1. Vanishing Gradients

Vanishing gradients are observed during model training when the gradients of the loss function used in gradient descent become very small, resulting in minimal updates of weights in the neural network during back propagation [53]. This slows the training process and can result in the early termination of generative network training [53]. In GANs, vanishing gradients can be observed if the discriminatory model is too effective: the generative model will always receive negative feedback, and a functional generative network will not be created as the network's training will terminate early [53]. A simple approach to prevent vanishing gradients in GANs includes the use of Wasserstein loss and modified minmax loss [142,143].

10.2. Failure to Converge

As described earlier, if the functionality of the generator model surpasses that of the discriminator model, then the generator model may receive inappropriate feedback. In order to counter this, a multitude of techniques have been employed [69]. Similar to vanishing gradients, the use of Wasserstein loss and modified minmax loss can be used to prevent failure to converge [142,143]. If failure to converge happens with Wasserstein loss or other models, one could include both the generator model outputs and known false negatives in discriminator training to provide the discriminator with a consistent flow of negative results [142,143]. Furthermore, regularization schemes can be used: they penalize the weights from the discriminator function as the functionality of the generator improves over the discriminator [144].

10.3. Mode Collapse

An ideal GAN is expected to produce a variety of images, and not only a few that are non-diversified [58,145]. Failure to achieve that capability can occur due to model collapse [58,145]. Model collapse can occur when the generative model, trying to outperform the discriminatory model, produces an output that is very close to ground truth [58,145]. If the discriminator is unable to discriminate between the fake and real image and the generator continues to present this same (or too similar) image over and over, only a few non-diverse images will be produced by the GAN [58,145]. For example, if a GAN developed to perform data augmentation by producing multiple variants of a rare form of retinal malignant tumor produces, after complete training, only a set of ten identical new images of the rare retinal malignancy, then we can assume that the model collapse happened to the model. To avoid undiversified output from a generative model, a large, diverse training and testing set should be employed [62,64,146,147]. Furthermore, the two following methods can be used [62,64,146,147]. First, using Wasserstein loss allows the discriminator to reject discriminators’ solutions on which the generators want overfit [62,64,146,147]. This rejection forces the generator to propose new outputs [62,64,146,147]. Second, some GANs, called unrolled GANs, employ more than a single discriminator network, which prevents the generative model from overfitting to a single discriminator [62,64,146,147].

10.4. Class Imbalance

Class imbalance occurs when a neural network is not exposed enough to one or several subclasses of the dataset due to insufficient variety in the input provided and learns to be biased towards the dominant class provided [69]. The model performs well when dealing with cases from the majority class and poorly when dealing with cases from the minority class [69]. Class imbalance significantly reduces the accuracy of multiple neural network architectures and can be a challenge in the medical field, where some diseases and conditions are rare [69]. In the case of a GAN built for predicting FA results, if the training set is overly sampled with cases of NPDR, then the output results of PDR would likely be few and of poor quality [69]. To avoid this, data should be collected as equally as possible for each of the target output classes. In non-GAN models, metrics other than total accuracy should be evaluated to ensure a network is not hampered by class imbalance [69]. A very interesting and important fact is that a major utility of GAN is data augmentation, which is to artificially create new data from available data with the objective of training AI models [69]. In other words, in the example of FA, if a dataset has few cases of PDR, a first GAN can perform data augmentation and generate synthetic images of PDR, which will serve a second GAN that will potentially predict FA results more accurately, minimizing class imbalance [69].

10.5. Unintended Bias

Unintended bias refers to learning unintentionally acquired by a neural network that negatively impacts its performance [148,149,150]. Unintended bias can occur in multiple steps of training, including during data input and data evaluation [148,149,150]. Bias is very similar to class imbalance. Data input has either underrepresented parts or has intrinsic biases if there is a discrepancy between training input and actual input for which the data are evaluated [148,149,150]. For example, a model trained for a specific retinal malignancy in today’s population may perform differently if used in 40 years in a population where the same specific retinal malignancy is manifesting itself differently due to a change in risk factors and exposures. Bias may occur when sample image datasets are small and without appropriate spatial alignment, in which case GANs may struggle to learn spatial alignment accurately. For this reason, spatial distortions are often incorporated into GAN outputs [148,149,150].

10.6. Hyperparameter Sensitivity

In machine learning, parameters refer to internal variables of the model that fluctuate and are determined through a training session [151]. In contrast, hyperparameters refer to variables that determine the neural network’s structure and need to be defined before training [151]. Hyperparameters include the number of nodes in the hidden layers, the learning rate, the number of iterations of training, etc. Some GANs can be sensitive, and small changes in hyperparameters can have significant implications for the model’s performance [151]. A GAN developer should be aware that slight variations in hyperparameters could drastically affect performance [151]. Thus, determining hyperparameters by trial and error is not always a good approach [151].

10.7. Data Dependence

Data dependence is the reliance of a neural network's performance on the underlying data [152,153]. As machine learning is based on developing an algorithm in response to data, models can be highly susceptible to poor data quality [152,153]. This is often more so in more complicated models. GANs are typically more data-dependent than simpler models such as DCNNs, CNN, and ANNs [152,153]. For example, an ANN may be able to function with relatively high accuracy with a 10% mislabeling rate in the training data [154,155]. However, such an error rate in a GAN training set may hinder performance significantly [154,155].

10.8. Summary

Vanishing gradients, failure to converge, mode collapse, class imbalance, unintended bias, hyperparameter sensitivity, and data dependence are key challenges in training GANs. Strategies such as using Wasserstein loss, unrolled GANs, diverse training data, and regularization can mitigate these issues. Additionally, addressing biases, hyperparameter tuning, and ensuring high-quality data are critical for optimizing GAN performance and avoiding unintended consequences. While physicians should be aware of these challenges, their resolution ideally requires collaboration with experts in machine learning and neural network optimization to ensure robust and reliable outcomes in medical AI applications.

11. Conclusions

GANs have significant potential in the field of medicine, mainly fields like medical and surgical retina, where imaging modalities play a key role in the diagnosis and prognosis of disease. The structure of GANs differs from that of typical machine learning programs used in the field, and, therefore, the flaws associated with these networks need to be acknowledged. There is no one-size-fits-all approach to determining the structures of GANs; therefore, the structure of a GAN implemented in a project should be determined on a case-by-case basis. The primary variables to consider in developing a GAN can include the data structure, the CNN or DCNN structure, the loss function, and the evaluative metrics. There are a multitude of GAN architectures that can be used in conjunction with each other to build a custom GAN for a specific application. Still, the most significant challenges associated with GANs are data collection and quality. Without an appropriate training dataset, a GAN is unlikely to succeed. Therefore, before a GAN is developed, researchers should be aware of their own data limitations. Furthermore, the differing structure of GANs results in differing challenges and limitations compared to other machine learning models. Awareness of these limitations is essential to avoid these limitations. One key limitation is in the evaluation of these networks as there has yet to be a gold standard metric to judge the quality and realism of GANs outputs. Despite this, GANs have significant promise in medical imaging, and it is likely that such technologies will be implemented in the field of ophthalmology in the near future.

Author Contributions

Conceptualization, R.R. and M.K.; writing—original draft preparation, R.R., A.S., A.A. and A.H.; writing—review and editing, M.K., A.V., S.K.A. and C.X.Q.; visualization, M.K.; supervision, C.X.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bohr, A.; Memarzadeh, K. The rise of artificial intelligence in healthcare applications. In Artificial Intelligence in Healthcare; Elsevier: Amsterdam, The Netherlands, 2020; pp. 25–60. [Google Scholar]
  2. Ahuja, A.S. The impact of artificial intelligence in medicine on the future role of the physician. PeerJ 2019, 7, e7702. [Google Scholar] [CrossRef] [PubMed]
  3. Aloysius, N.; Geetha, M. A review on deep convolutional neural networks. In Proceedings of the 2017 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 6–8 April 2017; pp. 0588–0592. [Google Scholar]
  4. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 6999–7019. [Google Scholar] [CrossRef] [PubMed]
  5. Gu, J.; Wang, Z.; Kuen, J.; Shahroudy, L.M.A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; Chen, T. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  6. Zou, J.; Han, Y.; So, S.S. Overview of artificial neural networks. Artif. Neural Netw. Methods Appl. 2009, 458, 14–22. [Google Scholar]
  7. Yegnanarayana, B. Artificial Neural Networks; PHI Learning Pvt. Ltd.: Delhi, India, 2009. [Google Scholar]
  8. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  9. Myles, A.J.; Feudale, R.N.; Liu, Y.; Woody, N.A.; Brown, S.D. An introduction to decision tree modeling. J. Chemom. A J. Chemom. Soc. 2004, 18, 275–285. [Google Scholar] [CrossRef]
  10. Çelik, E.; İnan, E. Artificial Intelligence in Ophthalmology Clinical Practices. Izmir Democr. Univ. Health Sci. J. 2023, 6, 445–459. [Google Scholar] [CrossRef]
  11. Chaurasia, A.K.; Greatbatch, C.J.; Hewitt, A.W. Diagnostic accuracy of artificial intelligence in glaucoma screening and clinical practice. J. Glaucoma 2022, 31, 285–299. [Google Scholar] [CrossRef]
  12. Ting DS, W.; Peng, L.; Varadarajan, A.V.; Keane, P.A.; Burlina, P.M.; Chiang, M.F.; Schmetterer, L.; Pasquale, L.R.; Bressler, N.M.; Webster, D.R.; et al. Deep learning in ophthalmology: The technical and clinical considerations. Prog. Retin. Eye Res. 2019, 72, 100759. [Google Scholar] [CrossRef]
  13. Padhy, S.K.; Takkar, B.; Chawla, R.; Kumar, A. Artificial intelligence in diabetic retinopathy: A natural step to the future. Indian J. Ophthalmol. 2019, 67, 1004–1009. [Google Scholar]
  14. Cleland, C.R.; Rwiza, J.; Evans, J.R.; Gordon, I.; MacLeod, D.; Burton, M.J.; Bascaran, C. Artificial intelligence for diabetic retinopathy in low-income and middle-income countries: A scoping review. BMJ Open Diabetes Res. Care 2023, 11, e003424. [Google Scholar] [CrossRef]
  15. Rajalakshmi, R. The impact of artificial intelligence in screening for diabetic retinopathy in India. Eye 2020, 34, 420–421. [Google Scholar] [CrossRef]
  16. Chen, J.S.; Coyner, A.S.; Ostmo, S.; Sonmez, K.; Bajimaya, S.; Pradhan, E.; Valikodath, N.; Cole, E.D.; Al-Khaled, T.; Chan RV, P.; et al. Deep Learning for the Diagnosis of Stage in Retinopathy of Prematurity. Ophthalmol. Retin. 2021, 5, 1027–1035. [Google Scholar] [CrossRef] [PubMed]
  17. Leng, X.; Shi, R.; Wu, Y.; Zhu, S.; Cai, X.; Lu, X.; Liu, R. Deep learning for detection of age-related macular degeneration: A systematic review and meta-analysis of diagnostic test accuracy studies. PLoS ONE 2023, 18, e0284060. [Google Scholar] [CrossRef] [PubMed]
  18. Nagasato, D.; Tabuchi, H.; Ohsugi, H.; Masumoto, H.; Enno, H.; Ishitobi, N.; Sonobe, T.; Kameoka, M.; Niki, M.; Hayashi, K.; et al. Deep Neural Network-Based Method for Detecting Central Retinal Vein Occlusion Using Ultrawide-Field Fundus Ophthalmoscopy. J. Ophthalmol. 2018, 2018, 1875431. [Google Scholar] [CrossRef] [PubMed]
  19. Ren, X.; Feng, W.; Ran, R.; Gao, Y.; Lin, Y.; Fu, X.; Tao, Y.; Wang, T.; Wang, B.; Ju, L.; et al. Artificial intelligence to distinguish retinal vein occlusion patients using color fundus photographs. Eye 2023, 37, 2026–2032. [Google Scholar] [CrossRef]
  20. Chen, Q.; Yu, W.H.; Lin, S.; Liu, B.S.; Wang, Y.; Wei, Q.J.; He, X.X.; Ding, F.; Yang, G.; Chen, Y.X.; et al. Artificial intelligence can assist with diagnosing retinal vein occlusion. Int. J. Ophthalmol. 2021, 14, 1895–1902. [Google Scholar] [CrossRef]
  21. Cai, L.; Hinkle, J.W.; Arias, D.; Gorniak, R.J.; Lakhani, P.C.; Flanders, A.E.; Kuriyan, A.E. Applications of Artificial Intelligence for the Diagnosis, Prognosis, and Treatment of Age-related Macular Degeneration. Int. Ophthalmol. Clin. 2020, 60, 147–168. [Google Scholar] [CrossRef]
  22. Bogunović, H.; Mares, V.; Reiter, G.S.; Schmidt-Erfurth, U. Predicting treat-and-extend outcomes and treatment intervals in neovascular age-related macular degeneration from retinal optical coherence tomography using artificial intelligence. Front. Med. 2022, 9, 958469. [Google Scholar] [CrossRef]
  23. Schmidt-Erfurth, U.; Waldstein, S.M.; Klimscha, S.; Sadeghipour, A.; Hu, X.; Gerendas, B.S.; Osborne, A.; Bogunovic, H. Prediction of Individual Disease Conversion in Early AMD Using Artificial Intelligence. Investig. Ophthalmol. Vis. Sci. 2018, 59, 3199. [Google Scholar] [CrossRef]
  24. Prahs, P.; Radeck, V.; Mayer, C.; Cvetkov, Y.; Cvetkova, N.; Helbig, H.; Märker, D. OCT-based deep learning algorithm for the evaluation of treatment indication with anti-vascular endothelial growth factor medications. Graefes Arch. Clin. Exp. Ophthalmol. 2018, 256, 91–98. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, C.; Cheng, J.; Li, C.; Tian, Q. Image-Specific Classification with Local and Global Discriminations. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 4479–4486. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, T.C.; Liu, M.Y.; Zhu, J.Y.; Tao, A.; Kautz, J.; Catanzaro, B. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8798–8807. [Google Scholar]
  27. Li, Z.; Xia, B.; Zhang, J.; Wang, C.; Li, B. A comprehensive survey on data-efficient GANs in image generation. arXiv 2022, arXiv:220408329. [Google Scholar]
  28. Osokin, A.; Chessel, A.; Carazo Salas, R.E.; Vaggi, F. GANs for biological image synthesis. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2233–2242. [Google Scholar]
  29. Liu, S.; Wang, T.; Bau, D.; Zhu, J.Y.; Torralba, A. Diverse image generation via self-conditioned gans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14286–14295. [Google Scholar]
  30. Siarohin, A.; Sangineto, E.; Lathuiliere, S.; Sebe, N. Deformable gans for pose-based human image generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3408–3416. [Google Scholar]
  31. Harshvardhan, G.; Gourisaria, M.K.; Pandey, M.; Rautaray, S.S. A comprehensive survey and analysis of generative models in machine learning. Comput. Sci. Rev. 2020, 38, 100285. [Google Scholar]
  32. Adams, L.C.; Busch, F.; Truhn, D.; Makowski, M.R.; Aerts, H.J.W.L.; Bressem, K.K. What Does DALL-E 2 Know About Radiology? J. Med. Internet Res. 2023, 25, e43110. [Google Scholar] [CrossRef]
  33. You, A.; Kim, J.K.; Ryu, I.H.; Yoo, T.K. Application of generative adversarial networks (GAN) for ophthalmology image domains: A survey. Eye Vis. 2022, 9, 6. [Google Scholar] [CrossRef]
  34. Niu, Y.; Wang, Y.D.; Mostaghimi, P.; Swietojanski, P.; Armstrong, R.T. An innovative application of generative adversarial networks for physically accurate rock images with an unprecedented field of view. Geophys. Res. Lett. 2020, 47, e2020GL089029. [Google Scholar] [CrossRef]
  35. Chen, Y.; Shi, F.; Christodoulou, A.G.; Xie, Y.; Zhou, Z.; Li, D. Efficient and accurate MRI super-resolution using a generative adversarial network and 3D multi-level densely connected network. In Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2018; pp. 91–99. [Google Scholar]
  36. Wolterink, J.M.; Mukhopadhyay, A.; Leiner, T.; Vogl, T.J.; Bucher, A.M.; Išgum, I. Generative Adversarial Networks: A Primer for Radiologists. RadioGraphics 2021, 41, 840–857. [Google Scholar] [CrossRef]
  37. Lim, S.; Nam, H.; Shin, H.; Jeong, S.; Kim, K.; Lee, Y. Noise Reduction for a Virtual Grid Using a Generative Adversarial Network in Breast X-ray Images. J. Imaging 2023, 9, 272. [Google Scholar] [CrossRef]
  38. Sun, Y.; Liu, X.; Cong, P.; Li, L.; Zhao, Z. Digital radiography image denoising using a generative adversarial network. J. X-Ray Sci. Technol. 2018, 26, 523–534. [Google Scholar] [CrossRef]
  39. Lyu, Q.; Wang, G. Conversion between CT and MRI images using diffusion and score-matching models. arXiv 2022, arXiv:220912104. [Google Scholar]
  40. Kadambi, S.; Wang, Z.; Xing, E. WGAN domain adaptation for the joint optic disc-and-cup segmentation in fundus images. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1205–1213. [Google Scholar] [CrossRef]
  41. Yang, J.; Dong, X.; Hu, Y.; Peng, Q.; Tao, G.; Ou, Y.; Cai, H.; Yang, X. Fully Automatic Arteriovenous Segmentation in Retinal Images via Topology-Aware Generative Adversarial Networks. Interdiscip. Sci. 2020, 12, 323–334. [Google Scholar] [CrossRef]
  42. Khan, Z.K.; Umar, A.I.; Shirazi, S.H.; Rasheed, A.; Qadir, A.; Gul, S. Image based analysis of meibomian gland dysfunction using conditional generative adversarial neural network. BMJ Open Ophthalmol. 2021, 6, e000436. [Google Scholar] [CrossRef] [PubMed]
  43. Tavakkoli, A.; Kamran, S.A.; Hossain, K.F.; Zuckerbrod, S.L. A novel deep learning conditional generative adversarial network for producing angiography images from retinal fundus photographs. Sci. Rep. 2020, 10, 21580. [Google Scholar] [CrossRef] [PubMed]
  44. Liu, Y.; Yang, J.; Zhou, Y.; Wang, W.; Zhao, J.; Yu, W.; Zhang , D.; Ding, D.; Li, X.; Chen, Y. Prediction of OCT images of short-term response to anti-VEGF treatment for neovascular age-related macular degeneration using generative adversarial network. Br. J. Ophthalmol. 2020, 104, 1735–1740. [Google Scholar] [CrossRef] [PubMed]
  45. Helm, J.M.; Swiergosz, A.M.; Haeberle, H.S.; Karnuta, J.M.; Schaffer, J.L.; Krebs, V.E.; Spitzer, A.I.; Ramkumar, P.N. Machine learning and artificial intelligence: Definitions, applications, and future directions. Curr. Rev. Musculoskelet. Med. 2020, 13, 69–76. [Google Scholar] [CrossRef]
  46. Packin, N.G.; Lev-Aretz, Y. Learning algorithms and discrimination. In Research Handbook on the Law of Artificial Intelligence; Edward Elgar Publishing: Cheltenham, UK, 2018; pp. 88–113. [Google Scholar]
  47. Castelli, M.; Manzoni, L. Generative models in artificial intelligence and their applications. Appl. Sci. 2022, 12, 4127. [Google Scholar] [CrossRef]
  48. Jin, L.; Tan, F.; Jiang, S. Generative Adversarial Network Technologies and Applications in Computer Vision. Comput. Intell. Neurosci. 2020, 2020, 1459107. [Google Scholar] [CrossRef]
  49. Munro, P. Backpropagation. In Encyclopedia of Machine Learning; Sammut, C., Webb, G.I., Eds.; Springer: Boston, MA, USA, 2010; p. 73. [Google Scholar] [CrossRef]
  50. Kornblith, S.; Chen, T.; Lee, H.; Norouzi, M. Why do better loss functions lead to less transferable features? Adv. Neural Inf. Process. Syst. 2021, 34, 28648–28662. [Google Scholar]
  51. Komatsuzaki, A. One epoch is all you need. arXiv 2019, arXiv:190606669. [Google Scholar]
  52. Ying, X. An overview of overfitting and its solutions. J. Phys. Conf. Ser. 2019, 1168, 022022. [Google Scholar] [CrossRef]
  53. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  54. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. Available online: https://arxiv.org/abs/1406.2661 (accessed on 4 September 2024).
  55. Feizi, S.; Farnia, F.; Ginart, T.; Tse, D. Understanding GANs: The LQG Setting. arXiv 2017, arXiv:1710.10793. Available online: https://arxiv.org/abs/1710.10793 (accessed on 4 September 2024).
  56. Sorin, V.; Barash, Y.; Konen, E.; Klang, E. Creating Artificial Images for Radiology Applications Using Generative Adversarial Networks (GANs)—A Systematic Review. Acad. Radiol. 2020, 27, 1175–1185. [Google Scholar] [CrossRef] [PubMed]
  57. Wang, Z.; Lim, G.; Ng, W.Y.; Keane, P.A.; Campbell, J.P.; Tan, G.S.W.; Schmetterer , L.; Wong , T.Y.; Liu, Y.; Ting, D.S.W. Generative adversarial networks in ophthalmology: What are these and how can they be used? Curr. Opin. Ophthalmol. 2021, 32, 459–467. [Google Scholar] [CrossRef] [PubMed]
  58. Arjovsky, M.; Bottou, L. Towards Principled Methods for Training Generative Adversarial Networks. arXiv 2017, arXiv:1701.04862. Available online: https://arxiv.org/abs/1701.04862 (accessed on 4 September 2024).
  59. Little, C.; Elliot, M.; Allmendinger, R.; Samani, S.S. Generative adversarial networks for synthetic data generation: A comparative study. arXiv 2021, arXiv:211201925. [Google Scholar]
  60. Chavdarova, T.; Fleuret, F. SGAN: An Alternative Training of Generative Adversarial Networks. arXiv 2017, arXiv:1712.02330. Available online: https://arxiv.org/abs/1712.02330 (accessed on 4 September 2024).
  61. Pan, Z.; Yu, W.; Wang, B.; Xie, H.; Sheng, V.S.; Lei, J. Loss Functions of Generative Adversarial Networks (GANs): Opportunities and Challenges. IEEE Trans. Emerg. Top Comput. Intell. 2020, 4, 500–522. [Google Scholar] [CrossRef]
  62. Borji, A. Pros and cons of GAN evaluation measures. Comput. Vis. Image Underst. 2019, 179, 41–65. [Google Scholar] [CrossRef]
  63. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  64. Barratt, S.; Sharma, R. A Note on the Inception Score. arXiv 2018, arXiv:1801.01973. Available online: https://arxiv.org/abs/1801.01973 (accessed on 4 September 2024).
  65. Salimans, T.; Goodfellow, I.; Zaremba, W.; Cheung, V.; Radford, A.; Chen, X. Improved Techniques for Training GANs. arXiv 2016, arXiv:1606.03498. Available online: https://arxiv.org/abs/1606.03498 (accessed on 4 September 2024).
  66. Lucic, M.; Kurach, K.; Michalski, M.; Gelly, S.; Bousquet, O. Are GANs Created Equal? A Large-Scale Study. arXiv 2017, arXiv:1711.10337. Available online: https://arxiv.org/abs/1711.10337 (accessed on 4 September 2024).
  67. Brock, A.; Donahue, J.; Simonyan, K. Large Scale GAN Training for High Fidelity Natural Image Synthesis. arXiv 2018, arXiv:1809.11096. Available online: https://arxiv.org/abs/1809.11096 (accessed on 4 September 2024).
  68. Soloveitchik, M.; Diskin, T.; Morin, E.; Wiesel, A. Conditional Frechet Inception Distance. arXiv 2021, arXiv:2103.11521. Available online: https://arxiv.org/abs/2103.11521 (accessed on 4 September 2024).
  69. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein GAN. arXiv 2017, arXiv:1701.07875. Available online: https://arxiv.org/abs/1701.07875 (accessed on 4 September 2024).
  70. Frogner, C.; Zhang, C.; Mobahi, H.; Araya-Polo, M.; Poggio, T. Learning with a Wasserstein Loss. arXiv 2015, arXiv:1506.05439. Available online: https://arxiv.org/abs/1506.05439 (accessed on 4 September 2024).
  71. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. arXiv 2016, arXiv:1609.04802. Available online: https://arxiv.org/abs/1609.04802 (accessed on 4 September 2024).
  72. Gupta, R.; Sharma, A.; Kumar, A. Super-Resolution using GANs for Medical Imaging. Procedia Comput. Sci. 2020, 173, 28–35. [Google Scholar] [CrossRef]
  73. Ha, A.; Sun, S.; Kim, Y.K.; Lee, J.; Jeoung, J.W.; Kim, H.C.; Park, K.H. Deep-learning-based enhanced optic-disc photography. PLoS ONE 2020, 15, e0239913. [Google Scholar] [CrossRef]
  74. Fu, J.; Cao, L.; Wei, S.; Xu, M.; Song, Y.; Li, H.; You, Y. A GAN-based deep enhancer for quality enhancement of retinal images photographed by a handheld fundus camera. Adv. Ophthalmol. Pract. Res. 2022, 2, 100077. [Google Scholar] [CrossRef] [PubMed]
  75. Das, V.; Dandapat, S.; Bora, P.K. Unsupervised Super-Resolution of OCT Images Using Generative Adversarial Network for Improved Age-Related Macular Degeneration Diagnosis. IEEE Sens. J. 2020, 20, 8746–8756. [Google Scholar] [CrossRef]
  76. Yeh, R.A.; Chen, C.; Lim, T.Y.; Schwing, A.G.; Hasegawa-Johnson, M.; Do, M.N. Semantic Image Inpainting with Deep Generative Models. arXiv 2016, arXiv:1607.07539. Available online: https://arxiv.org/abs/1607.07539 (accessed on 4 September 2024).
  77. DeVries, T.; Romero, A.; Pineda, L.; Taylor, G.W.; Drozdzal, M. On the Evaluation of Conditional GANs. arXiv 2019, arXiv:1907.08175. Available online: https://arxiv.org/abs/1907.08175 (accessed on 4 September 2024).
  78. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. Available online: https://arxiv.org/abs/1411.1784 (accessed on 4 September 2024).
  79. Sricharan, K.; Bala, R.; Shreve, M.; Ding, H.; Saketh, K.; Sun, J. Semi-supervised Conditional GANs. arXiv 2017, arXiv:1708.05789. Available online: https://arxiv.org/abs/1708.05789 (accessed on 4 September 2024).
  80. Lan, L.; You, L.; Zhang, Z.; Fan, Z.; Zhao, W.; Zeng, N.; Chen, Y.; Zhou, X. Generative Adversarial Networks and Its Applications in Biomedical Informatics. Front. Public Health 2020, 8, 164. [Google Scholar] [CrossRef]
  81. Agarwal, R.; Tripathi, A. Current Modalities for Low Vision Rehabilitation. Cureus 2021, 13, e16561. Available online: https://www.cureus.com/articles/64479-current-modalities-for-low-vision-rehabilitation (accessed on 4 September 2024). [CrossRef]
  82. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. arXiv 2016, arXiv:1611.07004. Available online: https://arxiv.org/abs/1611.07004 (accessed on 4 September 2024).
  83. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. arXiv 2017, arXiv:1703.10593. Available online: https://arxiv.org/abs/1703.10593 (accessed on 4 September 2024).
  84. Huang, D.; Swanson, E.A.; Lin, C.P.; Schuman, J.S.; Stinson, W.G.; Chang, W.; Hee, M.R.; Flotte, T.; Gregory, K.; Puliafito, C.A. Optical coherence tomography. Science 1991, 254, 1178–1181. [Google Scholar] [CrossRef]
  85. Fujimoto, J.G.; Pitris, C.; Boppart, S.A.; Brezinski, M.E. Optical Coherence Tomography: An Emerging Technology for Biomedical Imaging and Optical Biopsy-PMC. Neoplasia 2000, 2, 9–25. Available online: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1531864/ (accessed on 6 June 2024). [CrossRef]
  86. Gómez-Benlloch, A.; Garrell-Salat, X.; Cobos, E.; López, E.; Esteve-Garcia, A.; Ruiz, S.; Vázquez, M.; Sararols, L.; Biarnés, M. Optical Coherence Tomography in Inherited Macular Dystrophies: A Review. Diagnostics 2024, 14, 878. [Google Scholar] [CrossRef]
  87. Wang, J.; Li, W.; Chen, Y.; Fang, W.; Kong, W.; He, Y.; Shi, G. Weakly supervised anomaly segmentation in retinal OCT images using an adversarial learning approach. Biomed. Opt. Express. 2021, 12, 4713. [Google Scholar]
  88. Ouyang, J.; Mathai, T.S.; Lathrop, K.; Galeotti, J. Accurate tissue interface segmentation via adversarial pre-segmentation of anterior segment OCT images. Biomed. Opt. Express. 2019, 10, 5291. [Google Scholar] [CrossRef] [PubMed]
  89. Menten, M.J.; Holland, R.; Leingang, O.; Bogunović, H.; Hagag, A.M.; Kaye, R.; Riedl, S.; Traber, G.L.; Hassan, O.N.; Pawlowski, N.; et al. Exploring Healthy Retinal Aging with Deep Learning. Ophthalmol. Sci. 2023, 3, 100294. [Google Scholar] [CrossRef] [PubMed]
  90. Sun, L.C.; Pao, S.I.; Huang, K.H.; Wei, C.Y.; Lin, K.F.; Chen, P.N. Generative adversarial network-based deep learning approach in classification of retinal conditions with optical coherence tomography images. Graefes. Arch. Clin. Exp. Ophthalmol. 2023, 261, 1399–1412. [Google Scholar] [CrossRef]
  91. Assaf, J.F.; Abou Mrad, A.; Reinstein, D.Z.; Amescua, G.; Zakka, C.; Archer, T.J.; Yammine, J.; Lamah, E.; Haykal, M.; Awwad, S.T. Creating realistic anterior segment optical coherence tomography images using generative adversarial networks. Br. J. Ophthalmol. 2024, 108, bjo-2023-324633. [Google Scholar] [CrossRef]
  92. Assaf, J.F.; Yazbeck, H.; Reinstein, D.Z.; Archer, T.J.; Arbelaez, J.; Bteich, Y.; Arbelaez, M.C.; Abou Mrad, A.; Awwad, S.T. Enhancing the Automated Detection of Implantable Collamer Lens Vault Using Generative Adversarial Networks and Synthetic Data on Optical Coherence Tomography. J. Refract. Surg. 2024, 40, e199–e207. Available online: https://journals.healio.com/doi/10.3928/1081597X-20240214-01 (accessed on 22 May 2024). [CrossRef]
  93. Zheng, C.; Ye, H.; Yang, J.; Fei, P.; Qiu, Y.; Xie, X.; Wang, Z.; Chen, J.; Zhao, P. Development and Clinical Validation of Semi-Supervised Generative Adversarial Networks for Detection of Retinal Disorders in Optical Coherence Tomography Images Using Small Dataset. Asia-Pac. J. Ophthalmol. 2022, 11, 219–226. [Google Scholar] [CrossRef]
  94. He, X.; Fang, L.; Rabbani, H.; Chen, X.; Liu, Z. Retinal optical coherence tomography image classification with label smoothing generative adversarial network. Neurocomputing 2020, 405, 37–47. [Google Scholar] [CrossRef]
  95. Kugelman, J.; Alonso-Caneiro, D.; Read, S.A.; Vincent, S.J.; Chen, F.K.; Collins, M.J. Data augmentation for patch-based OCT chorio-retinal segmentation using generative adversarial networks. Neural Comput. Appl. 2021, 33, 7393–7408. [Google Scholar] [CrossRef]
  96. Ni, G.; Wu, R.; Zheng, F.; Li, M.; Huang, S.; Ge, X. Toward ground-truth optical coherence tomography via three-dimensional unsupervised deep learning processing and data. IEEE Trans. Med. Imaging 2024, 43, 2395–2407. [Google Scholar] [CrossRef]
  97. Mehdizadeh, M.; Saha, S.; Alonso-Caneiro, D.; Kugelman, J.; MacNish, C.; Chen, F. Employing texture loss to denoise OCT images using generative adversarial networks. Biomed. Opt. Express. 2024, 15, 2262. [Google Scholar] [CrossRef] [PubMed]
  98. Liang, K.; Liu, X.; Chen, S.; Xie, J.; Qing Lee, W.; Liu, L.; Kuan Lee, H. Resolution enhancement and realistic speckle recovery with generative adversarial modeling of micro-optical coherence tomography. Biomed. Opt. Express. 2020, 11, 7236. [Google Scholar] [CrossRef] [PubMed]
  99. Cheong, H.; Devalla, S.K.; Pham, T.H.; Zhang, L.; Tun, T.A.; Wang, X.; Perera, S.; Schmetterer, L.; Aung, T.; Boote, C.; et al. DeshadowGAN: A Deep Learning Approach to Remove Shadows from Optical Coherence Tomography Images. Trans. Vis. Sci. Tech. 2020, 9, 23. [Google Scholar] [CrossRef]
  100. Halupka, K.J.; Antony, B.J.; Lee, M.H.; Lucy, K.A.; Rai, R.S.; Ishikawa, H.; Wollstein, G.; Schuman, J.S.; Garnavi, R. Retinal optical coherence tomography image enhancement via deep learning. Biomed. Opt. Express. 2018, 9, 6205. [Google Scholar] [CrossRef] [PubMed]
  101. Ren, M.; Dey, N.; Fishbaugh, J.; Gerig, G. Segmentation-Renormalized Deep Feature Modulation for Unpaired Image Harmonization. IEEE Trans. Med. Imaging 2021, 40, 1519–1530. [Google Scholar] [CrossRef]
  102. Wu, Y.; Olvera-Barrios, A.; Yanagihara, R.; Kung, T.H.; Lu, R.; Leung, I.; Mishra, A.V.; Nussinovitch, H.; Grimaldi, G.; Blazes, M.; et al. Training Deep Learning Models to Work on Multiple Devices by Cross-Domain Learning with No Additional Annotations. Ophthalmology 2023, 130, 213–222. [Google Scholar] [CrossRef]
  103. Chen, S.; Ma, D.; Lee, S.; Yu TT, L.; Xu, G.; Lu, D.; Popuri, K.; Ju, M.J.; Sarunic, M.V.; Beg, M.F. Segmentation-guided domain adaptation and data harmonization of multi-device retinal optical coherence tomography using cycle-consistent generative adversarial networks. Comput. Biol. Med. 2023, 159, 106595. [Google Scholar] [CrossRef]
  104. Lazaridis, G.; Lorenzi, M.; Ourselin, S.; Garway-Heath, D. Improving statistical power of glaucoma clinical trials using an ensemble of cyclical generative adversarial networks. Med. Image Anal. 2021, 68, 101906. [Google Scholar] [CrossRef]
  105. Romo-Bucheli, D.; Seeböck, P.; Orlando, J.I.; Gerendas, B.S.; Waldstein, S.M.; Schmidt-Erfurth, U.; Bogunović, H. Reducing image variability across OCT devices with unsupervised unpaired learning for improved segmentation of retina. Biomed. Opt. Express. 2020, 1s1, 346. [Google Scholar] [CrossRef]
  106. Tripathi, A.; Kumar, P.; Mayya, V.; Tulsani, A. Generating OCT B-Scan DME images using optimized Generative Adversarial Networks (GANs). Heliyon 2023, 9, e18773. [Google Scholar] [CrossRef]
  107. Liu, S.; Hu, W.; Xu, F.; Chen, W.; Liu, J.; Yu, X.; Wang, Z.; Li, Z.; Li, Z.; Yang, X.; et al. Prediction of OCT images of short-term response to anti-VEGF treatment for diabetic macular edema using different generative adversarial networks. Photodiagnosis Photodyn. Ther. 2023, 41, 103272. [Google Scholar] [CrossRef] [PubMed]
  108. Lee, H.; Kim, S.; Kim, M.A.; Chung, H.; Kim, H.C. Post-treatment prediction of optical coherence tomography using a conditional generative adversarial network in age-related macular degeneration. Retina 2021, 41, 572–580. [Google Scholar] [CrossRef] [PubMed]
  109. Xu, F.; Yu, X.; Gao, Y.; Ning, X.; Huang, Z.; Wei, M.; Zhai, W.; Zhang, R.; Wang, S.; Li, J. Predicting OCT images of short-term response to anti-VEGF treatment for retinal vein occlusion using generative adversarial network. Front. Bioeng. Biotechnol. 2022, 10, 914964. [Google Scholar] [CrossRef]
  110. Ehrlich, R.; Harris, A.; Wentz, S.M.; Moore, N.A.; Siesky, B.A. Anatomy and Regulation of the Optic Nerve Blood Flow. In Reference Module in Neuroscience and Biobehavioral Psychology; Elsevier: Amsterdam, The Netherlands, 2017; p. 9780128093245013018. Available online: https://linkinghub.elsevier.com/retrieve/pii/B9780128093245013018 (accessed on 26 May 2024).
  111. Javed, A.; Khanna, A.; Palmer, E.; Wilde, C.; Zaman, A.; Orr, G.; Kumudhan, D.; Lakshmanan, A.; Panos, G.D. Optical coherence tomography angiography: A review of the current literature. J. Int. Med. Res. 2023, 51, 03000605231187933. [Google Scholar] [CrossRef] [PubMed]
  112. de Carlo, T.E.; Romano, A.; Waheed, N.K.; Duker, J.S. A review of optical coherence tomography angiography (OCTA). Int. J. Retin. Vitr. 2015, 1, 5. [Google Scholar] [CrossRef]
  113. Badhon, R.H.; Thompson, A.C.; Lim, J.I.; Leng, T.; Alam, M.N. Quantitative Characterization of Retinal Features in Translated OCTA. medRxiv 2024. Available online: http://medrxiv.org/lookup/doi/10.1101/2024.02.23.24303275 (accessed on 25 May 2024). [CrossRef]
  114. Coronado, I.; Pachade, S.; Trucco, E.; Abdelkhaleq, R.; Yan, J.; Salazar-Marioni, S.; Jagolino-Cole, A.; Bahrainian, M.; Channa, R.; Sheth, S.A.; et al. Synthetic OCT-A blood vessel maps using fundus images and generative adversarial networks. Sci. Rep. 2023, 13, 15325. [Google Scholar]
  115. Cao, J.; Xu, Z.; Xu, M.; Ma, Y.; Zhao, Y. A two-stage framework for optical coherence tomography angiography image quality improvement. Front. Med. 2023, 10, 1061357. [Google Scholar] [CrossRef]
  116. Jiang, Z.; Huang, Z.; Qiu, B.; Meng, X.; You, Y.; Liu, X.; Liu, G.; Zhou, C.; Yang, K.; Maier, A.; et al. Comparative study of deep learning models for optical coherence tomography angiography. Biomed. Opt. Express. 2020, 11, 1580. [Google Scholar] [CrossRef]
  117. Kornblau, I.S.; El-Annan, J.F. Adverse reactions to fluorescein angiography: A comprehensive review of the literature. Surv. Ophthalmol. 2019, 64, 679–693. [Google Scholar] [CrossRef]
  118. Maquire, A.M.; Bennett, J. Fluorescein elimination in human breast milk. Arch. Ophthalmol. 1988, 106, 718–719. [Google Scholar] [CrossRef]
  119. Huang, K.; Li, M.; Yu, J.; Miao, J.; Hu, Z.; Yuan, S.; Chen, Q. Lesion-aware generative adversarial networks for color fundus image to fundus fluorescein angiography translation. Comput. Methods Programs Biomed. 2023, 229, 107306. [Google Scholar] [CrossRef] [PubMed]
  120. Li, P.; He, Y.; Wang, P.; Wang, J.; Shi, G.; Chen, Y. Synthesizing multi-frame high-resolution fluorescein angiography images from retinal fundus images using generative adversarial networks. BioMed. Eng. OnLine 2023, 22, 16. [Google Scholar] [CrossRef] [PubMed]
  121. Li, W.; He, Y.; Kong, W.; Wang, J.; Deng, G.; Chen, Y. SequenceGAN: Generating Fundus Fluorescence Angiography Sequences from Structure Fundus Image. In Simulation and Synthesis in Medical Imaging; Svoboda, D., Burgos, N., Wolterink, J.M., Zhao, C., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; Volume 12965, pp. 110–120. Available online: https://link.springer.com/10.1007/978-3-030-87592-3_11 (accessed on 25 May 2024).
  122. Kamran, S.A.; Fariha Hossain, K.; Tavakkoli, A.; Zuckerbrod, S.; Baker, S.A.; Sanders, K.M. Fundus2Angio: A Conditional GAN Architecture for Generating Fluorescein Angiography Images from Retinal Fundus Photography. In Advances in Visual Computing; Bebis, G., Yin, Z., Kim, E., Bender, J., Subr, K., Kwon, B.C., Zhao, J., Kalkofen, D., Baciu, G., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2020; Volume 12510, pp. 125–138. Available online: https://link.springer.com/10.1007/978-3-030-64559-5_10 (accessed on 25 May 2024).
  123. Kamran, S.A.; Hossain, K.F.; Tavakkoli, A.; Zuckerbrod, S.L. Attention2AngioGAN: Synthesizing Fluorescein Angiography from Retinal Fundus Images using Generative Adversarial Networks. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 9122–9129. Available online: https://ieeexplore.ieee.org/document/9412428/ (accessed on 25 May 2024).
  124. Kamran, S.A.; Hossain, K.F.; Tavakkoli, A.; Zuckerbrod, S.L.; Baker, S.A. VTGAN: Semi-supervised Retinal Image Synthesis and Disease Prediction using Vision Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; Available online: https://arxiv.org/abs/2104.06757 (accessed on 25 May 2024).
  125. Kamran, S.A.; Hossain, K.F.; Ong, J.; Waisberg, E.; Zaman, N.; Baker, S.A.; Lee, A.G.; Tavakkoli, A. FA4SANS-GAN: A Novel Machine Learning Generative Adversarial Network to Further Understand Ophthalmic Changes in Spaceflight Associated Neuro-Ocular Syndrome (SANS). Ophthalmol. Sci. 2024, 4, 100493. [Google Scholar] [CrossRef]
  126. Shi, D.; Zhang, W.; He, S.; Chen, Y.; Song, F.; Liu, S.; Wang, R.; Zheng, Y.; He, M. Translation of Color Fundus Photography into Fluorescein Angiography Using Deep Learning for Enhanced Diabetic Retinopathy Screening. Ophthalmol. Sci. 2023, 3, 100401. [Google Scholar] [CrossRef]
  127. Shi, D.; He, S.; Yang, J.; Zheng, Y.; He, M. One-shot Retinal Artery and Vein Segmentation via Cross-modality Pretraining. Ophthalmol. Sci. 2024, 4, 100363. [Google Scholar] [CrossRef]
  128. Ge, R.; Fang, Z.; Wei, P.; Chen, Z.; Jiang, H.; Elazab, A.; Li, W.; Wan, X.; Zhang, S.; Wang, C. UWAFA-GAN: Ultra-Wide-Angle Fluorescein Angiography Transformation via Multi-scale Generation and Registration Enhancement. IEEE J. Biomed. Health Inform. 2024, 28, 4820–4829. [Google Scholar] [CrossRef]
  129. Wang, X.; Ji, Z.; Ma, X.; Zhang, Z.; Yi, Z.; Zheng, H.; Fan, W.; Chen, C. Automated Grading of Diabetic Retinopathy with Ultra-Widefield Fluorescein Angiography and Deep Learning. J. Diabetes Res. 2021, 2021, 2611250. [Google Scholar] [CrossRef]
  130. Abdelmotaal, H.; Sharaf, M.; Soliman, W.; Wasfi, E.; Kedwany, S.M. Bridging the resources gap: Deep learning for fluorescein angiography and optical coherence tomography macular thickness map image translation. BMC Ophthalmol. 2022, 22, 355. [Google Scholar] [CrossRef]
  131. Pole, C.; Ameri, H. Fundus Autofluorescence and Clinical Applications. J. Ophthalmic. Vis. Res. 2021, 16, 432–461. [Google Scholar] [CrossRef]
  132. Wu, M.; Cai, X.; Chen, Q.; Ji, Z.; Niu, S.; Leng, T.; Rubin, D.L.; Park, H. Geographic atrophy segmentation in SD-OCT images using synthesized fundus autofluorescence imaging. Comput. Methods Programs Biomed. 2019, 182, 105101. [Google Scholar] [CrossRef]
  133. Su, J.; She, K.; Song, L.; Jin, X.; Li, R.; Zhao, Q.; Xiao, J.; Chen, D.; Cheng, H.; Lu, F.; et al. In vivo base editing rescues photoreceptors in a mouse model of retinitis pigmentosa. Mol. Ther.-Nucleic Acids 2023, 31, 596–609. [Google Scholar] [CrossRef] [PubMed]
  134. Veturi, Y.A.; Woof, W.; Lazebnik, T.; Moghul, I.; Woodward-Court, P.; Wagner, S.K.; Cabral de Guimarães, T.A.; Daich Varela, M.; Liefers, B.; Patel, P.J.; et al. SynthEye: Investigating the Impact of Synthetic Data on Artificial Intelligence-assisted Gene Diagnosis of Inherited Retinal Disease. Ophthalmol. Sci. 2023, 3, 100258. [Google Scholar] [CrossRef] [PubMed]
  135. Asanad, S.; Karanjia, R. Full-Field Electroretinogram. In StatPearls [Internet]; StatPearls Publishing: Treasure Island, FL, USA, 2024. Available online: http://www.ncbi.nlm.nih.gov/books/NBK557483/ (accessed on 9 June 2024).
  136. Asanad, S.; Karanjia, R. Multifocal Electroretinogram. In StatPearls [Internet]; StatPearls Publishing: Treasure Island, FL, USA, 2024. Available online: https://www.ncbi.nlm.nih.gov/books/NBK564322/ (accessed on 9 June 2024).
  137. Kulyabin, M.; Zhdanov, A.; Maier, A.; Loh, L.; Estevez, J.J.; Constable, P.A. Generating synthetic electroretinogram waveforms using Artificial Intelligence to improve classification of retinal conditions in under-represented populations. arXiv 2024, arXiv:2404.11842. Available online: https://arxiv.org/abs/2404.11842 (accessed on 27 May 2024). [CrossRef]
  138. Muraleedharan, S.; Tripathy, K. Indocyanine Green (ICG) Angiography. In StatPearls [Internet]; StatPearls Publishing: Treasure Island, FL, USA, 2024. Available online: http://www.ncbi.nlm.nih.gov/books/NBK580479/ (accessed on 9 June 2024).
  139. Chen, R.; Zhang, W.; Song, F.; Yu, H.; Cao, D.; Zheng, Y.; He, M.; Shi, D. Translating color fundus photography to indocyanine green angiography using deep-learning for age-related macular degeneration screening. npj Digit. Med. 2024, 7, 34. [Google Scholar] [CrossRef]
  140. Jiang, H.; Chen, X.; Shi, F.; Ma, Y.; Xiang, D.; Ye, L.; Su, J.; Li, Z.; Chen, Q.; Hua, Y.; et al. Improved cGAN based linear lesion segmentation in high myopia ICGA images. Biomed. Opt. Express. 2019, 10, 2355. [Google Scholar] [CrossRef]
  141. Coscas, G.; Soubrane, G. Pathologic Myopia. In Retinal Imaging; Elsevier: Amsterdam, The Netherlands, 2006; pp. 164–174. Available online: https://linkinghub.elsevier.com/retrieve/pii/B9780323023467500181 (accessed on 26 May 2024).
  142. Rao, H.L.; Yadav, R.K.; Begum, V.U.; Addepalli, U.K.; Choudhari, N.S.; Senthil, S.; Garudadri, C.S. Role of Visual Field Reliability Indices in Ruling Out Glaucoma. JAMA Ophthalmol. 2015, 133, 40. [Google Scholar] [CrossRef]
  143. Kang, H.; Ko, S.; Kim, J.-C.; Le, D.T.; Bum, J.; Han, J.C.; Choo, H. Visual Field Prediction for Fundus Image with Generative AI. In Proceedings of the 2024 18th International Conference on Ubiquitous Information Management and Communication (IMCOM), Kuala Lumpur, Malaysia, 3–5 January 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–3. Available online: https://ieeexplore.ieee.org/document/10418344/ (accessed on 27 May 2024).
  144. Ranga, V.; Dave, M.; Verma, A.K. Modified Max-Min Algorithm for Game Theory. In Proceedings of the 2015 Fifth International Conference on Advanced Computing & Communication Technologies, Haryana, India, 20–21 February 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 153–156. Available online: http://ieeexplore.ieee.org/document/7079070/ (accessed on 4 September 2024).
  145. Roth, K.; Lucchi, A.; Nowozin, S.; Hofmann, T. Stabilizing Training of Generative Adversarial Networks through Regularization. arXiv 2017, arXiv:1705.09367. Available online: https://arxiv.org/abs/1705.09367 (accessed on 4 September 2024).
  146. Chong, P.; Ruff, L.; Kloft, M.; Binder, A. Simple and Effective Prevention of Mode Collapse in Deep One-Class Classification. arXiv 2020, arXiv:2001.08873. Available online: https://arxiv.org/abs/2001.08873 (accessed on 4 September 2024).
  147. Thanh-Tung, H.; Tran, T. On Catastrophic Forgetting and Mode Collapse in Generative Adversarial Networks. arXiv 2018, arXiv:1807.04015. Available online: https://arxiv.org/abs/1807.04015 (accessed on 4 September 2024).
  148. Metz, L.; Poole, B.; Pfau, D.; Sohl-Dickstein, J. Unrolled Generative Adversarial Networks. arXiv 2016, arXiv:1611.02163. Available online: https://arxiv.org/abs/1611.02163 (accessed on 4 September 2024).
  149. Guo, X.; Yin, Y.; Dong, C.; Yang, G.; Zhou, G. On the Class Imbalance Problem. In Proceedings of the 2008 Fourth International Conference on Natural Computation, Jinan, China, 18–20 October 2018; IEEE: Piscataway, NJ, USA, 2008; pp. 192–201. Available online: http://ieeexplore.ieee.org/document/4667275/ (accessed on 4 September 2024).
  150. Dixon, L.; Li, J.; Sorensen, J.; Thain, N.; Vasserman, L. Measuring and Mitigating Unintended Bias in Text Classification. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, New Orleans, LA, USA, 27 December 2018; ACM: New York, NY, USA, 2018; pp. 67–73. Available online: https://dl.acm.org/doi/10.1145/3278721.3278729 (accessed on 4 September 2024).
  151. Hutchinson, B.; Prabhakaran, V.; Denton, E.; Webster, K.; Zhong, Y.; Denuyl, S. Unintended machine learning biases as social barriers for persons with disabilitiess. SIGACCESS Access. Comput. 2020, 125, 1558–2337. [Google Scholar] [CrossRef]
  152. Suresh, H.; Guttag, J.V. A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle. arXiv 2019, arXiv:1901.10002. Available online: https://arxiv.org/abs/1901.10002 (accessed on 4 September 2024).
  153. Dumont, V.; Ju, X.; Mueller, J. Hyperparameter Optimization of Generative Adversarial Network Models for High-Energy Physics Simulations. arXiv 2022, arXiv:2208.07715. Available online: https://arxiv.org/abs/2208.07715 (accessed on 4 September 2024).
  154. Krähenbühl, P.; Doersch, C.; Donahue, J.; Darrell, T. Data-dependent Initializations of Convolutional Neural Networks. arXiv 2015, arXiv:1511.06856. Available online: https://arxiv.org/abs/1511.06856 (accessed on 4 September 2024).
  155. Olden, J.D.; Joy, M.K.; Death, R.G. An accurate comparison of methods for quantifying variable importance in artificial neural networks using simulated data. Ecol. Model. 2004, 178, 389–397. [Google Scholar] [CrossRef]
Figure 1. Applications of Generative Adversarial Networks in Ophthalmology. GANs can enhance image quality by reducing noise and artifacts (a), perform image segmentation for structure identification, such as cup (red) and disc (blue) assessment when assessing an optic nerve (nerve) (b), and generate imaging modalities from other inputs (c), supporting clinical diagnosis, disease management, and physician training.
Figure 1. Applications of Generative Adversarial Networks in Ophthalmology. GANs can enhance image quality by reducing noise and artifacts (a), perform image segmentation for structure identification, such as cup (red) and disc (blue) assessment when assessing an optic nerve (nerve) (b), and generate imaging modalities from other inputs (c), supporting clinical diagnosis, disease management, and physician training.
Biomedicines 13 00284 g001
Figure 2. Schematic representation of a generative adversarial network architecture applied to ophthalmology. GANs consist of two primary components: a generator and a discriminator. The generator produces synthetic images starting from random noise and aims to generate outputs that resemble real images. Real images are sourced from the training dataset, while the discriminator is tasked with distinguishing between real images and synthetic images created by the generator. The discriminator evaluates each input image and calculates the likelihood of it being real or fake. Feedback, or backpropagation (indicated by red arrows), flows from the discriminator to the generator, allowing the generator to improve its image synthesis over successive iterations. The goal is for the generator to produce images that are indistinguishable from real ones. The fake images shown here were generated using a Stable Diffusion model (accessible at https://app.aitubo.ai/create/ accessed on 10 October 2024). The schematic captures the interplay between the generator and discriminator in a GAN setup, particularly in the context of ophthalmic imaging. The Figure was created based on the cited reference [36].
Figure 2. Schematic representation of a generative adversarial network architecture applied to ophthalmology. GANs consist of two primary components: a generator and a discriminator. The generator produces synthetic images starting from random noise and aims to generate outputs that resemble real images. Real images are sourced from the training dataset, while the discriminator is tasked with distinguishing between real images and synthetic images created by the generator. The discriminator evaluates each input image and calculates the likelihood of it being real or fake. Feedback, or backpropagation (indicated by red arrows), flows from the discriminator to the generator, allowing the generator to improve its image synthesis over successive iterations. The goal is for the generator to produce images that are indistinguishable from real ones. The fake images shown here were generated using a Stable Diffusion model (accessible at https://app.aitubo.ai/create/ accessed on 10 October 2024). The schematic captures the interplay between the generator and discriminator in a GAN setup, particularly in the context of ophthalmic imaging. The Figure was created based on the cited reference [36].
Biomedicines 13 00284 g002
Figure 3. Generative adversarial networks improve imaging quality in the field of ophthalmology. Generative adversarial networks (GANs) can be applied to optical coherence tomography images and fundus photos by decreasing speckle noise induced by media opacities during acquisition, as well as removing artifacts induced during acquisition.
Figure 3. Generative adversarial networks improve imaging quality in the field of ophthalmology. Generative adversarial networks (GANs) can be applied to optical coherence tomography images and fundus photos by decreasing speckle noise induced by media opacities during acquisition, as well as removing artifacts induced during acquisition.
Biomedicines 13 00284 g003
Figure 4. Inpainting application illustrations applied to ophthalmology. Inpainting can be used for various applications in ophthalmology, such as filling in photos with black areas (a) and generating intermediate synthetic slices between optical coherence tomography acquisition images (b).
Figure 4. Inpainting application illustrations applied to ophthalmology. Inpainting can be used for various applications in ophthalmology, such as filling in photos with black areas (a) and generating intermediate synthetic slices between optical coherence tomography acquisition images (b).
Biomedicines 13 00284 g004
Table 1. Basic terminology of generative adversarial networks and other artificial intelligence-based algorithms.
Table 1. Basic terminology of generative adversarial networks and other artificial intelligence-based algorithms.
TermExplanationReferences
AI a modelA program that has been trained to perform a task without further human intervention. For example, this task can be as simple as telling if the number provided (called input) is greater or less than a certain integer (i.e., 5).[44]
Neural network/artificial neural networks Type of AI model that uses several layers of nodes (artificial neurons) to perform a task. The first node layer is the input layer, followed by one or several layers of “hidden nodes”, and the last layer is the output. Nodes in one layer are related to nodes in previous layers by weights and biases that summarize the input signal and an activation function that determines whether that signal should be transmitted.[6,7]
Convolutional Neural Network (CNN) aA CNN a is a type of neural network adapted for image-type inputs and data that is in the form of a grid. Their major particularity is that they apply a matrix (i.e., a 3 × 3 grid matrix) and convolutional operations to input data.[4,5]
Deep convolutional neural networksA type of neural network similar to CNNs, but with a larger architecture allowing the model to perform more complex tasks. This gain in terms of accuracy and complexity costs more computer power.[3]
Discriminatory modelsAI models that perform classifications and output labels. These models help with discriminatory tasks, such as, but not limited to, distinguishing healthy versus pathological conditions, benign versus malignant lesions, disease A versus disease B, etc. [45]
Generative models AI models that learn patterns from a dataset and output new but similar data. In imaging generative models, the output is often a new image rather than a predefined label.[46,47]
GeneratorThe first part of a GAN that uses an internal distribution of imaging data is to create a candidate image close to a concrete real-world counterpart. The objective of training is to maximize the generator’s performance.[47]
DiscriminatorThe second part of a GAN tries to distinguish between real and fake data provided by the generator. The objective of the training is to have a discriminator that is unable to distinguish between real versus fake data provided by the generator while maximizing the discriminator’s performance. [47]
BackpropagationA mathematical update of weights from the last to the first layer of the AI model based on the loss function. The partial derivatives of the loss function to each weight are used in the weight update. This mathematic type of optimization is called gradient descent. The objective of backpropagation and gradient descent is to maximize the models’ accuracy by minimizing the loss function.[48]
Loss functionA function that evaluates the algorithm’s performance when comparing its output to the answers (ground truth). Incorrect answers provide a high loss value. The objective of training is to minimize the loss function.[49]
EpochCorresponds to one algorithm training session with a complete pass of the training dataset.[50]
OverfittingUndesirable behavior of an algorithm that fails to generalize from input data by being too specific and fitting too closely. The algorithm is thus unable to correctly perform accurate prediction.[51]
ConvergenceCorresponds to the point where the loss function is successfully minimized, and training parameters reach a stable state. The algorithm is thus able to accurately perform predictions with the obtained parameters.[52]
a Abbreviations: AI, artificial intelligence; CNN, convolutional neural networks; GAN, generative adversarial network.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Remtulla, R.; Samet, A.; Kulbay, M.; Akdag, A.; Hocini, A.; Volniansky, A.; Kahn Ali, S.; Qian, C.X. A Future Picture: A Review of Current Generative Adversarial Neural Networks in Vitreoretinal Pathologies and Their Future Potentials. Biomedicines 2025, 13, 284. https://doi.org/10.3390/biomedicines13020284

AMA Style

Remtulla R, Samet A, Kulbay M, Akdag A, Hocini A, Volniansky A, Kahn Ali S, Qian CX. A Future Picture: A Review of Current Generative Adversarial Neural Networks in Vitreoretinal Pathologies and Their Future Potentials. Biomedicines. 2025; 13(2):284. https://doi.org/10.3390/biomedicines13020284

Chicago/Turabian Style

Remtulla, Raheem, Adam Samet, Merve Kulbay, Arjin Akdag, Adam Hocini, Anton Volniansky, Shigufa Kahn Ali, and Cynthia X. Qian. 2025. "A Future Picture: A Review of Current Generative Adversarial Neural Networks in Vitreoretinal Pathologies and Their Future Potentials" Biomedicines 13, no. 2: 284. https://doi.org/10.3390/biomedicines13020284

APA Style

Remtulla, R., Samet, A., Kulbay, M., Akdag, A., Hocini, A., Volniansky, A., Kahn Ali, S., & Qian, C. X. (2025). A Future Picture: A Review of Current Generative Adversarial Neural Networks in Vitreoretinal Pathologies and Their Future Potentials. Biomedicines, 13(2), 284. https://doi.org/10.3390/biomedicines13020284

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop