Next Article in Journal
Deep Learning in Medical Image Analysis
Previous Article in Journal
Psychophysical Determination of the Relevant Colours That Describe the Colour Palette of Paintings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Iterative-Trained Semi-Blind Deconvolution Algorithm to Compensate Straylight in Retinal Images

Departamento de Física Aplicada, Universidad de Zaragoza, 50009 Zaragoza, Spain
*
Author to whom correspondence should be addressed.
J. Imaging 2021, 7(4), 73; https://doi.org/10.3390/jimaging7040073
Submission received: 26 February 2021 / Revised: 13 April 2021 / Accepted: 14 April 2021 / Published: 16 April 2021
(This article belongs to the Special Issue Blind Image Restoration)

Abstract

:
The optical quality of an image depends on both the optical properties of the imaging system and the physical properties of the medium in which the light travels from the object to the final imaging sensor. The analysis of the point spread function of the optical system is an objective way to quantify the image degradation. In retinal imaging, the presence of corneal or cristalline lens opacifications spread the light at wide angular distributions. If the mathematical operator that degrades the image is known, the image can be restored through deconvolution methods. In the particular case of retinal imaging, this operator may be unknown (or partially) due to the presence of cataracts, corneal edema, or vitreous opacification. In those cases, blind deconvolution theory provides useful results to restore important spatial information of the image. In this work, a new semi-blind deconvolution method has been developed by training an iterative process with the Glare Spread Function kernel based on the Richardson-Lucy deconvolution algorithm to compensate a veiling glare effect in retinal images due to intraocular straylight. The method was first tested with simulated retinal images generated from a straylight eye model and applied to a real retinal image dataset composed of healthy subjects and patients with glaucoma and diabetic retinopathy. Results showed the capacity of the algorithm to detect and compensate the veiling glare degradation and improving the image sharpness up to 1000% in the case of healthy subjects and up to 700% in the pathological retinal images. This image quality improvement allows performing image segmentation processing with restored hidden spatial information after deconvolution.

1. Introduction

The eye, as a non-ideal optical system, presents intrinsic optical physiological imperfections throughout the ocular media limiting the quality of the image formed at the retina due to the combined presence of aberrations and scattering [1]. Multiple scattering contributions within the ocular media give rise to intraocular straylight (IS) and veiling glare (VG) generation that affects the contrast sensitivity [2], perception of colors, and general visual function [3]. IS results from different sources, such as cornea, lens, aqueous and vitreous humors, fundus reflectance, retinal pigmentation, and age are also multiplicative factors [4]. Some studies have reported an increase in IS after implantation of intraocular lenses [5,6,7] or poorly fitting contact lenses [8].
In addition to the dramatic impact of IS in visual disability, ocular straylight sources convert the eye in a scattering medium that degrades the retinal imaging modalities in opthalmoscopy, such as Optical Computerized Tomography (OCT), confocal laser scanning opthalmoscopy, or digital fundus images [9,10], among others. Retinal imaging analysis plays a fundamental role in the diagnosis of ophthalmic diseases. In that context, image segmentation processing have been applied in retinal fundus images for the early detection of glaucoma [11], diabetic retinopathy [12], or age-related macular degeneration [13]. If the angular distribution of the IS is wide enough, the sharpness, contrast, and image resolution will be decreased, making the features extraction harder even with advanced post-processing techniques.
In this sense, different optical methods have been proposed to reduce IS effects such as Fourier analysis [14] or adaptive optics (AO) deep learning techniques [15]. However, in the presence of advanced cataracts, corneal distrophies, edema, or vitreous opacifications can lead to an unavoidable VG generation degrading the fundus images and, therefore, making segmentation and image analysis difficult and leading to erroneous estimates.
Although the mathematical degradation function may be unknown, blind deconvolution algorithms have been successfully applied to restore degraded retinal images [16], to improve high-resolution scanning laser ophthalmoscopy [17], or to reveal hidden structures in pathological retinas [18].
Iterative blind deconvolution algorithms have been previously reported [19], obtaining the object and point spread function (PSF) estimates after each iteration. Years later, the concept of semi-blind deconvolution was introduced in the implementation of a Richardson-Lucy algorithm, improving the restoration previously reported [20] to incorporate functional forms assuming previous knowledge of the PSF. The purpose of this work has been to develop a new semi-blind deconvolution algorithm based on the Richardson-Lucy deconvolution theory incorporating the Glare Spread Function (GSF) to restore retinal images from intraocular straylight. The algorithm also incorporates new computational costs’ optimization process based on an iterative-trained non-reference image quality score method. In our work, the GSF is considered as the mathematical IS operator that blur retinal images as a consequence of VG generation.
The algorithm was tested by optical simulations performed in a scattering eye model and validated with a public retinal image’s dataset from healthy, glaucomatous and diabetic retinopathy patients. Results showed the capability of the algorithm to detect the VG effects caused by IS within the images with strong image quality and spatial resolution restoration quantified by Image Quality Assessment (IQA). Finally, an image segmentation processing was carried out, revealing hidden blood vessels and retinal structures as well as an improved structural differentiation of the optic disk, drusen, and other retinal fundus findings.

2. Materials and Methods

2.1. Image Dataset

We used a public database (data set can be found in Reference [21]) of fundus images composed of healthy subjects (15), patients with glaucoma (15), and diabetic retinopathy (15) recorded by a group retinal image analysis of clinicians with a Canon CR-1 fundus camera (Field of view 45º). Figure 1 shows representative retinal images from every group.

2.2. Image Quality Assessment and Segmentation

IS arises from multiple scatterings of light while propagating through the ocular media (equivalently optical system) before reaching the retina (equivalently optical sensor) generating a global illumination effect that fogs the observed scene (or the field of view of the sensor camera). Straylight generates VG that reduces local contrast and sharpness, and blur the image. In our work, the image quality assessment (IQA) has been carried out by computing five parameters: Image Sharpness (S), Trained Natural Image Quality Evaluator Score (NIQE), Structural Similarity Index measure (SSIM), and Blind Reference Image Spatial Quality Evaluator (BRISQUE). IQA and segmentation processing was performed in a custom MATLAB 2019b script (The Mathworks Inc., Natick, MA, USA).

2.2.1. Image Sharpness

In Computer Vision, acutance [22] is related to the amplitude of the gradient of an image that measures the subjective sharpness. In this work, we define the Image Sharpness (S) parameter as follows:
S x , y = G x 2 + G y 2
where Gx and Gy are the horizontal (x) and vertical (y) directional gradients of the image.

2.2.2. Trained NIQE Score

Natural Image Quality Evaluator (NIQE) computes blind image quality assessment comparing the restored image to a predefined feature model [23] in order to evaluate the restoration in every iteration, allowing establishing a stopping criterion of the deconvolution when the maximum NIQE score is reached. In this work, we have developed our own trained NIQE model by extracting information from a reference image dataset. This process starts by partitioning every image of the dataset selected as a reference in small blocks that can allow significant spatial segmentation. In our case, blocks of 40 × 40 pixels of a total image size of 400 × 400 pixels were selected for extracting the covariance, mean luminance, and image sharpness properties. These parameters provide the reference of the model, where the outputs of the fitting are the multivariate mean and standard deviation of the Gaussian. The model stores the statistical outputs of those blocks that are used as a reference. Once the trained NIQE model is performed, the same statistical feature extraction is applied to the testing images from the dataset to create the testing Gaussian model. Finally, the NIQE score is calculated by computing the distance between trained (reference) and testing Gaussians as schematized in Figure 2. In this work, we have used the image dataset described in Section 2.1. This dataset is composed of three groups of subjects: healthy, glaucoma, and diabetic retinopathy patients. Focusing on the goal of restoring IS in retinal images, in this work, we have established the healthy group as a reference due to the absence of straylight sources that originated by ocular pathologies. In this sense, as referenced in the Introduction, the pathological groups (glaucoma and retinopathy) are susceptible of high generation VG effects due to pathological IS sources, which were then set as testing images.

2.2.3. Structural Similarity Index

SSIM is a method for measuring the similarity between pairs of images A and B, based on an optical quality comparison by considering one of them as a reference or ideally free of distortion. The calculation of SSIM relates three components of the image: luminance (l), contrast (c), and structure (s) trying to take into account their perceptual relations [24].
S S I M A , B   = f ( l A , B ,   c A , B ,   s A , B )  

2.2.4. Blind Reference Image Spatial Quality Evaluator

Blind Reference Image Spatial Quality Evaluator (BRISQUE) [24] is a referenceless image quality evaluator that is used in our work to evaluate the performance of the restoration after deconvolution of real retinal images. Smaller values of BRISQUE corresponds to a better perceptual quality. BRISQUE can be computed from both a pair of reference and test images or in a single image. In this work, we have performed single calculations to allow a quantitative comparison between the original and restored images. BRISQUE was computed by employing MATLAB 2019b “Computer Vision Toolbox”.

2.2.5. Segmentation

Retinal image segmentation was carried out using a classical image processing procedure summarized in Figure 3. The RGB input image is converted to a gray scale and binarized, once the intensity image is created. The “Canny” filter [25] was applied using the same thresholding sensitivity for egde detection.

2.3. Glare Spread Function Generation

Considering the eye as an optical system, the VG concept is equivalent to the IS [26,27]. That is, the field of view of an optical system undergoes a constant superimposed luminance equivalent to the visual effect experienced while looking at an object scene in the presence of a glare source at the same angular field [24,25]. In this sense, the GSF according to the “Commission Internationale de l’eclairage” (CIE) standard is defined as [28]:
G S F α = 1 0.08 A g e 70 4 9.2 10 6 1 + α 0.046 2 1.5 + 1.5 10 5 1 + α 0.045 2 1.5 + 1 + 1.6 A g e 70 4 400 1 + α 0.1 2 + 3 10 8 α 2 + P F 1300 1 + α 0.1 2 1.5 + 0.8 1 + α 0.1 2 0.5 + 2.5 10 3 P F
where Age and PF are the age and the pigment factor of the observer, and α is the glare angle. This equation basically relates the VG with respect to the relative illuminance. Thus, it is outstanding to notice that parameters Age and PF are considered constants in this work when computing the illuminance of a scene instead of that of an observer’s retina. In that case, GSF will only depend on the glare angle. From the filter of Equation (3), the bi-dimensional filter kernel GSF is then generated by storing the values of the GSF as a function of the glare angle α, for each pixel of the original image (of which the size is m x n pixels), that is, the radius of the kernel is double the retinal image size. The values of the kernel, computed from Equation (3), are always positive and are normalized by the total sum, ensuring that the area under the 1D-GSF curve sum the unit.

2.4. Iterative-Trained Semi-Blind Deconvolution Algorithm Description

Assume that acquired fundus images i′ are modelled by the following mathematical convolution expression:
i = G S F i                        
where i is the original retinal area image, GSF represents the wide-angle point spread function degrading the retinal image (see description in Section 2.3), and denotes the convolution operation.
Richardson-Lucy deconvolution algorithm was derived from Bayesian theory and described in an iterative method for image restoration [29,30] that provides the following solution to estimate i:
î N + 1 = î N · i î N G S F G S F T        
where î is the estimates of i, N is the iteration number, and G S F T is the spatially reversed GSF built by means of horizontal and vertical flip operations. In this paper, we ensure non-negativity of i by avoiding additional constraints. If the iterative process converges, it converges to the maximum likelihood estimation [31].
In iterative deconvolution algorithms, the stopping criterion is essential to optimize the deconvolution and processing time. In this sense, the combination of regularization functions to Richardson-Lucy algorithms has proved to have enhanced performance [32]. On the other hand, we are assuming that the GSF is the “known” PSF that degrades the retinal images. However, no previous estimation of the ocular straylight were performed. This fact defines the semi-blind approach of our algorithm that employs the iteration as an optimization process in a similar way to blind deconvolution theory.
The Iterative-Trained Semi-Blind Deconvolution (ITSD) algorithm procedure carries two main steps:
Step #1: GSF optimization by means of direct search of the maximum image sharpness.
As mentioned in the algorithm description, the amount IS during the image acquisition is unknown and this lack of information is used in our method as an optimization process, as follows.
The program reads the grayscale image and generates an initial GSF for a null glare angle (α), performs a non-iterative deconvolution operation, and then calculates the sharpness (S) of the restored image. The process keeps working by increasing α in steps of 1° in each iteration of the deconvolution until the maximum S is found by employing the Hill-Climbing searching concept [33]. This point generates the optimal GSF as a function of the α value at the stopping process, as shown in Figure 4. The optimal GSF is given when the iteration as a function of the α range is stopped at the maximum found S value.
Step #2: Deconvolution optimization by iteration of trained image quality metric.
Once the optimized GSF is computed in Step #1, the algorithm keeps running Step #2 by comparing the NIQE image quality metric with the trained model (detailed in Section 2.4) as a function of the iteration number (IT). Once the maximum NIQE score is found, the iteration stops and the program generates the final restored image. The optimal IT corresponds to the maximum found NIQE score that established the optimized stopping criterion at which the final deconvolution can be performed as a function of two optimal parameters: glare angle of the GSF and iteration number of the deconvolution, as shown in Figure 5. The algorithm was written and compiled in MATLAB 2019b (The Mathworks Inc., Natick, MA, USA).

3. Results

3.1. Testing the ITSD Algorithm with a Straylight Eye Model Optical Simulation

Optical human eye models are usually employed to evaluate the performance and design of intraocular lenses [34], soft contact lenses, or to explore ocular aberrations [35]. Our group recently reported a scattering eye model to quantify IS by direct optical methods [36]. In this work, we modified this model (details on the straylight eye model can be found elsewhere [36]) to incorporate cataracts as straylight source, given the impact of cataracts grade on retinal fundus imaging. Figure 6 shows a shaded representation of the IS model (Figure 6a) incorporating a ray-tracing of a collimated gaussian beam focused on the retina of an emmetropic eye. The straylight model was applied to the entire anterior surface of the anterior crystalline lens. Figure 6b–d show the retinal focused beam for null, 5° and 10° glare angles, respectively. Optical simulations were performed using Zemax optical design software (Zemax OpticStudio, LCC, Arlington Capital Partners, Washington, DC, USA).
The modified eye model with a cataract straylight source was employed to generate retinal images of a Siemens star (Figure 7a) by employing the retinal surface as a detector. Figure 7b–d shows the retinal images of the visual test corresponding to different straylight angles (400 × 400 pixels size). The reduction in image contrast, relative illuminance, and sharpness is visually evident. The perceptual differences in the degraded images reveal drastic impoverishment of the image quality because of the veiling glare originated from the straylight cataract source.
After applying Step #1, the algorithm was compared with and without applying the optimization process, as described in Step #2. For a direct comparison of our method with “regular” deconvolution processing, Figure 8a was carried out without iteration-optimization (Step #2 “OFF”) and Figure 8c shows the results after optimization activation (Step #2 “ON”). Notice that both non-optimized (Figure 8b) and iteration-optimized (Figure 8d) restored images show similar SSIM (0.94 and 0.94) and Sharpness (0.17 and 0.17) values, respectively, when compared to the original reference. However, without a stopping criterion, the iteration number must be set by the user and, in the case of IT = 2000, the computational cost exceeds the requirements for deconvolution by 21 times when using iteration optimization.
Table 1 shows the results of the algorithm testing by optical simulations of straylight retinal images as a function of the glare angle. Induced and detected glare angle α, optimal iteration number IT, Sharpness, and SSIM before and after deconvolution (S_Original, SSIM_Degraded, S_Restored, and SSIM_Restored, respectively) are compared. The computational costs (CC) were tested running MATLAB 2019b version installed in a Intel® Core™ i5-6500 processor at 3.20 GHz and 16 GB RAM. For example, the CC to restore Figure 9d was 50.51 s compared to Figure 9e that required only 2.39 s when the optimization process (Step #2) is activated.
For the sake of completeness of the analysis, four different popular image enhancement functions available in MATLAB Image Processing toolbox 2019b: “imsharpen,” “localcontrast,” “imadjust,” and “histeq” were used to increase the image quality of the Siemens Star degraded with stray light α = 5° (see Figure 7b). Optimal setting parameters for “imsharpen,” “localcontrast,” and “imadjust” functions were found by direct search to maximize the SSIM value of the enhanced image. The enhanced images and corresponding SSIM maps calculated in reference to an original undegraded Siemens Star are shown in Figure 9.

3.2. ITSD of Retinal Images

The previous section showed the performance assessment of the ITSD algorithm with theoretical images generated in a cataract straylight source eye model. Both the detection of the straylight angle and optimization of the IT allow restoring the degraded images from the VG effect. This section tests the ITSD algorithm in real fundus images acquired from healthy subjects and glaucoma and diabetic retinopathy patients.

3.2.1. ITSD of Retinal Images from Healthy Subjects

Figure 10a shows the original image acquired from healthy subjects with calculated sharpness of S = 0.01. Even assuming that no straylight pathological sources are affecting the image quality, the algorithm detected a glare angle of α = 1° and optimal IT = 20. The result of the deconvolution returns a sharpness enhancement of 1000% (S = 0.11), as shown in Table 2. Visually, the perception of fine details, resolution, and contrast is prominently improved. The segmentation processing (Figure 10b,d) was carried out by employing the same thresholding, providing a direct comparison of the structures that can be edge-detected. The optical quality assessment of the original and restored images quantified by the BRISQUE metric and the corresponding enhancement are shown in Table 2.

3.2.2. ITSD of Retinal Images from Glaucomatous Eyes

The results of the deconvolution applied in a representative case of a glaucomatous eye are shown in Figure 11. The original image (Figure 11a) does not provide resolution enough to perform a visual analysis without missing important fundus features, vessels, and the optic disk. In this pathological case, the ITSD algorithm detected a greater glare angle (α = 4°) compared to the control sample that also required higher iterations (IT = 70) to achieve the optimal deconvolution. The restored image provided 63.66% enhanced sharpness (Soriginal = 0.041 and Srestored = 0.067, respectively) (see Table 2). The edge detection processing revealed a segmented image after deconvolution with a significant spatially-resolved structure including the optic disk and vasculature. In the case of the glaucomatous sample, the restoration process achieved an image quality improvement of 923% measured by the BRISQUE metric.

3.2.3. ITSD of Retinal Images from Diabetic Retinopathy Patients

Finally, the results for a diabetic retinopathy eye provides spatially-resolved information about hidden fundamental structures in retinopathy, such as drusen. The ITSD algorithm detected a straylight of α = 6° and needed 90 iterations to optimize the deconvolution process that returned a 700% improved (Table 2) image sharpness (Original = 0.01 and Restored = 0.08). It is especially noteworthy that the achieved enhancement can be translated into automatic detection of drusen through segmentation processing, as shown in edge-detection images (Figure 12b,d). In this pathological sample, we found the maximum image quality enhancement (1263%) assessed by the BRISQUE metric.
Table 2 summarizes the results of the restoration of the three analyzed groups (healthy, glaucoma, and diabetic retinopathy patients). The ITSD algorithm detected higher glare angles in the pathological samples that correspond to higher computational processing (i.e., higher IT and CC). However, the sharpness improvement seems to not be correlated with improved optical quality measured by BRISQUE.
Finally, Figure 13 shows the synthetic GSF reconstructions after finding, in Step #1, the optical glare angle for the samples of Figure 11, Figure 12 and Figure 13 (upper row) and how the reconstructed GSFs (named Learned GSFs) look like in the end of the iterations (i.e., when the optimization converges in Step #2).

4. Discussion and Conclusions

The optical quality of retinal images is affected mainly by optical aberrations and scattering effects that blur the imaged plane and degrade the relative illuminance and image contrast and resolution [1,2,3,4]. Adaptive optics (AO) has been widely employed for the compensation of wave-front errors [1] potentially improving the resolution of retinal images. However, even using AO, uncorrected residual aberrations can lead to errors when the retinal fundus is being analyzed and classified [37]. In this sense, deconvolution methods have been employed to restore retinal images from optical aberrations and defocus [18] effects, multichannel blind deconvolution for color retinal images restoration [16], to enhance AO retinal images [37,38], or to compensate uneven retinal illumination [39].
Unlike image enhancement techniques, restoration techniques, as deconvolution algorithms are, have the potential to improve image quality without adding image artifacts that can be harmful for clinical diagnosis purposes. In general, this aspect can be clearly seen in the results exposed in Figure 9. In particular, Figure 9g,h show how histogram equalization (“histeq”) increases the dynamical range (in comparison with the degraded image (7b)), but it also shows the existence of undesirable halos around the central and peripheral areas of the processed image (in comparison with Figure 8a).
Nevertheless, Figure 9a–e shows how, even when function parameters are optimized to achieve the highest SSIM in reference to the original undegraded image, enhancement techniques are not able to achieve the image quality of restoration techniques (see Figure 8). Special mention must be done to enhance with “localcontrast” (Figure 9c,d) where it can be seen how global contrast enhancement and a very good SSIM (0.89) comes with undesirable image artifacts in the central region of the processed Siemens Star.
Considering the intraocular scattering as an important degradation factor, it is clear the increased IS in the presence of cataracts, lens opacifications, or even after cataract surgery [40,41]. However, retinal manifestations of systemic diseases such as diabetes mellitus can increase IS in a less clear way [42,43]. Considering the relationship between increased IS and the severity of diabetic retinopathy makes straylight measurement a sensitive test for detecting diabetes mellitus through retinal images. In this sense, our method provides a powerful tool to detect veiling glare degradation in retinal images and may allow clinicians to suspect unnoticed intraocular straylight sources in those cases in which an intraocular straylight meter is not available. In our work, the GSF [27,28] is proposed as a degradation model within a Richardson-Lucy [29] deconvolution method defined as a semi-blind approach as the degradation operator is IS but of unknown parametrization. The algorithm incorporates an optimization process that searches for both the optimal glare angle and iteration number by employing a no-reference image quality metric iterative-trained procedure.
Our results have revealed the proposed algorithm as a useful tool to compensate VG in retinal images whose origin lies on IS. The ITSD algorithm have been tested with both simulated retinal images generated from an IS eye model [36] and retinal image dataset from healthy subjects and patients with retinal pathologies (glaucoma and diabetic retinopathy).
For all, the iterative-trained deconvolution process, through the compensation for IS, was able to significantly improve the image sharpness up to 1000% in healthy samples and up to 700% in diabetic retinopathy fundus images, unveiling hidden retinal structures unobservable in the original images. On the other hand, the image quality of the restored images was evaluated with a non-reference image quality metric (BRISQUE) [23] since it is usually employed in blind image quality evaluation. BRISQUE metric revealed an image quality improvement after restoration up to 1263% in pathological retinal images (see Table 2).
The ITSD algorithm detected higher glare angles in the pathological samples, which is consistent with the findings reported in the literature [40,41]. Results showed that the higher the amount of detected IS, the higher the computational costs for a successful restoration is (Table 2). In addition, the sharpness improvement, which can be associated with human perception, seems to not be correlated with a non-reference image quality assessment (BRISQUE), as shown in Table 2, which means that human vision does not necessarily correlate computer vision and this fact becomes relevant in clinical diagnosis, since human perception can confuse image artifacts with better resolution. In this sense, our optimization process based on the stopping criterion ensures reducing the risk of generating artifacts in the final restored image as a consequence of the deconvolution executed in a semi-blind approach.
To conclude, in this work, we have developed a new Richardson-lucy based algorithm by means of iterative-trained semi-blind deconvolution approach to restore the veiling glare from retinal images as a consequence of intraocular straylight, defining theoretical GSFs, according to CIE standards [25,26]. Future improvements of the current version of the algorithm could incorporate more sophisticated Artificial Intelligence optimization modules for modelling realistic and predictive PSFs of the complex degradation operator of the eye.

Author Contributions

Conceptualization, F.J.Á. Methodology, F.J.Á., J.A., M.V.C., and L.R. Software, F.J.Á. and J.A. Validation, F.J.Á. and J.A. Formal analysis, F.J.Á. Investigation, F.J.Á., J.A., M.C.M., and L.R. Writing—original draft preparation, F.J.Á. Writing—review and editing, F.J.Á., J.A., M.V.C. and L.R. Supervision, F.J.Á. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by “Gobierno de Aragón. Grupo E44_20R.”

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly archived datasets can be found in: https://www5.cs.fau.de/research/data/fundus-images/ (Accessed on 1 January 2021).

Acknowledgments

Authors thank the Pattern Recognition Lab (CS5), the Department of Ophthalmology, Friedrich-Alexander University Erlangen-Nuremberg (Germany), and the Brno University of Technology, Faculty of Electrical Engineering and Communication, Department of Biomedical Engineering, Brno (Czech Republic) for providing the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wanek, J.; Mori, M.; Shahidi, M. The effect of aberrations and scatter on image resolution assessed by adaptive optics retinal section imaging. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 2010, 24, 1296–1304. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Hope, G.M. Night Myopia. Sur. Ophthalmol. 1984, 29, 129–136. [Google Scholar] [CrossRef]
  3. Marin-Franch, I.; Xu, R.; Bradley, A.; Thibos, L.N.; Lopez-Gil, N. The effect of spherical aberration on visual performance and refractive state for stimuli and tasks typical of night viewing. J. Optom. 2018, 11, 144–152. [Google Scholar] [CrossRef] [PubMed]
  4. Van der Berg, T.J. Analysis of intraocular straylight, especially in relation to age. Optom. Vis. Sci. 1995, 72, 52–59. [Google Scholar] [CrossRef]
  5. Lapid-Gortzak, R.; Labuz, G.; van der Meulen, I.J.; van der Linden, J.W.; Mourits, M.P.; van den Berg, T.J. Straylight Measurements in Two Different Apodized Diffractive Multifocal Intraocular Lenses. J. Refract. Surg. 2015, 31, 746–751. [Google Scholar] [CrossRef] [Green Version]
  6. Łabuz, G.; Reus, N.J.; van den Berg, T.J. Comparison of ocular straylight after implantation of multifocal intraocular lenses. J. Cataract Refract. Surg. 2016, 42, 618–625. [Google Scholar] [CrossRef]
  7. Łabuz, G.; Reus, N.J.; van den Berg, T.J. Light scattering levels from intraocular lenses extracted from donor eyes. J. Cataract Refract. Surg. 2017, 43, 1207–1212. [Google Scholar] [CrossRef]
  8. Gaurisankar, Z.S.; van Rijn, G.A.; Luyten, G.P.; van den Berg, T.J. Straylight As The Result Of Refractive Correction. Clin. Ophthalmol. 2019, 12, 2195–2201. [Google Scholar] [CrossRef] [Green Version]
  9. Uji, A.; Ooto, S.; Hangai, M.; Arichika, S.; Yoshimura, N. Image Quality Improvement in Adaptive Optics Scanning Laser Ophthalmoscopy Assisted Capillary Visualization Using B-spline-based Elastic Image Registration. PLoS ONE 2013, 8, e80106. [Google Scholar] [CrossRef] [Green Version]
  10. Burns, S.A.; Elsner, A.E.; Sapoznik, K.A.; Warner, R.L.; Gast, T.J. Adaptive optics imaging of the human retina. Prog. Retin. Eye Res. 2019, 68, 1–30. [Google Scholar] [CrossRef]
  11. Kanse, S.S.; Yadav, D.M. Retinal fundus image for glaucoma detection: A review and study. J. Intell. Syst. 2019, 28, 43–56. [Google Scholar] [CrossRef]
  12. Hassan, G.; El-Bendary, N.; Hassanien, A.E.; El-Bendary, N.; Fahmy, A. Retinal blood vessel segmenetation approach based on mathematical morphology. Proc. Comp. Sci. 2015, 65, 612–622. [Google Scholar] [CrossRef] [Green Version]
  13. Köse, C.; Sevik, U.; Calioglu, G. Automatic segmentation of age-related macular degeneration in retinal fundus images. Comp. Biol. Med. 2008, 38, 611–619. [Google Scholar] [CrossRef]
  14. García-Guerra, C.E.; Martínez-Roda, J.; Aldaba, M.; Diaz-Douton, F.; Vilaseca, M.; Vohnsen, B.; Pujol, J. Method to reduce undesired multiple fundus scattering effects in double-pass systems. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 2019, 36, 918–924. [Google Scholar] [CrossRef]
  15. Fei, X.; Zhao, J.; Zhao, H.; Yun, D.; Zhang, Y. Deblurring adaptive optics retinal images using deep convolutional neural networks. Biomed. Opt. Express 2017, 8, 5675–5687. [Google Scholar] [CrossRef] [Green Version]
  16. Marrugo, A.; Millan, M.; Sorel, M.; Sroubek, F.; Millan, M.S. Retinal image restoration by means of blind deconvolution. J. Biomed. Opt. 2011, 16, 116016. [Google Scholar] [CrossRef] [Green Version]
  17. Nourrit, V.; Vohnsen, B.; Artal, P. Blind deconvolution for high-resolution confocal scanning laser ophthalmoscopy. J. Opt. A Pure Appl. Opt. 2005, 7, 585–592. [Google Scholar] [CrossRef]
  18. Yang, S.; Barriga, S.; Erry, G.; Raman, B.; Russell, S.; Mitra, S.; Soliz, P. Retinal image deconvolution: Revealing hidden structures and pathology. Investig. Ophthalmol. Vis. Sci. 2005, 46, 361. [Google Scholar]
  19. Ayers, G.R.; Dainty, J.C. Iterative blind deconvolution method and its applications. Opt. Lett. 1988, 13, 547–549. [Google Scholar] [CrossRef] [Green Version]
  20. Fish, D.; Brinicombe, A.M.; Pike, E.R. Blind Deconvolution by means of Richardson-Lucy algorithm. J. Opt. Soc. Am. A 1995, 2, 58–65. [Google Scholar] [CrossRef] [Green Version]
  21. Budai, A.; Bock, R.; Maier, A.; Hornegger, J.; Michelson, G. Robust Vessel Segmentation in Fundus Images. Int. J. Biomed. Imaging 2013, 2013, 154860. [Google Scholar] [CrossRef] [Green Version]
  22. Maître, H. Image Quality. In From Photon to Pixel: The Digital Camera Handbook, 2nd ed.; Wiley: Hoboken, NJ, USA, 2015; pp. 201–205. [Google Scholar]
  23. Mittal, A.; Soundararajan, R.; Bovi, A.C. Making a Completely Blind Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 22, 209–212. [Google Scholar] [CrossRef]
  24. Zhou, W.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.C. Image Qualifty Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2014, 13, 600–612. [Google Scholar]
  25. Canny, J. A Computational Approach to Edge Detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef]
  26. Van den Berg, T.J.; IJspeert, J.K.; de Waard, P. Dependence of intraocular straylight on pigmentation and light transmission through the ocular wall. Vision Res. 1991, 31, 1361–1367. [Google Scholar] [CrossRef]
  27. Vos, J.J. Disability glare—A state of the art report. Comm. Int. l’Eclairage J. 1984, 3, 39–53. [Google Scholar]
  28. Vos, J.; van den Berg, T.J. Report on Disability Glare. CIE Collect. 1999, 135, 1–9. [Google Scholar]
  29. Richardson, W. Bayesian-Based Iterative Method of Image Restoration. J. Opt. Soc. Am. 1972, 62, 55–59. [Google Scholar] [CrossRef]
  30. Lucy, L.B. An Iterative Technique for the Rectification of Observed Distributions. Astron. J. 1974, 79, 745. [Google Scholar] [CrossRef] [Green Version]
  31. Shepp, L.A.; Vardi, Y. Maximum Likelihood Reconstruction for Emission Tomography. IEEE Trans. Med. Imaging 1974, 1, 113–122. [Google Scholar] [CrossRef]
  32. Laasmaa, M.; Vendelin, M.; Peterson, P. Application of regularized Richardson-Lucy algorithm deconvolution of confocal microscopy images. J. Microsc. 2011, 243, 124–140. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Lozano, M.; Herrera, F.; Krasnogor, N.; Molina, D. Real-Coded Memetic Algorithms with Crossover Hill-Climbing. Evol. Comput. 2004, 12, 273–302. [Google Scholar] [CrossRef] [PubMed]
  34. Pérez-Gracia, J.; Ávila, F.; Ares, J.; Vallés, J.; Remón, L. Misaligment and tilt effect on aspheric intraocular lens designs after a corneal refractive surgery. PLoS ONE 2020, 15, e0243740. [Google Scholar] [CrossRef] [PubMed]
  35. Smith, G.; Bedggood, P.; Ashman, R.; Daaboul, M.; Metha, A. Exploring ocular aberrations with a schematic human eye model. Optom. Vis. Sci. 2008, 85, 330–340. [Google Scholar] [CrossRef]
  36. Ávila, F.; Collados, M.V.; Ares, J.; Remón, L. Wide-field direct ocular straylight meter. Opt. Express 2020, 28, 11237–11242. [Google Scholar] [CrossRef]
  37. Christou, J.C.; Roorda, A.; Williams, D. Deconvolution of adaptive optics retinal images. J. Opt. Soc. Am. A 2004, 21, 1393–1401. [Google Scholar] [CrossRef]
  38. Blanco, L.; Mugnier, L.M. Marginal blind deconvolution of adaptive optics retinal images. Opt. Express 2011, 19, 23227–23239. [Google Scholar] [CrossRef] [Green Version]
  39. Marrugo, A.G.; Vargas, R.; Contreras, S.; Millán, M. On the compensation of uneven illumination in retinal images for restoration by means of blind deconvolution. In Proceedings of the 016 XXI Symposium on Signal Processing, Images and Artificial Vision (STSIVA), Bucaramanga, Colombia, 31 August–2 September 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–4. [Google Scholar] [CrossRef]
  40. Kretz, F.T.; Tandogan, T.; Khoramnia, R.; Auffarth, G.U. High order aberration and straylight evaluation after cataract surgery with implantation of an aspheric, aberration correcting monofocal intraocular lens. Int. J. Ophthalmol. 2015, 8, 736–741. [Google Scholar]
  41. Vlasman, J.M.; van den Berg, T.J.; Reus, N.J. Straylight due to intraocular lens opacification in a patient with asteroid hyalosis. Am. J. Ophthalmol. Case Rep. 2020, 19, 1000857. [Google Scholar]
  42. Hwang, H.B.; Yim, H.B.; Chung, S.K. Effect of diabetic retinopathy and diabetes on the intraocular straylight in pseudophakic eyes. BMC Ophthalmol. 2015, 15, 1–7. [Google Scholar] [CrossRef] [Green Version]
  43. Liu, J.; Wang, X.; Wang, J.; Guo, H. Optical quality and intraocular scattering in the diabetic eye wihtout diabetic retinopathy. Optom. Vis. Sci. 2019, 96, 248–255. [Google Scholar] [CrossRef]
Figure 1. Retinal images from the dataset corresponding to a healthy (a), glaucomatous (b), and diabetic retinopathy (c) eyes.
Figure 1. Retinal images from the dataset corresponding to a healthy (a), glaucomatous (b), and diabetic retinopathy (c) eyes.
Jimaging 07 00073 g001
Figure 2. Schematic diagram of the trained Natural Image Quality Evaluator (NIQE) score computation.
Figure 2. Schematic diagram of the trained Natural Image Quality Evaluator (NIQE) score computation.
Jimaging 07 00073 g002
Figure 3. Image processing segmentation procedure for edge detection of retinal images. First, the RGB is converted to grayscale and binarized and then the Canny filter is applied.
Figure 3. Image processing segmentation procedure for edge detection of retinal images. First, the RGB is converted to grayscale and binarized and then the Canny filter is applied.
Jimaging 07 00073 g003
Figure 4. Step #1 of the algorithm.
Figure 4. Step #1 of the algorithm.
Jimaging 07 00073 g004
Figure 5. Step #2 of the algorithm.
Figure 5. Step #2 of the algorithm.
Jimaging 07 00073 g005
Figure 6. Shaded representation of the IS eye model. (a) Retinal projection of an incident collimated Gaussian beam of the IS-free eye. (b) Cataract modelled eye with α = 0°. (b) α = 5° (c) and α = 10°. (d) Glare angles.
Figure 6. Shaded representation of the IS eye model. (a) Retinal projection of an incident collimated Gaussian beam of the IS-free eye. (b) Cataract modelled eye with α = 0°. (b) α = 5° (c) and α = 10°. (d) Glare angles.
Jimaging 07 00073 g006
Figure 7. Theoretical eye model retinal images corresponding to different straylight angles. (a): Original; (b): α = 5º; (c): α = 7º; (d): α = 10º.
Figure 7. Theoretical eye model retinal images corresponding to different straylight angles. (a): Original; (b): α = 5º; (c): α = 7º; (d): α = 10º.
Jimaging 07 00073 g007
Figure 8. Restored image without iteration optimization (a), restored image after iteration optimization (c), and SSIM maps (b) computed from Figure 7a and Figure 8a and (d) from Figure 7a and Figure 8c.
Figure 8. Restored image without iteration optimization (a), restored image after iteration optimization (c), and SSIM maps (b) computed from Figure 7a and Figure 8a and (d) from Figure 7a and Figure 8c.
Jimaging 07 00073 g008
Figure 9. (a,c,e,g) Digitally enhanced images with popular MATLAB functions from a Siemens Star image degraded with stray-light (α = 5º, Figure 7b). (b,d,f,h) Corresponding SSIM (see Section 2.2.3) maps in relation with the original undegraded Siemens Star image shown in Figure 7a. (a) Image enhanced with function “imsharpen” with parameters: ‘amount’ = 2, ‘radius’ = 30, ‘threshold’ = 1. (c) Image enhanced with function “localcontrast” with parameters: ‘edgethreshold’ = 0.7, ‘amount’ = 1. (e) Image enhanced with function “imadjust” with parameters: [low_in high_in] = strechlim, ‘gamma’ = 6.1, (f) Image enhanced with histogram equalization function “histeq”. The function parameters for “imsharpen,” “localcontrast,” and “imadjust” were optimized to achieve a maximum SSIM.
Figure 9. (a,c,e,g) Digitally enhanced images with popular MATLAB functions from a Siemens Star image degraded with stray-light (α = 5º, Figure 7b). (b,d,f,h) Corresponding SSIM (see Section 2.2.3) maps in relation with the original undegraded Siemens Star image shown in Figure 7a. (a) Image enhanced with function “imsharpen” with parameters: ‘amount’ = 2, ‘radius’ = 30, ‘threshold’ = 1. (c) Image enhanced with function “localcontrast” with parameters: ‘edgethreshold’ = 0.7, ‘amount’ = 1. (e) Image enhanced with function “imadjust” with parameters: [low_in high_in] = strechlim, ‘gamma’ = 6.1, (f) Image enhanced with histogram equalization function “histeq”. The function parameters for “imsharpen,” “localcontrast,” and “imadjust” were optimized to achieve a maximum SSIM.
Jimaging 07 00073 g009
Figure 10. Original image from a healthy subject. (a) Edge segmentation of original image. (b) Restored image (c) and edge segmentation of restored image (d).
Figure 10. Original image from a healthy subject. (a) Edge segmentation of original image. (b) Restored image (c) and edge segmentation of restored image (d).
Jimaging 07 00073 g010
Figure 11. Original image from a glaucomatous eye. (a) Edge segmentation of original image. (b) Restored image (c) and edge segmentation of restored image (d).
Figure 11. Original image from a glaucomatous eye. (a) Edge segmentation of original image. (b) Restored image (c) and edge segmentation of restored image (d).
Jimaging 07 00073 g011
Figure 12. Original image from a diabetic retinopathy eye. (a) Edge segmentation of original image. (b) Restored image (c) and edge segmentation of restored image (d).
Figure 12. Original image from a diabetic retinopathy eye. (a) Edge segmentation of original image. (b) Restored image (c) and edge segmentation of restored image (d).
Jimaging 07 00073 g012
Figure 13. Reconstructed Glare Spread Function (GSFs) for the optical α values found in Step #1 for the healthy (a), glaucoma (b), and diabetic retinopathy (c) samples and generated GSFs after the optimization converges (optimal iteration number in Step #2) (df).
Figure 13. Reconstructed Glare Spread Function (GSFs) for the optical α values found in Step #1 for the healthy (a), glaucoma (b), and diabetic retinopathy (c) samples and generated GSFs after the optimization converges (optimal iteration number in Step #2) (df).
Jimaging 07 00073 g013
Table 1. Results of the algorithm testing in simulated images. Induced glare angle (Induced α). Detected glare angle: (Detected α). Optimal iteration number. (IT); Original image sharpness: (S_Original) Original. SSIM of the degraded image: (SSIM_Degraded). Sharpness of the restored image (S_Restored). Restored SSIM: (SSIM_Restored). Computational costs: (CC).
Table 1. Results of the algorithm testing in simulated images. Induced glare angle (Induced α). Detected glare angle: (Detected α). Optimal iteration number. (IT); Original image sharpness: (S_Original) Original. SSIM of the degraded image: (SSIM_Degraded). Sharpness of the restored image (S_Restored). Restored SSIM: (SSIM_Restored). Computational costs: (CC).
Induced αDetected αITS_OriginalSSIM_DegradedS_RestoredSSIM_RestoredCC (s)
1190.1400.6300.180.971.87
22150.0690.2000.240.921.96
32220.0670.1900.270.932.07
43310.0100.1500.190.922.22
55400.0040.1100.170.952.39
66640.0030.0130.180.912.83
76640.0290.0160.170.762.83
87700.0030.0090.160.753.05
99440.0020.0080.100.572.46
109740.0010.0070.070.362.96
Table 2. Results of the deconvolution processing in real retinal images. Detected glare angle: (Detected α). Optimal iteration number (IT). Sharpness of the original image (S_Orig.). Sharpness of the restored image (S_Rest). BRISQUE (see Section 2.2.4) of the original image (BRISQUE_Orig.). BRISQUE of the restored image (BRISQUE_Rest). Computational costs (CC).
Table 2. Results of the deconvolution processing in real retinal images. Detected glare angle: (Detected α). Optimal iteration number (IT). Sharpness of the original image (S_Orig.). Sharpness of the restored image (S_Rest). BRISQUE (see Section 2.2.4) of the original image (BRISQUE_Orig.). BRISQUE of the restored image (BRISQUE_Rest). Computational costs (CC).
ParameterHealthyGlaucomaD. Ret.
Detected α1.004.006.00
IT20.0070.0090.00
S_Orig.0.010.040.01
S_Rest.0.110.070.08
S_Improv. (%)1000.0063.41700.00
Brisque_Orig.38.4229.0229.54
Brisque_Rest.43.3439.2543.17
Brisque_Improv. (%)12.80923.001263.00
CC (s)2.233.683.87
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ávila, F.J.; Ares, J.; Marcellán, M.C.; Collados, M.V.; Remón, L. Iterative-Trained Semi-Blind Deconvolution Algorithm to Compensate Straylight in Retinal Images. J. Imaging 2021, 7, 73. https://doi.org/10.3390/jimaging7040073

AMA Style

Ávila FJ, Ares J, Marcellán MC, Collados MV, Remón L. Iterative-Trained Semi-Blind Deconvolution Algorithm to Compensate Straylight in Retinal Images. Journal of Imaging. 2021; 7(4):73. https://doi.org/10.3390/jimaging7040073

Chicago/Turabian Style

Ávila, Francisco J., Jorge Ares, María C. Marcellán, María V. Collados, and Laura Remón. 2021. "Iterative-Trained Semi-Blind Deconvolution Algorithm to Compensate Straylight in Retinal Images" Journal of Imaging 7, no. 4: 73. https://doi.org/10.3390/jimaging7040073

APA Style

Ávila, F. J., Ares, J., Marcellán, M. C., Collados, M. V., & Remón, L. (2021). Iterative-Trained Semi-Blind Deconvolution Algorithm to Compensate Straylight in Retinal Images. Journal of Imaging, 7(4), 73. https://doi.org/10.3390/jimaging7040073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop