Next Article in Journal
Driver Drowsiness Detection: A Machine Learning Approach on Skin Conductance
Previous Article in Journal
Integrating Visual and Network Data with Deep Learning for Streaming Video Quality Assessment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Historical Text Image Enhancement Using Image Scaling and Generative Adversarial Networks

1
Multimedia Information Processing Lab, Department of Information and Communication Engineering, Chosun University, Gwangju 61452, Republic of Korea
2
Department of Information System, Prince Sattam bin Abdulaziz University, Al-Kharj 16278, Saudi Arabia
3
Department of Computer Engineering, Gachon University, Seongnam-si 13120, Republic of Korea
4
Department of Robotics, Hanyang University, Ansan-si 15558, Republic of Korea
5
Faculty of Engineering, University Malaysia Sarawak, Kota Samarahan 94300, Malaysia
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(8), 4003; https://doi.org/10.3390/s23084003
Submission received: 1 February 2023 / Revised: 30 March 2023 / Accepted: 3 April 2023 / Published: 14 April 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Historical documents such as newspapers, invoices, contract papers are often difficult to read due to degraded text quality. These documents may be damaged or degraded due to a variety of factors such as aging, distortion, stamps, watermarks, ink stains, and so on. Text image enhancement is essential for several document recognition and analysis tasks. In this era of technology, it is important to enhance these degraded text documents for proper use. To address these issues, a new bi-cubic interpolation of Lifting Wavelet Transform (LWT) and Stationary Wavelet Transform (SWT) is proposed to enhance image resolution. Then a generative adversarial network (GAN) is used to extract the spectral and spatial features in historical text images. The proposed method consists of two parts. In the first part, the transformation method is used to de-noise and de-blur the images, and to increase the resolution effects, whereas in the second part, the GAN architecture is used to fuse the original and the resulting image obtained from part one in order to improve the spectral and spatial features of a historical text image. Experiment results show that the proposed model outperforms the current deep learning methods.

1. Introduction

Document Analysis and Recognition (DAR) research is highly focused on protecting and improving the readability of historical text images. Historical documents are an invaluable part of our cultural heritage, having a significant scientific and cultural value. Historical documents typically contain important information and are extremely valuable [1]. The digitization of ancient texts is an essential step toward protecting the literature. However, manual processing of these large volumes of documents takes time and effort. As a result, it is necessary to use cutting-edge image processing and machine learning methods to process historical text documents automatically. Recent text documents can easily be enhanced. However, due to degradation such as noise, fading, artifacts, molds, water stains, stamps, contrast, brightness, and so on, historical text images are much more difficult to enhance. These distortions may have severe impact on post-processing steps such as Optical Character Recognition (OCR), segmentation, indexing, and information spotting.
The Convolutional Neural Network (CNN) model has recently gained popularity in DIE-related subtasks such as binarization [2,3], de-fading [4], and de-blurring [5]. Despite the fact that CNN and its family of algorithms outperform previous image processing techniques, they have some flaws. CNNs restore different regions using the same convolution filters, resulting in spectral information loss. Additionally, CNNs fail to extract low-level spatial features from text images.
With the advent of deep learning (DL), several research studies applied cutting-edge deep-learning-based models to various image processing and computer vision tasks, such as image enhancement [3,6,7], object detection [8], image text segmentation [9], etc. DL methods have been shown to deliver promising results and outperform traditional methods. Likewise, in recent years, DL-based methods for text image enhancement have received a lot of attention.
The motivation behind this study is to present a hybrid image processing and DL-based method to enhance historical text images. Our contributions are twofold. First, the transformation method is used to reduce noise and blurring effects while increasing intensity resolution. A new bi-cubic interpolation of Lifting Wavelet Transform (LWT) and Stationary Wavelet Transform (SWT) is employed to enhance image resolution and to reduce noise and blurring effects. Second, the GAN architecture is employed to extract the spatial and spectral features of a historical text image.
The rest of the paper is organized as follow: Section 2 introduces the related work on text image enhancement. Our proposed method is explained in Section 3. Section 4 illustrates qualitative and quantitative experiments and provides the discussion, and Section 5 presents the conclusion and future directions.

2. Related Work

Text in historical documents is difficult to read due to degradation. The Document Analysis and Recognition (DAR) community has paid much attention to text image enhancement. Documents enhancement problems consist of several tasks such as binarization, de-blurring, de-noising, de-fading, watermark removal, and shadow removal. This research work is aimed to solve the problem of image de-blurring, image de-noising, and image de-fading. Figure 1 visually represents these problems.
Similarly, different datasets are available for text document enhancement tasks. Table 1 illustrates the specification of these datasets.
There are numerous challenges associated with historical text image enhancement, such as text enhancement without affecting the background because the background contains important information or objects that must be preserved. Some higher-resolution images because the algorithm to slow down and we need a memory-efficient algorithm to process the image. A flash light is now an essential part of image acquisition. When a flash strikes on an object, the light reflection distorts the image and reduces readability. Furthermore, various algorithms present ideas on how to improve the image’s edges and curves, allowing for enhanced text. Text images with low resolution require enhancement for readability as well, but we must first improve the image’s resolution.
A simple thresholding method was first used for text image enhancement. In this method, a threshold value was used to segment or to classify the background and foreground images [10,11]. These techniques have been progressing in the past few years, by using machine learning algorithms such as support vector machines (SVM) [12]. Subsequently, methods based on energy were presented to monitor the text pixels by optimizing their energy function and minimizing the effect of the degraded image background. However, the results obtained using those methods were unsatisfactory [13,14].
An idea was introduced in [15] proposing a method to enhance the contrast of the image by preventing the over-enhancement problem. In this method, the original image was decomposed using Discrete Wavelet Transform (DWT), and then the CLAHE technique was applied to the low-frequency component. Finally, weighted average was taken from the original and the resultant image to obtain an enhanced image. However, this method is not suitable for grayscale images and fails in the image de-fading task.
Using deep learning approaches, the problems of image de-blurring, de-fading, and de-nosing were resolved by learning the enhancement directly from the original data. A pixel’s classification method was proposed in [16] to classify each pixel as white or black depending on the neighboring pixels. For this task, a 2D long- and short-term memory was trained. However, it is a time-consuming process.
Convolutional Neural Network (CNN) auto-encoders provide a more practical solution for image de-noising. As a result, they have become the leading edge of recent advances in image de-noising [17] and, more specifically, in text image enhancement operations such as de-blurring [5] and image binarization [2].
A full CNN model was suggested in [18] to binarize deteriorated images at various image scales. Likewise, a pre-trained U-Net model was presented in [19] to learn binarization with less data. A conditional generative adversarial network (GAN) was introduced in [20] for various image enhancement tasks such as binarization, de-blurring, and watermarks removal. A modified version of [20] was presented in [21] by using dual discriminator to evaluate text readability. Similar cGAN’s techniques were presented in [22,23,24].
Recently, transformer-based architectures were used in few papers for natural image restoration and image de-warping [25]. However, the architectures that were used kept relying on the CNN method before moving to the transformer stage. A Document Image Transformer (DocTr) was proposed in [26] to solve the problem of geometry and illumination distortion of the document images. In this method, the global context of the document image was captured and pixel wise detection method was used to correct the distortion. Another method was proposed in [27] quantify the downstream analysis effects using empirical analysis. In this study, bais-variance decomposition of an empirical risk function and mean square error were measured using bootstrap data simulation. Again in [28] provides a report on the methodology and performance of the five submitted algorithms by various research groups across the world. Binarization and Classification were done using well know evaluation metrics. Similarly, in [29], a software framework, Binarization Shop, that combines a series of binarization approaches that have been tailored to exploit user assistance was presented. This method provides a practical approach for converting difficult documents to black and white. However, it was a time consuming process and takes long time for training. Contrarily, we proposed a novel model to reduce noise and blurring effects with increased pixels resolution. A new bi-cubic interpolation of Lifting Wavelet Transform (LWT) and Stationary Wavelet Transform (SWT) is employed to enhance pixels resolution and to reduce noise and blurring effects. Second, the GAN architecture is employed to extract the spatial and spectral features of a historical text image.

3. Materials and Methods

3.1. Data Collection

Four datasets were used in this study. These datasets are Noisy Office [30] database, S-MS (Synchromedia Multispectral Ancient Document Images) [27], Bishop Bickley diary [29], Tobacco 800 [31], and Blurry document Images. These datasets are also publicly available online. The details are given in Table 1.
Table 1. Datasets used for image enhancement tasks.
Table 1. Datasets used for image enhancement tasks.
DatabaseTaskNumber of ImagesReal/SyntheticImage Size (Resolution)
Noisy Office [30]Image de-noising288MixedVariable
S-MS [27]Mixed (de-noising, de-blurring, de-fading)3.4309Synthetic1001 × 330
Bishop Bickley diary [29]Image binarization07Real1050 × 1350
Tobacco 800 [31]Image de-noising1290Real1200 × 1600
Smart Doc-QA [32]Image de-blurring4260 351 × 292
Blurry document images [5]Image de-blurring35,000Synthetic300 × 300

3.2. Methodology

The proposed method is divided into two steps, which are explained below.

3.2.1. Pixels Resolution Enhancement

LWT Bi-Cubic Interpolation Method

In this part, the primary concept which is being followed to enhance resolution is wavelet transformation of the image. Three key concepts, Lifting Wavelet Transform (LWT), Stationary Wavelet Transform (SWT), and image interpolation, altogether sum up to contribute to the resolution enhancement. Initially, the input image is decomposed using a one-level LWT. The Daubechies transformation approach is used to decompose the image which is empirically determined for obtaining the best results. After decomposition, vertical, horizontal, diagonal, and approximation coefficients are obtained denoted by LH, HL, HH, and LL. Figure 2 shows the LWT coefficients representation.
Initially, the input image with resolution m×n is subjected to the LWT function which yields three high-frequency sub-bands, i.e., horizontal, vertical, and diagonal details of the image as shown in Figure 2. We chose the level of decomposition as “1”, because it produces good results. The only one low-frequency sub-band contains approximation coefficient. Then the high-frequency sub-bands are processed using bi-cubic interpolation process. The interpolation function is chosen as “2”. It zooms the image without decreasing the resolution. Figure 3 depicts the effects of bi-cubic interpolation.
LWT decomposes the m × n band into m × n/2 sub-bands. The interpolation yields the resultant sub-bands with m × n resolution.

Stationary Wavelet Transform (SWT)

LWT down-samples the sub-bands so as to generate high-frequency components, which cause information loss. To mitigate this loss, SWT is employed to minimize the down-sampling loss. The input image undergoes SWT and yields four sub-bands as generated by LWT. The sub-bands generated by SWT are of m × n resolution comprising of horizontal, vertical, and diagonal details.

Addition of Sub-Bands and Interpolation

The high-frequency sub-bands obtained from LWT and SWT are summed together. These sub-bands include interpolated H, V, D from LWT and H, V, D from SWT, such that the corresponding components are summed together. The resultant sub-bands obtained after adding the corresponding sub-bands from LWT and SWT are then subjected to further interpolation by some factor “γ”. Simultaneously the original image also undergoes interpolation by factor “γ”.

Inverse Lifting Wavelet Transform

After obtaining sub-bands by adding corresponding LWT and SWT sub-bands and applying an interpolation function, the inverse LWT function is applied to the interpolated original image and resulting sub-bands. A new image is obtained with the resolution of 2γ(m × n). Figure 4 shows the resolution enhancement architecture.
When the image is up-sampled, it surely loses its quality because of missing pixels. With bi-cubic interpolation, the image is filled with neighboring pixels and resized. In this method, the strength of bi-cubic interpolation and strengths of wavelet function ensure that the image is scaled up and the quality of the image is preserved. This method enhances the quality without the staircase effect and reduces noise, blurriness, and jaggedness.

3.2.2. Generative Adversarial Network (GAN)

The purpose of the GAN model is to extract the spatial and spectral features of a historical text image. It is employed to fuse the original and the resulting image achieved from part one to improve the spectral and spatial features of a historical text image. GAN is made up of two modules: generator and discriminator. The prime objective of GAN is to generate data patterns that are similar to the input data. The goal of the discriminator is to distinguish the resulted image from the source image. When the generator fools the discriminator, it indicates that the generator is well trained. Following the application of the GAN model, the resulting fused image contains all of the complementary information present in the input images, namely the fused image obtained from part 1 and the original image. Figure 5 depicts the entire image fusion process.
Initially, the input original and fused (obtained from part 1) images are given to generator. Then, fused and input images are compared. This approach generates a minimum/maximum game between the GAN modules. The generated fused image is again sent to discriminators. The discriminator receives the fused image and the original image as an input.
The generator uses a five-layer CNN with 3 × 3 filters in each layer. Each of these five convolution layers has a stride value of “1”. Convolution layers are used to extract feature maps from an input image. In generator architecture, batch normalization is used to make the model more stable. In all layers of the generator, “Leaky Relu” is employed except the last layer for activation. In the last layer, “tanh” function is used. The “tanh” and “Leaky Relu” functions are expressed in Equations (1) and (2).
L k y R l u = max ( 0.1 x , x ) ,
t n h = t n h ( x )
The discriminator is made up of five layers of CNN with 3 × 3 filters. The final filter has only one 1 × 1 filter. Each of these five convolution layers has a stride value of “2”. In discriminator architecture, batch normalization is used between the first and the last layers. In all layers of the generator, “Leaky Relu” is employed except the last layer. Similar to the generator, in all layers of the discriminator, “Leaky Relu” is employed except the last layer for activation. In the last layer, “tanh” function is used.

4. Results

The experiments have been carried out, and it has been found out that our proposed method outclasses the existing state-of-the-art methods. All experiments were conducted using MATLAB 2021 installed on a Core i7 (9th Generation) processor with 1.9 GHz and 16 GB of RAM. To validate our proposed method, the datasets mentioned in the Materials and Methods section are used. The quantitative and qualitative results show significant performance as compared to the existing methods. We have noticed the superiority performance of our proposed method in recovering highly degraded historical text images. We used the same criteria in our experimental results as in the existing methods.
The performance assessment metrics employed in this study are F-measures (FM), Peak Signal-to-Noise Ratio (PSNR), Word Error Rate (WER), and Distance-Reciprocal Distortion Metric (DRD). Figure 6, Figure 7 and Figure 8 show the qualitative results of our proposed and the existing deep learning methods published in 2021 and 2022.
Figure 6, Figure 7 and Figure 8 clearly depict the superiority of our proposed algorithm. However, the experimental results may not be perfect all the time, especially where the input image has a light text and dull background. During the experiments, it was observed that the resulting image from [2] was de-blurred well, but it was contaminated with background noise. However, its performance was satisfactory. The output achieved form [3] was free from noise but had blurring effects. Our proposed method initially de-blurred and de-noised the image with enhanced pixels resolution. Using the bi-cubic interpolation function, the input image was zoomed without any distortion.
Almost in all cases, the results achieved using S. Kang’s approach [2] are satisfactory and corporately higher than those of all other methods. This approach even still provides some better results as compared to our proposed method in terms of FM. This is because some images available from Noisy Office and Images datasets have dull backgrounds. S. K. Jemni’s [3] results are satisfactory; however, the high brightness causes some text distortion. Similarly, the proposed bi-cubic interpolation of LWT and SWT has better performance but with lower contrast. The application of GAN architecture enhances the visual quality due to complementary information transformation from input to output of images without distortion. GAN has the ability to generate equal distribution of frequencies received from input images. If we look at the final images in Figure 6, Figure 7 and Figure 8, the intensity variations are relatively optimal. The VID is much higher than for the rest of the methods due to the application of the GAN architecture, as it transforms the complementary information with any distortion of the text to the final fuse image.
Additionally, we have performed quantitative analysis to further evaluate our proposed method. Table 2, Table 3 and Table 4 illustrate the quantitative results of our proposed and the comparative methods. In few cases, when the images are very dull, our algorithm fails to enhance the input image and the images lose their details.
Table 2, Table 3 and Table 4 depict that our proposed algorithm achieved maximum results except in few cases. In the case of Noisy Office dataset, S. Kang [2] achieved higher FM than that of our proposed method due to few images with dull backgrounds. For the mixed dataset, the WER result of S. K. Jemni [3] is better than that of S. Kang [2] and shows competitive performance.
PSNR is most widely used as a quantitative metric which provides a pixel-wise evaluation and shows the effectiveness of document enhancement method in terms of visual quality. The primary objective of the first part of our proposed method is to enhance intensity variations, which increases the PSNR rate. Similarly, with the usage of the GAN architecture, the WER and DRD rates increase dramatically due to complementary information transformation from input to output of images without distortion.
To further evaluate our proposed method, images from datasets were analyzed and the results showed that our proposed method achieved satisfactory results. Table 5 and Table 6 illustrate the quantitative results of our proposed method and the existing comparative methods.
It was observed from Table 5 and Table 6 that our proposed algorithm achieved maximum results except in the case of WER for “Images” dataset due to some images having dull backgrounds, and because our method could not de-blur them well. For Tobacco dataset, our proposed method shows better performance.

5. Conclusions and Future Direction

This article proposed a new method towards the improvement of historical text images. Initially, the image is scaled up using a new bi-cubic interpolation of Lifting Wavelet Transform (LWT) and Stationary Wavelet Transform (SWT) to enhance pixels resolution and to reduce noise and blurring effects without degrading the quality of the image. Second, the GAN model is employed to extract the spatial and spectral features of a historical text image. The proposed method may be implemented in future for X-ray and CT images to assist clinical process. There is, however, still room for improvement.
Text readability is improved by sharpening the edges and lowering the contrast of the image’s background.
Overexposure and underexposure correction, super resolution, and OCR performance evaluation are a few key areas open for future work. Overexposure arises when excessive light is captured during the digitization process, particularly when using a smart phone. The camera flash adds an excessive amount of reflection or glare to the image, and as far as we know, no study is available to overcome this issue for historical text image enhancement. The same is true for underexposure when the lighting is poor. The underexposure problem received high attention recently. Moreover, fading is also a problem of attention and under-studied task. It occurs due to aging, washing out and illumination, etc.
Super-resolution of documents is a burning area in the field of image analysis and understanding. Low-resolution images create uncertainties for character recognition methods. Super-resolution is more challenging, especially when historical documents have noises and artifacts. In the future, effective deep learning methods, particularly transformer-based super-resolution methods, will be required to improve OCR performance.
Enhancing Optical Character Recognition, or OCR, is another crucial subject that needs attention in order to help automated text analysis of documents. Existing practices either overlook assessing their techniques in terms of OCR, or reveal OCR improved performance only on a small number of images, which is insufficient to demonstrate the practicality of their methodologies. Figures S1 and S2 show the examples of these tasks in the Supplementary Materials file. Figure S3 depicts an interpolation example.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/1424-8220/23/8/4003/s1, Figure S1: Image fading (image source [5]); Figure S2: Severe fading making text illegible; Figure S3: Interpolation example (when a = 0.5, we simply have the average of two parameters).

Author Contributions

Conceptualization, S.U.K., I.U. and S.U.; investigation, S.U.K. and I.U.; data creation, F.K.; formal analysis, I.U.; methodology, S.U.K. and I.U.; software, F.K.; validation, Y.L.; visualization, Y.L.; writing—original draft, S.U.K. and S.U. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported partially by the National Research Foundation of Korea (NRF), grant 2022R1G1A1003531, and Institute of Information and Communications Technology Planning and Evaluation (IITP), grant IITP-2022-2020-0-101741, and RS-2022-00155885 funded by the Korean government (MSIT).

Institutional Review Board Statement

I certified that I have followed the study protocol and the method of obtaining informed consent as approved by the IRB during the period of the research work.

Data Availability Statement

The datasets generated and/or analyzed during the current study are publicly available as mentioned in the Section 3. However, datasets may be provided by the corresponding author on reasonable request.

Conflicts of Interest

The authors have no conflicts of interest.

References

  1. Megyei, B.; Blomqvist, N.; Pettersson, E. The decode database: Collection of historical ciphers and keys. In Proceedings of the the 2nd International Conference on Historical Cryptology, Histo-Crypt 2019, Mons, Belgium, 23–26 June 2019; pp. 69–78. [Google Scholar]
  2. Kang, S.; Iwana, B.K.; Uchida, S. Complex image processing with less data document image binarization by integrating multiple pre train edu-net modules. Pattern Recognit. 2021, 109, 107577. [Google Scholar] [CrossRef]
  3. Jemni, S.K.; Souibgui, M.A.; Kessentini, Y.; Fornes, A. Enhance to read better: Amulti-task adversarial network for hand written document image enhancement. Pattern Recognit. 2022, 123, 108370. [Google Scholar] [CrossRef]
  4. Wang, B.; Chen, C.L.P. An effective background estimation method for shadows removal of document images. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 3611–3615. [Google Scholar]
  5. Hradis, M.; Kotera, J.; Zemcık, P.; Šroubek, F. Convolutional neural networks for direct text deblurring. In Proceedings of the BMVC, Swansea, UK, 7–10 September 2015. [Google Scholar]
  6. Gu, S.; Zuo, W.; Guo, S.; Chen, Y.; Chen, C.; Zhang, L. Learning dynamic guidance for depth image enhancement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3769–3778. [Google Scholar]
  7. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1780–1789. [Google Scholar]
  8. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  9. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  10. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  11. Sauvola, J.; Pietikainen, M. Adaptive document image binarization. Pattern Recognit. 2000, 33, 225–236. [Google Scholar] [CrossRef]
  12. Xiong, W.; Xu, J.; Xiong, Z.; Wang, J.; Liu, M. Degraded historical document image binarization using local features and support vector machine(svm). Optik 2018, 164, 218–223. [Google Scholar] [CrossRef]
  13. Hedjam, R.; Cheriet, M.; Kalacska, M. Constrained energy maximization and self-referencing method for invisible ink detection from multi spectral historical document images. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 3026–3031. [Google Scholar]
  14. Pratikakis, I.; Zagoris, K.; Barlas, G.; Gatos, B. Icdar 2017 competition on document image binarization (dibco2017). In Proceedings of the 2017 International Conference on Document Analysis and Recognition, Kyoto, Japan, 9–15 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1395–1403. [Google Scholar]
  15. Zhao, W.; Jun, W.; Lidong, H.; Zebin, S. Combination of contrast limited adaptive histogram equalization and discrete wavelet transform for image enhancement. IET Image Process. 2015, 9, 908–915. [Google Scholar]
  16. Afzal, M.Z.; Pastor-Pellicer, J.; Shafait, F.; Breuel, T.M.; Dengel, A.; Liwicki, M. Document image binarization using lstm: A sequence learning approach. In Proceedings of the 3rd International Workshop on Historical Document Imaging and Processing, Nancy, France, 22 August 2015; pp. 79–84. [Google Scholar]
  17. Mao, X.-J.; Shen, C.; Yang, Y.-B. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2810–2818. [Google Scholar]
  18. Akbari, Y.; Al-Maadeed, S.; Adam, K. Binarization of degraded document images using convolutional neural networks and wavelet based multi channel images. IEEE Access 2020, 8, 153517–153534. [Google Scholar] [CrossRef]
  19. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  20. Souibgui, M.A.; Kessentini, Y. De-gan: A conditional generative adversarial network for document enhancement. IEEE Trans.Pattern Anal. Mach. Intell. 2020, 44, 1180–1191. [Google Scholar] [CrossRef] [PubMed]
  21. Zhao, J.; Shi, C.; Jia, F.; Wang, Y.; Xiao, B. Document image binarization with cascaded generators of conditional generative adversarial networks. Pattern Recognit. 2019, 96, 106968. [Google Scholar] [CrossRef]
  22. Bhunia, A.K.; Bhunia, A.K.; Sain, A.; Roy, P.P. Improving document binarization via adversarial noise-texture augmentation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2721–2725. [Google Scholar]
  23. Tamrin, M.O.; Ech-Cherif, M.E.-A.; Cheriet, M. A two stage unsupervised deep learning framework for degradation removal in ancient documents. In Proceedings of the CPR International Workshops and Challenges, Virtual Event, 10–15 January 2021; pp. 292–303. [Google Scholar]
  24. Souibgui, M.A.; Kessentini, Y.; Fornes, A. A Conditional GAN based Approach for Distorted Camera Captured Documents Recovery. In Proceedings of the Mediterranean Conference on Pattern Recognition and Artificial Intelligence, Hammamet, Tunisia, 20–22 December 2020; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  25. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; VanGool, L.; Timofte, R. Swinir: Image Restoration Using Swin Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  26. Feng, H.; Wang, Y.; Zhou, W.; Deng, J.; Li, H. Doctr: Document image transformer for geometric unwarping and illumination correction. arXiv 2021, arXiv:2110.12942. [Google Scholar]
  27. Hejam, R.; Nafchi, H.Z.; Moghaddam, R.F.; Kalacska, M.; Cheriet, M. Icdar2015 contest on multi spectral text extraction (ms-tex2015). In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR) 2015, Tunis, Tunisia, 23–26 August 2015; pp. 1181–1185. [Google Scholar]
  28. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef] [PubMed]
  29. Deng, F.; Wu, Z.; Lu, Z.; Brown, M.S. Binarization shop: A user-assisted software suite for converting old documents to black-and-white. In Proceedings of the 10th Annual Joint Conference on Digital Libraries, Gold Coast, Australia, 21–25 June 2010; pp. 255–258. [Google Scholar]
  30. Dua, D.; Graff, C. UCL Machine Learning Repository. Available online: http://archive.ics.uci.edu/ml (accessed on 31 January 2023).
  31. Lai, W.-S.; Huang, J.-B.; Ahuja, N.; Yang, M.-H. Deep laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 624–632. [Google Scholar]
  32. Nayef, N.; Luqman, M.M.; Prum, S.; Eskenazi, S.; Chazalon, J.; Ogier, J.M. Smartdoc-qa: A dataset for quality assessment of smart phone captured document images-single and multiple distortions. In Proceedings of the 2015 13th International Conference on Document Analysis and Recognition (ICDAR), Tunis, Tunisia, 23–26 August 2015; pp. 1231–1235. [Google Scholar]
Figure 1. Visual representation of text image enhancement tasks: (a,c,e) input images, (b) de-noised image, (d) de-blurred image, (f) de-faded image.
Figure 1. Visual representation of text image enhancement tasks: (a,c,e) input images, (b) de-noised image, (d) de-blurred image, (f) de-faded image.
Sensors 23 04003 g001aSensors 23 04003 g001b
Figure 2. LWT coefficients.
Figure 2. LWT coefficients.
Sensors 23 04003 g002
Figure 3. Bi-cubic interpolation effects: (a) input image, (b) before interpolation, (c) after interpolation, and (d) no interpolation.
Figure 3. Bi-cubic interpolation effects: (a) input image, (b) before interpolation, (c) after interpolation, and (d) no interpolation.
Sensors 23 04003 g003
Figure 4. Intensity resolution.
Figure 4. Intensity resolution.
Sensors 23 04003 g004
Figure 5. GAN fusion process.
Figure 5. GAN fusion process.
Sensors 23 04003 g005
Figure 6. Qualitative results using Noisy Office dataset [30]: (a) input image, (b) S. Kang [2], (c) S. K. Jemni [3], (d) proposed LWT and SWT bi-cubic interpolation result, (e) proposed final image.
Figure 6. Qualitative results using Noisy Office dataset [30]: (a) input image, (b) S. Kang [2], (c) S. K. Jemni [3], (d) proposed LWT and SWT bi-cubic interpolation result, (e) proposed final image.
Sensors 23 04003 g006
Figure 7. Qualitative results S-MS [27]: (a) input image, (b) S. Kang [2], (c) S. K. Jemni [3], (d) proposed LWT and SWT bi-cubic interpolation result, (e) proposed final image.
Figure 7. Qualitative results S-MS [27]: (a) input image, (b) S. Kang [2], (c) S. K. Jemni [3], (d) proposed LWT and SWT bi-cubic interpolation result, (e) proposed final image.
Sensors 23 04003 g007
Figure 8. Qualitative results using Blurry document images dataset [5]: (a) input image, (b) S. Kang [2], (c) S. K. Jemni [3], (d) proposed LWT and SWT bi-cubic interpolation result, (e) proposed final image.
Figure 8. Qualitative results using Blurry document images dataset [5]: (a) input image, (b) S. Kang [2], (c) S. K. Jemni [3], (d) proposed LWT and SWT bi-cubic interpolation result, (e) proposed final image.
Sensors 23 04003 g008
Table 2. Comparative results on Noisy Office [30] dataset.
Table 2. Comparative results on Noisy Office [30] dataset.
MethodPSNRFMWERDRD
S. Kang [2] (CNN)14.4296.5485.451.23
S. K. Jemni [3] (CGAN)14.0994.2184.951.18
LWT &SWT13.9594.1883.121.01
Final Proposed15.0296.5186.171.31
Table 3. Comparative results on S-MS [27] dataset.
Table 3. Comparative results on S-MS [27] dataset.
MethodPSNRFMWERDRD
S. Kang [2] (CNN)17.3185.6174.212.02
S. K. Jemni [3] (CGAN)16.9985.2174.191.99
LWT and SWT16.9184.0273.911.88
Final Proposed18.0185.7173.322.44
Table 4. Comparative results on Blurry document images dataset [5].
Table 4. Comparative results on Blurry document images dataset [5].
MethodPSNRFMWERDRD
S. Kang [2] (CNN)19.4298.1281.282.11
S. K. Jemni [3] (CGAN)19.0297.8481.312.10
LWT and SWT18.2194.0280.212.01
Final Proposed19.4997.1382.122.17
Table 5. Comparative results on Tobacco 800 [31] dataset.
Table 5. Comparative results on Tobacco 800 [31] dataset.
MethodPSNRFMWERDRD
S. Kang [2] (CNN)21.0293.2183.212.18
S. K. Jemni [3] (CGAN)20.5193.0183.022.07
LWT and SWT19.7992.8182.881.98
Final Proposed21.7493.2483.842.57
Table 6. Comparative results on S-MS [27] dataset.
Table 6. Comparative results on S-MS [27] dataset.
MethodPSNRFMWERDRD
S. Kang [2] (CNN)19.4287.1279.511.87
S. K. Jemni [3] (CGAN)18.9186.7479.111.59
LWT and SWT18.0285.0778.931.02
Final Proposed19.7787.0579.591.91
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khan, S.U.; Ullah, I.; Khan, F.; Lee, Y.; Ullah, S. Historical Text Image Enhancement Using Image Scaling and Generative Adversarial Networks. Sensors 2023, 23, 4003. https://doi.org/10.3390/s23084003

AMA Style

Khan SU, Ullah I, Khan F, Lee Y, Ullah S. Historical Text Image Enhancement Using Image Scaling and Generative Adversarial Networks. Sensors. 2023; 23(8):4003. https://doi.org/10.3390/s23084003

Chicago/Turabian Style

Khan, Sajid Ullah, Imdad Ullah, Faheem Khan, Youngmoon Lee, and Shahid Ullah. 2023. "Historical Text Image Enhancement Using Image Scaling and Generative Adversarial Networks" Sensors 23, no. 8: 4003. https://doi.org/10.3390/s23084003

APA Style

Khan, S. U., Ullah, I., Khan, F., Lee, Y., & Ullah, S. (2023). Historical Text Image Enhancement Using Image Scaling and Generative Adversarial Networks. Sensors, 23(8), 4003. https://doi.org/10.3390/s23084003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop