Next Article in Journal
Double-Deck Metal Solenoids 3D Integrated in Silicon Wafer for Kinetic Energy Harvester
Next Article in Special Issue
Application of Machine Learning Algorithm on MEMS-Based Sensors for Determination of Helmet Wearing for Workplace Safety
Previous Article in Journal
Experimental Study of Bubble Formation from a Micro-Tube in Non-Newtonian Fluid
Previous Article in Special Issue
Optimization of Machine Learning in Various Situations Using ICT-Based TVOC Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Pre-Processing Method of Machine Learning for Edge Detection with Image Signal Processor Enhancement

Department of Electronic Engineering, Soongsil University, Seoul 06978, Korea
*
Author to whom correspondence should be addressed.
Micromachines 2021, 12(1), 73; https://doi.org/10.3390/mi12010073
Submission received: 20 November 2020 / Revised: 4 January 2021 / Accepted: 6 January 2021 / Published: 11 January 2021
(This article belongs to the Special Issue Artificial Intelligence on MEMS/Microdevices/Microsystems)

Abstract

:
Even though computer vision has been developing, edge detection is still one of the challenges in that field. It comes from the limitations of the complementary metal oxide semiconductor (CMOS) Image sensor used to collect the image data, and then image signal processor (ISP) is additionally required to understand the information received from each pixel and performs certain processing operations for edge detection. Even with/without ISP, as an output of hardware (camera, ISP), the original image is too raw to proceed edge detection image, because it can include extreme brightness and contrast, which is the key factor of image for edge detection. To reduce the onerousness, we propose a pre-processing method to obtain optimized brightness and contrast for improved edge detection. In the pre-processing, we extract meaningful features from image information and perform machine learning such as k-nearest neighbor (KNN), multilayer perceptron (MLP) and support vector machine (SVM) to obtain enhanced model by adjusting brightness and contrast. The comparison results of F1 score on edgy detection image of non-treated, pre-processed and pre-processed with machine learned are shown. The pre-processed with machine learned F1 result shows an average of 0.822, which is 2.7 times better results than the non-treated one. Eventually, the proposed pre-processing and machine learning method is proved as the essential method of pre-processing image from ISP in order to gain better edge detection image. In addition, if we go through the pre-processing method that we proposed, it is possible to more clearly and easily determine the object required when performing auto white balance (AWB) or auto exposure (AE) in the ISP. It helps to perform faster and more efficiently through the proactive ISP.

1. Introduction

After the invention of camera, the quality of image from machinery has been continuously improved and it is easy to access the image data. It is recognized as the main data itself and is used to extract additional information through complex data processing using artificial intelligence (AI) [1].
The CMOS Image Sensor is one of the microelectromechanical systems (MEMS) related image data expected to combine with different devices such as visible light communication (VLC), light detection and ranging (LiDAR), Optical ID tags, etc. With CMOS Image Sensor, image signal processor (ISP) treats attributes of image and produces an output image. However, traditional ISP system is not able to perfectly solve the problems such as detail loss, high noise and color rendering and not being appropriate for edge detection [2].
In image processing, edge detection is fundamentally important because they can quickly determine the boundaries of objects in an image [3]. Furthermore, edge detection is performed to simplify the image in order to minimize the amount of data to be processed. Moreover, computer vision technology has been developing, edge detection is considered essential for more challenging task such as object detection [4], object proposal [5] and image segmentation [6]. Therefore, it is necessary to develop suitable processor or method only for edge detection.
There are a variety of edge detection methods that are classified by different calculations and generates different error models. Prewitt, Canny, Sobel and Laplacian of Gaussian (LoG) are well-used operators of edge detection [7]. They are sensitive of noise so as to deal with the shortcomings, edge detection filters or soft computing approaches are introduced [8]. Computer vision technology can supplement deficiencies with machine learning. A lot of algorithms have been previously introduced to perform edge detection; gPb-UCM [9], CEDN [10], RCF [11], BDCN [12] and so on. As a part of these efforts, we propose pre-processing method to determine optimized contrast and brightness for edge detection with improved accuracy. We performed three types of machine learning models including MLP, SVM and KNN; all machine learning methods showed better F1 score than non-machine learned one, while pre-processing also scored better than non-treated one.

2. Materials and Methods

2.1. MEMS on Image Sensor and Processer

MEMS technology is used as a key sensor element required to the internet of things (IoT)-based smart home, innovative production system of smart factory, and plant safety vision system. In addition, intelligent sensors that are used in various fields, such as autonomous vehicles, robots, unmanned aerial vehicles and smartphones, where the smaller devices have more advantage. Accordingly, system-in-package (SiP) technology, which aggregates sensors and semiconductor circuits on one chip using MEMS technology, is used to develop intelligent sensors [13].
The CMOS image sensor can be mass-produced through the application of a logic large scale integration (LSI) manufacturing processor; it has the advantage of low manufacturing cost and low power consumption due to its small device size compared to a charge coupled device (CCD) image sensor having a high voltage analog circuit. With those factors driving the growth, the current image sensor market is expected to grow at an annual rate of about 8.6% from 2020 to 2025 to reach 28 billion in 2025 [14].
A typical smart image sensor system implements the image-capturing device and the image processor into separate functional units: an array of pixel sensors and an off-array processing unit. A standard pixel array architecture includes the photodiode, gate switch, source follower and readout transistor. The reset gate resets the photodiode at the beginning of each capture phase. The source follower isolates the photodiode from the data bus. The analog signals from the sensor array take raw pixel values for further image processing as shown in Figure 1 [15].
The ISP is a processing block that converts the raw digital image output from the AFE into an image that can be used for a given application. This processing is very complex and include a number of discrete processing blocks that can be arranged in a different order depending on the ISP [16]. ISP consists of Lens shading, Defective Pixel Correction (DPC), denoise, color filter array (CFA), auto white balance (AWB), auto exposure (AE), color correction matrix (CCM), Gamma correction, Chroma Resampler and so on as shown in Figure 2.
ISP has the information that can explain the image variation and computer vision can learn to compensate through that variation. Through this, computer vision can complement the function of ISP and if the function of ISP is used for low-level operations such as denosing, and computer vision is used for high-level operation; this can secure capacity and lower processing power [17].
Basic AE algorithms are a system which divides the image into five areas and place the main object on center, the background on top, and weights each area [18]. This approach is appropriate when the overall image is mid tone while proper exposure has not been performed with mixed contrast. To overcome this problem, study for judging the condition of the light source and auto selection of the method for targeted contrast. In detail, the algorithm terminates with normal contrast values between the background and object [19]. On the other hand, the algorithm continues when the state of light is backward or forwarded, compared to the average, and center values of the brightness levels of the entire image the illumination condition was divided into the brightness under sunshine and the darkness during night and according to each illumination condition experiment were performed with exposure, without exposure, and contrast stretch. As a result, when the image was with exposure, the edge detection was good and when the contrast stretch was performed, the edge detection value further increased [20].

2.2. Edge Detection

Edges are curves in which sudden changes in brightness or spatial derivatives of brightness occur [21]. Changes in brightness are where the surface direction changes discontinuously, where one object obscures another, where shadow lines appear or where the surface reflection properties are discontinuous. In each case, you need to find the discontinuity of the image brightness or its derivatives. Edge detection is a technique that produces pixels that are only on the border between areas and Laplacian of Gaussian (LoG), Prewitt, Sobel and Canny are widely used operators for edge detection.
LoG uses the 2D Gaussian function to reduce noise and operate the Laplacian function to find the edge by performing second order differentiation in the horizontal and vertical directions [22].
Prewitt is used for vertical and horizontal edge detection. Compared to the Sobel mask, the edge comes out less but the speed is much faster. The operator uses two masks that provide detailed information about the edge direction when considering the characteristics of the data on the other side of the mask center point. The two masks are convolutional, with the original image to obtain separate approximations of the derivatives for the horizontal and vertical edge changes [23].
Sobel detects the amount of change by comparing each direction values based on the center using mask. It extracts vertical, horizontal and diagonal edges and is resistant to noise and as the mask gets bigger, the edges become thicker and sharper. However, change in contrast occurs frequently and is not effective in complex images [24]. A method of combining Sobel operator with soft-threshold wavelet denoising has also been proposed [25].
Canny edge detection is smoothed using a Gaussian filter to remove noise. After that, the size and direction are found using the gradient the maximum value of the edge is determined through the non-maximum suppression process and the last edge is classified through hysteresis edge tracking [26]. In recent research, a median filter was used instead of Gaussian filtering to reduce the effect of noise and remove isolated points [27].
We used canny because it has the advantages of improving signal to noise ratio and better detection specially in noise condition compared to other operators mentioned above [28].

2.3. Dataset

Many works to make dataset for object and edge detection and image segmentation are known like BSDS500 [2] by Arbelaez et al., NYUD [29] by Silberman et al., Multicue [30] by Mely et al., BIPED [31] by Soria et al., etc. Although BSDS500 dataset, which is composed of 500 images for 200 training, 100 validation and 200 test images, is well-known in computer vision field, the ground truth (GT) of this dataset contains both the segmentation and boundary. BIPED, Barcelona Images for Perceptual Edge Detection, is a dataset with annotated thin edges. It is composed of 250 outdoor images of 1280 × 720 pixels and annotated by experts on the computer vision. This dataset is generated by the lack of edge detection datasets and available as a benchmark for evaluating edge detection. The dataset used in our study was performed using not only BIPED but also actual images taken using a camera of a Samsung Galaxy Note 9 driven by BSDS500 and CMOS image sensor. However, in the process of extracting the features of the histogram, BIPED was the most appropriate in the method mentioned above, so only BIPED was used. Using BIPED dataset, we carried out the image-transformation on brightness and contrast to augment the input image data as shown in Figure 3.
As BIPED has only 50 images for test data, we also need to increase the amount of them. Same task is applied to augment the test data.

2.4. Image Characteristics

Images are generated by the combination of an illumination source and reflection or absorption of energy from various elements of the scene being imaged [32]. We indicate images by two-dimensional functions of the form f (x, y). the value of f at spatial coordinates (x, y) is a scalar quantity that is characterized by two components: (x) is the amount of source illumination incident on the scene being viewed and (y) is the amount of illumination reflected by the objects in the scene. To interpret this information, we see an image histogram which is graphical representation of pixel intensity for the x-axis and number of pixels for y-axis. We analyze the histogram to extract the meaningful analysis for effective image processing.
We indicate images by two-dimensional functions of the form f (x, y). the value of f at spatial coordinates (x, y) is a scalar quantity that is characterized by two components: (x) is the amount of source illumination incident on the scene being viewed and (y) is the amount of illumination reflected by the objects in the scene. To interpret this information, we see an image histogram which is graphical representation of pixel intensity for the x-axis and number of pixels for y-axis. We analyze the histogram to extract the meaningful analysis for effective image processing.
We convert to RGB image data to grayscale and get the histogram. The x-axis has all available gray level from 0 to 255 and y-axis has the number of pixels that have a particular gray level value. We can get the information of brightness by observing the spatial distribution of the values. If the values are concentrated toward to the left, the image is darker. In contrast, if they are focused toward to the right, the image is lighter. Intensity levels is closely associated with the image contrast. Which is defined as the difference in intensity between the highest and lowest intensity levels in an image. When an appreciable number of pixels in an image have a high dynamic range, we typically expect the image to high contrast. Conversely, an image with low dynamic range especially the middle of the intensity scale indicates low contrast.

2.4.1. Pixel Feature Normalization

We did process for normalization, which is a process to view the meaningful data patterns or rules when data units do not match as shown in Figure 4. In most of applications, each image has a different range of pixel value, therefore normalization of the pixel is essential process of image processing. We need to transform features by scaling them to a given range between 0 and 1 by Min–Max-Scaler from sklearn.

2.4.2. Histogram Information

To look through the characteristics of the training image, we investigated the histogram of image each. As shown in Table 1 and Figure 5, we categorize them into some distribution types of brightness and contrast according to concentration of peak, pixel intensity etc. In order to obtain the appropriate threshold in actual image with various illumination, it is estimated as an important task. The number of peaks and intensities is considered in divided zone of histogram, as shown in Figure 5. The intensity of each zone is scored as Izone, while the peak of each zone is scored as Pzone, as follow,
I zone = Intensity   of   each   zone total   Intensity ,   P zone = peak   number   of   each   zone total   peak   number

2.5. Proposing Machine Learning Method

Supervised Learning is a method of machine learning for inferring a function from training data, and supervised learners accurately guess predicted values for a given data from training data [33]. The training data contain the characteristics of the input object in vector format, and the desired result is labeled for each vector. Supervised learning is divided into a predefined classification that predicts one of several possible class labels and a regression that extracts a continuous value from a given function [34].
In order to predict brightness and contrast for better edge detection, we label the collected data using histograms and apply supervised learning. Types of classification methods that produce not continuous results including Support Vector Machine (SVM), K-Nearest Neighbor (KNN), Multilayer Perceptron (MLP), etc.
First, SVM is known as one of the most powerful classification tools [35]. The general concept of SVM is to classify training samples by hyperplane in the space where the samples are mapped. Therefore, SVM only requires training samples close to the class boundary, so high-dimensional data can be processed using a small number of training samples [36].
KNN is one of the most basic and simple classification methods. When there is little or no prior knowledge of data distribution, the KNN method is one of the first choices for classification. It is a nonparametric classification system that bypasses the probability density problem [37].
MLP is the most common choice and corresponds to a functional model where the hidden unit is a sigmoid function [38]. These are feed-forward networks where the input flows only in one direction to the output, and each neuron in the layer connects to all neurons in the successive layer, but there is no feedback for the neurons in the previous layer. As far as hidden layers and the number of units are concerned, you should choose a topology that provides optimal performance [39]. We carry out machine learning as shown in Figure 6.

2.6. Performance Evaluation

Mean square error (MSE) is the average of the square of the error and it calculates the variance of the data values at the same location between two images. It measures the average difference of pixels in the entire original ground truth image with the edge detection image. Higher MSE means there is a greater difference between the original image and the processed image.
The peak signal-to-noise ratio represents the maximum signal-to-noise ratio and peak signal-to-noise ratio (PSNR) is an objective measurement method to evaluate the degree of change in an image. PSNR is generally expressed in decibel (dB) scale and higher PSNR indicates higher quality [40].
Furthermore, the Structural similarity index measure (SSIM) was not used in the measurement method. Because our method performs edge detection by adjusting the brightness and contrast of the original image. SSIM evaluates how similar the brightness, contrast, and structural differences are compared to the original image. So, it is not suitable for evaluating our image [41].
We perform edge detection of the image applying the canny algorithm to the pre-processed image. Next, we measure the MSE and PSNR between each resulting edge detection image and the ground truth image.

2.7. Model Evaluation Method

Describes the metrics used to evaluate the classification performance of a model or pattern in machine learning.
As a performance evaluation index, we selected the following items. First, Precision is the ratio of the actual object edge among those classified as object edges and the ratio of those classified as object edges among those classified as object edges by the model was designated as the Recall value.
Lastly, the F1 score is the harmonic average of Precision and Recall. When the data label is unbalanced, it is possible to accurately evaluate the performance of the model and the performance can be evaluated with a single number.

3. Results

In the experiment, the most of testing set is categorized in type F, H, E, B therefore we compare F1 score of these types to test the performance of our method comparing original image without pre-processing with pre-processing in BIPED dataset. Not only the scores but also the edge detection result of the image is shown in Figure 7. It can be seen from Figure 7c that only Canny algorithm without pre-processing is too sensitive to noise. Compared with only Canny edge detection, our method maintains meaningful edge by overcoming the noise.
As shown in Figure 8, the MSE was 0.168 and the PSNR was 55.991 dB. Standard deviation was 0.04 for MSE and 1.05 dB for PSNR and the difference in results between the images was small. Table 2 shows the results of MSE and PSNR according to the edge detection method. It was confirmed that adjusting the brightness and contrast increases the function of edge detection according to the image characteristics through the PSNR value. Furthermore, Table 2 lists the PSNR of the different methods. For the dataset used in each paper, “Rena”, “Baboon”, and “Pepper” were mainly used, and the number of pixel arrays that can affect the value of PSNR and the number of datasets used were entered.
As shown in Figure 9, our method obtained the best F-measure values in BIPED dataset. It is proved that our method improve performance on F-measure from 0.235 to 0.823. It clearly illustrates the importance of preprocessing task in various illumination image and the performance can be enhanced through learning.

4. Discussion

The pre-processing method uses the basic information like brightness and contrast of the image, so you can simply select the characteristics of the data. In addition, if image pre-processing is performed using this method, ISP can find ROI more easily and faster than before. Furthermore, the phenomenon caused by not finding an object, such as flickering of AF seen when the image is bright or the boundary line is ambiguous, will also be reduced. Although testing was conducted with many image samples and data sets, there was a limitation in deriving various information because it was limited to the histogram type used in the data set. Therefore, afterwards, it is necessary to diversify and extract characteristics such as brightness and contrast by securing its own data set. The processing speed of pre-processing takes several minutes to the final step of receiving the image of the dataset, analyzing the histogram, applying the feature, and detecting the edge. In the case of processing speed, the speed can be sufficiently reduced by upgrading the graphic processor unit (GPU). It is necessary to run it on a real board and get the result.
Furthermore, the method we propose is to facilitate edge detection by using the basic information of the image as a pre-process to complement the ISP function of the CMOS image sensor when the brightness is strong or the contrast is low, the image itself appears hazy like a watercolor technique, it is possible to find the object necessary for AWB or AE at the ISP more clearly and easily using pre-processing we suggest. In addition, power consumption or noise can be reduced. In the case of hardware complexity, the method we used is image pre-processing for edge detection. Since the image was processed by the edge detection algorithm after receiving the existing image in the form of a file, it is necessary to consider proceeding the overall process of edge detection using the value input to the CMOS image sensor using a board equipped with an actual processor.

5. Conclusions

In this research, we a propose pre-processing method on light control in image with various illumination environments for optimized edge detection with high accuracy. Our method can improve the quality of image by adjusting brightness and contrast, which results in effective edge detection than implementation without light control. So, we see that our edge result achieves the best F-measure. It would be interesting to study further on detection of textures and roughness in images with varying illumination. In addition, the pre-processing we propose can respond more quickly and effectively to the perception of an object by detecting the edge of the image. In particular, it is used for ISP pre-processing so that it can recognize the boundary lines required for operation faster and more accurately, which improves the speed of data processing compared to the existing ISP. It will be useful for autonomous cars, medical information, aviation and defense industries, etc.

Author Contributions

Conceptualization, K.P. and M.C.; methodology, K.P.; software, M.C.; validation, M.C., K.P. and J.H.C.; formal analysis, M.C.; investigation, J.H.C.; resources, K.P.; data curation, K.P.; writing—original draft preparation, M.C.; writing—review and editing, J.H.C.; visualization, M.C.; supervision, J.H.C.; project administration, J.H.C.; funding acquisition, J.H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Institute of Korea Health Industry Development Institute (KHIDI), grant number HI19C1032 and The APC was funded by Ministry of Health and Welfare (MOHW).

Acknowledgments

This work was supported by Institute of Korea Health Industry Development Institute (KHIDI) grant funded by the Korea government (Ministry of Health and Welfare, MOHW) (No. HI19C1032, Development of autonomous defense-type security technology and management system for strengthening cloud-based CDM security).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ignatov, A.; Van Gool, L.; Timofte, R. Replacing mobile camera isp with a single deep learning model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 16–18 June 2020; pp. 536–537. [Google Scholar]
  2. Rafati, M.; Arabfard, M.; Rafati-Rahimzadeh, M. Comparison of different edge detections and noise reduction on ultrasound images of carotid and brachial arteries using a speckle reducing anisotropic diffusion filter. Iran. Red Crescent Med. J. 2014, 16, e14658. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Öztürk, S.; Akdemir, B. Comparison of edge detection algorithms for texture analysis on glass production. Procedia Soc. Behav. Sci. 2015, 195, 2675–2682. [Google Scholar] [CrossRef] [Green Version]
  4. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1627–1645. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Zitnick, C.L.; Dollár, P. Edge Boxes: Locating Object Proposals from Edges. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 391–405. [Google Scholar]
  6. Pal, N.R.; Pal, S.K. A review on image segmentation techniques. Pattern Recognit. 1993, 26, 1277–1294. [Google Scholar] [CrossRef]
  7. Singh, S.; Singh, R. Comparison of various edge detection techniques. In Proceedings of the IEEE 2015 2nd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 11–13 March 2015; pp. 393–396. [Google Scholar]
  8. Li, H.; Liao, X.; Li, C.; Huang, H.; Li, C. Edge detection of noisy images based on cellular neural networks. Commun. Nonlinear Sci. Numer. Simul. 1999, 16, 3746–3759. [Google Scholar] [CrossRef]
  9. Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 898–916. [Google Scholar] [CrossRef] [Green Version]
  10. Yang, J.; Price, B.; Cohen, S.; Lee, H.; Yang, M.-H. Object contour detection with a fully convolutional encoder-decoder network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 193–202. [Google Scholar]
  11. Liu, Y.; Cheng, M.-M.; Hu, X.; Wang, K.; Bai, X. Richer convolutional features for edge detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3000–3009. [Google Scholar]
  12. He, J.; Zhang, S.; Yang, M.; Shan, Y.; Huang, T. Bi-directional cascade network for perceptual edge detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–21 June 2019; pp. 3828–3837. [Google Scholar]
  13. Lee, K.; Kim, M.S.; Shim, P.; Han, I.; Lee, J.; Chun, J.; Cha, S. Technology advancement of laminate substrates for mobile, iot, and automotive applications. In Proceedings of the IEEE 2017 China Semiconductor Technology International Conference (CSTIC), Shanghai, China, 12–13 March 2017; pp. 1–4. [Google Scholar]
  14. Image Sensor Market. Available online: https://www.marketsandmarkets.com/Market-Reports/Image-Sensor-Semiconductor-Market-601.html?gclid=CjwKCAjwwab7BRBAEiwAapqpTDKqQhaxRMb7MA6f9d_mQXs4cJrjtZxg_LVMkER9m4eSUkmS_f3J_BoCvRcQAvD_BwE (accessed on 8 January 2020).
  15. Zhang, M.; Bermak, A. Cmos image sensor with on-chip image compression: A review and performance analysis. J. Sens. 2010. [Google Scholar] [CrossRef]
  16. Yahiaoui, L.; Horgan, J.; Deegan, B.; Yogamani, S.; Hughes, C.; Denny, P. Overview and empirical analysis of isp parameter tuning for visual perception in autonomous driving. J. Imaging 2019, 5, 78. [Google Scholar] [CrossRef] [Green Version]
  17. Wu, C.-T.; Isikdogan, L.F.; Rao, S.; Nayak, B.; Gerasimow, T.; Sutic, A.; Ain-kedem, L.; Michael, G. Visionisp: Repurposing the image signal processor for computer vision applications. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 4624–4628. [Google Scholar]
  18. Lee, J.-S.; Jung, Y.-Y.; Kim, B.-S.; Ko, S.-J. An advanced video camera system with robust af, ae, and awb control. IEEE Trans. Consum. Electron. 2001, 47, 694–699. [Google Scholar]
  19. Liang, J.; Qin, Y.; Hong, Z. An auto-exposure algorithm for detecting high contrast lighting conditions. In Proceedings of the IEEE 2007 7th International Conference on ASIC, Guilin, China, 22–25 October 2007; pp. 725–728. [Google Scholar]
  20. Nguyen, T.T.; Dai Pham, X.; Kim, D.; Jeon, J.W. Automatic exposure compensation for line detection applications. In Proceedings of the 2008 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems, Seoul, Korea, 20–22 August 2008; pp. 68–73. [Google Scholar]
  21. Ahmad, M.B.; Choi, T.-S. Local threshold and boolean function based edge detection. IEEE Trans. Consum. Electron. 1999, 45, 674–679. [Google Scholar] [CrossRef]
  22. Marr, D.; Hildreth, E. Theory of edge detection. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1980, 207, 187–217. [Google Scholar]
  23. Prewitt, J.M. Object enhancement and extraction. Pict. Process. Psychopictorics 1970, 10, 15–19. [Google Scholar]
  24. Sobel, M.E. Asymptotic confidence intervals for indirect effects in structural equation models. Sociol. Methodol. 1982, 13, 290–312. [Google Scholar] [CrossRef]
  25. Gao, W.; Zhang, X.; Yang, L.; Liu, H. An improved sobel edge detection. In Proceedings of the IEEE 2010 3rd International Conference on Computer Science and Information Technology, Chengdu, China, 9–11 July 2010; pp. 67–71. [Google Scholar]
  26. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 6, 679–698. [Google Scholar] [CrossRef]
  27. Xuan, L.; Hong, Z. An improved canny edge detection algorithm. In Proceedings of the 2017 8th IEEE international conference on software engineering and service science (ICSESS), Beijing, China, 24–26 November 2017; pp. 275–278. [Google Scholar]
  28. Maini, R.; Aggarwal, H. Study and comparison of various image edge detection techniques. Int. J. Image Process. IJIP 2009, 3, 1–11. [Google Scholar]
  29. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
  30. Mély, D.A.; Kim, J.; McGill, M.; Guo, Y.; Serre, T. A systematic comparison between visual cues for boundary detection. Vis. Res. 2016, 120, 93–107. [Google Scholar] [CrossRef]
  31. Poma, X.S.; Riba, E.; Sappa, A. Dense extreme inception network: Towards a robust cnn model for edge detection. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1923–1932. [Google Scholar]
  32. Gonzalez, R.C.; Woods, R.E.; Eddins, S.L. Digital Image Processing, 4th ed.; Pearson Education: London, UK, 2018; pp. 57–63. [Google Scholar]
  33. Russell, S.; Norvig, P. Artificial Intelligence: A Modern Approach, 2nd ed.; Prentice Hall: Upper Saddle River, NJ, USA, 2002. [Google Scholar]
  34. Mohri, M.; Rostamizadeh, A.; Talwalkar, A. Foundations of Machine Learning; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  35. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  36. Cavallaro, G.; Riedel, M.; Richerzhagen, M.; Benediktsson, J.A.; Plaza, A. On understanding big data impacts in remotely sensed image classification using support vector machine methods. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2015, 8, 4634–4646. [Google Scholar] [CrossRef] [Green Version]
  37. Darasay, B. Nearest Neighbor Pattern Classification Techniques; IEEE Computer Society Press: Las Alamitos, LA, USA, 1991. [Google Scholar]
  38. Hsu, S.Y.; Masters, T.; Olson, M.; Tenorio, M.F.; Grogan, T. Comparative analysis of five neural network models. Remote Sens. Rev. 1992, 6, 319–329. [Google Scholar] [CrossRef]
  39. Del Frate, F.; Pacifici, F.; Schiavon, G.; Solimini, C. Use of neural networks for automatic classification from high-resolution images. IEEE Trans Geosci. Remote. Sens. 2007, 45, 800–809. [Google Scholar] [CrossRef]
  40. Poobathy, D.; Chezian, R. Manicka. Edge detection operators: Peak signal to noise ratio based comparison. IJ Image Graph. Signal Process. 2014, 10, 55–61. [Google Scholar]
  41. Pambrun, J.F.; Rita, N. Limitations of the SSIM quality metric in the context of diagnostic imaging. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015. [Google Scholar]
  42. Gaurav, K.; Ghanekar, U. Image steganography based on Canny edge detection, dilation operator and hybrid coding. J. Inf. Secur. Appl. 2018, 41, 41–51. [Google Scholar] [CrossRef]
  43. Ellinas, J.N. A robust wavelet-based watermarking algorithm using edge detection. World Acad. Sci. Eng. Technol. 2007, 291–296. [Google Scholar]
  44. Zhang, X.; Wang, S. Vulnerability of pixel-value differencing steganography to histogram analysis and modification for enhanced security. Pattern Recognit. Lett. 2004, 25, 331–339. [Google Scholar] [CrossRef]
  45. Al-Dmour, H.; Al-Ani, A. A steganography embedding method based on edge identification and XOR coding. Expert Syst. Appl. 2016, 46, 293–306. [Google Scholar] [CrossRef]
  46. Wu, D.-C.; Tsai, W.-H. A steganographic method for images by pixel-value differencing. Pattern Recognit. Lett. 2003, 24, 1613–1626. [Google Scholar] [CrossRef]
  47. Yang, C.H.; Weng, C.Y.; Wang, S.J.; Sun, H.M. Adaptive data hiding in edge areas of images with spatial LSB domain systems. IEEE Trans. Inf. Forensics Secur. 2008, 3, 488–497. [Google Scholar] [CrossRef]
  48. Bhardwaj, K.; Mann, P.S. Adaptive Neuro-Fuzzy Inference System (ANFIS) Based Edge Detection Technique. Int. J. Sci. Emerg. Technol. Latest Trends 2013, 8, 7–13. [Google Scholar]
  49. Singh, S.; Datar, A. Improved hash based approach for secure color image steganography using canny edge detection method. Int. J. Comput. Sci. Netw. Secur. (IJCSNS) 2015, 15, 92. [Google Scholar]
  50. Singla, K.; Kaur, S. A Hash Based Approach for secure image stegnograpgy using canny edge detection method. Int. J. Comput. Sci. Commun. 2012, 3, 156–157. [Google Scholar]
  51. Xu, J.; Wang, L.; Shi, Z. A switching weighted vector median filter based on edge detection. Signal Process. 2014, 98, 359–369. [Google Scholar] [CrossRef]
  52. Gambhir, D.; Rajpal, N. Fuzzy edge detector based blocking artifacts removal of DCT compressed images. In Proceedings of the IEEE 2013 International Conference on Circuits, Controls and Communications (CCUBE), Bengaluru, India, 27–28 December 2013; pp. 1–6. [Google Scholar]
  53. Kumar, S.; Saxena, R.; Singh, K. Fractional Fourier transform and fractional-order calculus-based image edge detection. Circuits Syst. Signal Process. 2017, 36, 1493–1513. [Google Scholar] [CrossRef]
  54. Ryu, Y.; Park, Y.; Kim, J.; Lee, S. Image edge detection using fuzzy c-means and three directions image shift method. IAENG Int. J. Comput. Sci. 2018, 45, 1–6. [Google Scholar]
  55. Topno, P.; Murmu, G. An Improved Edge Detection Method based on Median Filter. In Proceedings of the IEEE 2019 Devices for Integrated Circuit (DevIC), Kalyani, India, 23–24 March 2019; pp. 378–381. [Google Scholar]
  56. Anwar, S.; Raj, S. A neural network approach to edge detection using adaptive neuro-fuzzy inference system. In Proceedings of the IEEE 2014 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Noida, India, 24–27 September 2014; pp. 2432–2435. [Google Scholar]
  57. Srivastava, G.K.; Verma, R.; Mahrishi, R.; Rajesh, S. A novel wavelet edge detection algorithm for noisy images. In Proceedings of the IEEE 2009 International Conference on Ultra Modern Telecommunications & Workshops, St. Petersburg, Russia, 12–14 October 2009; pp. 1–8. [Google Scholar]
  58. Singh, H.; Kaur, T. Novel method for edge detection for gray scale images using VC++ environment. Int. J. Adv. Comput. Res. 2013, 3, 193. [Google Scholar]
  59. Shi, Q.; An, J.; Gagnon, K.K.; Cao, R.; Xie, H. Image Edge Detection Based on the Canny Edge and the Ant Colony Optimization Algorithm. In Proceedings of the IEEE 2019 12th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Suzhou, China, 19–21 October 2019; pp. 1–6. [Google Scholar]
  60. Ali, M.M.; Yannawar, P.; Gaikwad, A.T. Study of edge detection methods based on palmprint lines. In Proceedings of the IEEE 2016 International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT), Chennai, India, 3–5 March 2016; pp. 1344–1350. [Google Scholar]
Figure 1. Complementary metal oxide semiconductor (CMOS) Image Sensor: (a) CMOS Sensor for industrial vision (Canon Inc., Tokyo, Japan); (b) Circuit of one pixel; (c) Pixel array and Analog Frontend (AFE).
Figure 1. Complementary metal oxide semiconductor (CMOS) Image Sensor: (a) CMOS Sensor for industrial vision (Canon Inc., Tokyo, Japan); (b) Circuit of one pixel; (c) Pixel array and Analog Frontend (AFE).
Micromachines 12 00073 g001
Figure 2. Conventional Structure of CMOS Image Sensor.
Figure 2. Conventional Structure of CMOS Image Sensor.
Micromachines 12 00073 g002
Figure 3. We augment input image data by putting differential in brightness and contrast using BIPED dataset. For more augmentation, it can be adjusted each and simultaneously on original image: (a) original image; (b) controlled image (darker); (c) controlled image (brighter); (d) controlled image (low contrast); (e) controlled image (high contrast).
Figure 3. We augment input image data by putting differential in brightness and contrast using BIPED dataset. For more augmentation, it can be adjusted each and simultaneously on original image: (a) original image; (b) controlled image (darker); (c) controlled image (brighter); (d) controlled image (low contrast); (e) controlled image (high contrast).
Micromachines 12 00073 g003
Figure 4. Example of normalization: (a) Original image; (b) Histogram of original image; (c) Normalized histogram of original image.
Figure 4. Example of normalization: (a) Original image; (b) Histogram of original image; (c) Normalized histogram of original image.
Micromachines 12 00073 g004
Figure 5. Definition of Zone in the normalized histogram of brightness.
Figure 5. Definition of Zone in the normalized histogram of brightness.
Micromachines 12 00073 g005
Figure 6. Proposed Framework.
Figure 6. Proposed Framework.
Micromachines 12 00073 g006
Figure 7. We can see the edge result images without our method (pre-processing about brightness and contrast control) and them with: (a) original image; (b) Ground Truth; (c) Edge detection result with only Canny algorithm; (d) Edge detection result with our method.
Figure 7. We can see the edge result images without our method (pre-processing about brightness and contrast control) and them with: (a) original image; (b) Ground Truth; (c) Edge detection result with only Canny algorithm; (d) Edge detection result with our method.
Micromachines 12 00073 g007
Figure 8. Result of mean square error (MSE), peak signal-to-noise ratio (PSNR) per image.
Figure 8. Result of mean square error (MSE), peak signal-to-noise ratio (PSNR) per image.
Micromachines 12 00073 g008
Figure 9. Evaluation result of four images (F1 score): (a) Image without pre-processing; (b) Image with pre-processing before learning; (c) Image with pre-processing after learning.
Figure 9. Evaluation result of four images (F1 score): (a) Image without pre-processing; (b) Image with pre-processing before learning; (c) Image with pre-processing after learning.
Micromachines 12 00073 g009
Table 1. Type by the brightness and contrast.
Table 1. Type by the brightness and contrast.
IzoneII > 0.5
I II ~ V 0.5
IV > 0.5
I I ~ IV 0.5
Other
Pzone
PI > 0.5
P II ~ V 0.5
ABC
PV > 0.5
P I ~ IV 0.5
DEF
OtherGHI
Table 2. The Comparison with other edge detection methods.
Table 2. The Comparison with other edge detection methods.
MethodMSEPSNRResolutionMethodMSEPSNRResolution
Proposed0.16855.9911280 × 720X-OR [42]0.12257.240512 × 512
Robust Wavelet [43]-54512 × 512IPVD [44]0.27253.785512 × 512
V-bpp Edge-XOR [45]0.28853.532512 × 512PVD [46]0.45952.511512 × 512
AE_LSB [47]0.40952.011512 × 512ANFIS [48]0.45451.559-
Improved Hash [49]-47.559-Hash [50]-46.774512 × 512
weighted vector median filter [51]24.66034.210256 × 256Fuzzy Edge Detection [52]51.17031.040256 × 256
Fractional Fourier Transform [53]171.58025.786256 × 256Fuzzy C-means [54]6714.75922.708-
Median Filter [55]-18.850-Neural Network Approach [56]-16.340-
Novel Wavelet Edge Detection [57]-15.670-Novel Method [58]5911.66310.413-
Ant Colony Optimization Algorithm [59]8233.0918.975-D. Poobathy [40]19567.4425.216-
Mouad, M.H.Ali [60]20073.8525.127-----
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Park, K.; Chae, M.; Cho, J.H. Image Pre-Processing Method of Machine Learning for Edge Detection with Image Signal Processor Enhancement. Micromachines 2021, 12, 73. https://doi.org/10.3390/mi12010073

AMA Style

Park K, Chae M, Cho JH. Image Pre-Processing Method of Machine Learning for Edge Detection with Image Signal Processor Enhancement. Micromachines. 2021; 12(1):73. https://doi.org/10.3390/mi12010073

Chicago/Turabian Style

Park, Keumsun, Minah Chae, and Jae Hyuk Cho. 2021. "Image Pre-Processing Method of Machine Learning for Edge Detection with Image Signal Processor Enhancement" Micromachines 12, no. 1: 73. https://doi.org/10.3390/mi12010073

APA Style

Park, K., Chae, M., & Cho, J. H. (2021). Image Pre-Processing Method of Machine Learning for Edge Detection with Image Signal Processor Enhancement. Micromachines, 12(1), 73. https://doi.org/10.3390/mi12010073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop