1. Introduction
After the invention of camera, the quality of image from machinery has been continuously improved and it is easy to access the image data. It is recognized as the main data itself and is used to extract additional information through complex data processing using artificial intelligence (AI) [
1].
The CMOS Image Sensor is one of the microelectromechanical systems (MEMS) related image data expected to combine with different devices such as visible light communication (VLC), light detection and ranging (LiDAR), Optical ID tags, etc. With CMOS Image Sensor, image signal processor (ISP) treats attributes of image and produces an output image. However, traditional ISP system is not able to perfectly solve the problems such as detail loss, high noise and color rendering and not being appropriate for edge detection [
2].
In image processing, edge detection is fundamentally important because they can quickly determine the boundaries of objects in an image [
3]. Furthermore, edge detection is performed to simplify the image in order to minimize the amount of data to be processed. Moreover, computer vision technology has been developing, edge detection is considered essential for more challenging task such as object detection [
4], object proposal [
5] and image segmentation [
6]. Therefore, it is necessary to develop suitable processor or method only for edge detection.
There are a variety of edge detection methods that are classified by different calculations and generates different error models. Prewitt, Canny, Sobel and Laplacian of Gaussian (LoG) are well-used operators of edge detection [
7]. They are sensitive of noise so as to deal with the shortcomings, edge detection filters or soft computing approaches are introduced [
8]. Computer vision technology can supplement deficiencies with machine learning. A lot of algorithms have been previously introduced to perform edge detection; gPb-UCM [
9], CEDN [
10], RCF [
11], BDCN [
12] and so on. As a part of these efforts, we propose pre-processing method to determine optimized contrast and brightness for edge detection with improved accuracy. We performed three types of machine learning models including MLP, SVM and KNN; all machine learning methods showed better F1 score than non-machine learned one, while pre-processing also scored better than non-treated one.
2. Materials and Methods
2.1. MEMS on Image Sensor and Processer
MEMS technology is used as a key sensor element required to the internet of things (IoT)-based smart home, innovative production system of smart factory, and plant safety vision system. In addition, intelligent sensors that are used in various fields, such as autonomous vehicles, robots, unmanned aerial vehicles and smartphones, where the smaller devices have more advantage. Accordingly, system-in-package (SiP) technology, which aggregates sensors and semiconductor circuits on one chip using MEMS technology, is used to develop intelligent sensors [
13].
The CMOS image sensor can be mass-produced through the application of a logic large scale integration (LSI) manufacturing processor; it has the advantage of low manufacturing cost and low power consumption due to its small device size compared to a charge coupled device (CCD) image sensor having a high voltage analog circuit. With those factors driving the growth, the current image sensor market is expected to grow at an annual rate of about 8.6% from 2020 to 2025 to reach 28 billion in 2025 [
14].
A typical smart image sensor system implements the image-capturing device and the image processor into separate functional units: an array of pixel sensors and an off-array processing unit. A standard pixel array architecture includes the photodiode, gate switch, source follower and readout transistor. The reset gate resets the photodiode at the beginning of each capture phase. The source follower isolates the photodiode from the data bus. The analog signals from the sensor array take raw pixel values for further image processing as shown in
Figure 1 [
15].
The ISP is a processing block that converts the raw digital image output from the AFE into an image that can be used for a given application. This processing is very complex and include a number of discrete processing blocks that can be arranged in a different order depending on the ISP [
16]. ISP consists of Lens shading, Defective Pixel Correction (DPC), denoise, color filter array (CFA), auto white balance (AWB), auto exposure (AE), color correction matrix (CCM), Gamma correction, Chroma Resampler and so on as shown in
Figure 2.
ISP has the information that can explain the image variation and computer vision can learn to compensate through that variation. Through this, computer vision can complement the function of ISP and if the function of ISP is used for low-level operations such as denosing, and computer vision is used for high-level operation; this can secure capacity and lower processing power [
17].
Basic AE algorithms are a system which divides the image into five areas and place the main object on center, the background on top, and weights each area [
18]. This approach is appropriate when the overall image is mid tone while proper exposure has not been performed with mixed contrast. To overcome this problem, study for judging the condition of the light source and auto selection of the method for targeted contrast. In detail, the algorithm terminates with normal contrast values between the background and object [
19]. On the other hand, the algorithm continues when the state of light is backward or forwarded, compared to the average, and center values of the brightness levels of the entire image the illumination condition was divided into the brightness under sunshine and the darkness during night and according to each illumination condition experiment were performed with exposure, without exposure, and contrast stretch. As a result, when the image was with exposure, the edge detection was good and when the contrast stretch was performed, the edge detection value further increased [
20].
2.2. Edge Detection
Edges are curves in which sudden changes in brightness or spatial derivatives of brightness occur [
21]. Changes in brightness are where the surface direction changes discontinuously, where one object obscures another, where shadow lines appear or where the surface reflection properties are discontinuous. In each case, you need to find the discontinuity of the image brightness or its derivatives. Edge detection is a technique that produces pixels that are only on the border between areas and Laplacian of Gaussian (LoG), Prewitt, Sobel and Canny are widely used operators for edge detection.
LoG uses the 2D Gaussian function to reduce noise and operate the Laplacian function to find the edge by performing second order differentiation in the horizontal and vertical directions [
22].
Prewitt is used for vertical and horizontal edge detection. Compared to the Sobel mask, the edge comes out less but the speed is much faster. The operator uses two masks that provide detailed information about the edge direction when considering the characteristics of the data on the other side of the mask center point. The two masks are convolutional, with the original image to obtain separate approximations of the derivatives for the horizontal and vertical edge changes [
23].
Sobel detects the amount of change by comparing each direction values based on the center using mask. It extracts vertical, horizontal and diagonal edges and is resistant to noise and as the mask gets bigger, the edges become thicker and sharper. However, change in contrast occurs frequently and is not effective in complex images [
24]. A method of combining Sobel operator with soft-threshold wavelet denoising has also been proposed [
25].
Canny edge detection is smoothed using a Gaussian filter to remove noise. After that, the size and direction are found using the gradient the maximum value of the edge is determined through the non-maximum suppression process and the last edge is classified through hysteresis edge tracking [
26]. In recent research, a median filter was used instead of Gaussian filtering to reduce the effect of noise and remove isolated points [
27].
We used canny because it has the advantages of improving signal to noise ratio and better detection specially in noise condition compared to other operators mentioned above [
28].
2.3. Dataset
Many works to make dataset for object and edge detection and image segmentation are known like BSDS500 [
2] by Arbelaez et al., NYUD [
29] by Silberman et al., Multicue [
30] by Mely et al., BIPED [
31] by Soria et al., etc. Although BSDS500 dataset, which is composed of 500 images for 200 training, 100 validation and 200 test images, is well-known in computer vision field, the ground truth (GT) of this dataset contains both the segmentation and boundary. BIPED, Barcelona Images for Perceptual Edge Detection, is a dataset with annotated thin edges. It is composed of 250 outdoor images of 1280 × 720 pixels and annotated by experts on the computer vision. This dataset is generated by the lack of edge detection datasets and available as a benchmark for evaluating edge detection. The dataset used in our study was performed using not only BIPED but also actual images taken using a camera of a Samsung Galaxy Note 9 driven by BSDS500 and CMOS image sensor. However, in the process of extracting the features of the histogram, BIPED was the most appropriate in the method mentioned above, so only BIPED was used. Using BIPED dataset, we carried out the image-transformation on brightness and contrast to augment the input image data as shown in
Figure 3.
As BIPED has only 50 images for test data, we also need to increase the amount of them. Same task is applied to augment the test data.
2.4. Image Characteristics
Images are generated by the combination of an illumination source and reflection or absorption of energy from various elements of the scene being imaged [
32]. We indicate images by two-dimensional functions of the form f (
x,
y). the value of f at spatial coordinates (
x,
y) is a scalar quantity that is characterized by two components: (
x) is the amount of source illumination incident on the scene being viewed and (
y) is the amount of illumination reflected by the objects in the scene. To interpret this information, we see an image histogram which is graphical representation of pixel intensity for the
x-axis and number of pixels for
y-axis. We analyze the histogram to extract the meaningful analysis for effective image processing.
We indicate images by two-dimensional functions of the form f (x, y). the value of f at spatial coordinates (x, y) is a scalar quantity that is characterized by two components: (x) is the amount of source illumination incident on the scene being viewed and (y) is the amount of illumination reflected by the objects in the scene. To interpret this information, we see an image histogram which is graphical representation of pixel intensity for the x-axis and number of pixels for y-axis. We analyze the histogram to extract the meaningful analysis for effective image processing.
We convert to RGB image data to grayscale and get the histogram. The x-axis has all available gray level from 0 to 255 and y-axis has the number of pixels that have a particular gray level value. We can get the information of brightness by observing the spatial distribution of the values. If the values are concentrated toward to the left, the image is darker. In contrast, if they are focused toward to the right, the image is lighter. Intensity levels is closely associated with the image contrast. Which is defined as the difference in intensity between the highest and lowest intensity levels in an image. When an appreciable number of pixels in an image have a high dynamic range, we typically expect the image to high contrast. Conversely, an image with low dynamic range especially the middle of the intensity scale indicates low contrast.
2.4.1. Pixel Feature Normalization
We did process for normalization, which is a process to view the meaningful data patterns or rules when data units do not match as shown in
Figure 4. In most of applications, each image has a different range of pixel value, therefore normalization of the pixel is essential process of image processing. We need to transform features by scaling them to a given range between 0 and 1 by Min–Max-Scaler from sklearn.
2.4.2. Histogram Information
To look through the characteristics of the training image, we investigated the histogram of image each. As shown in
Table 1 and
Figure 5, we categorize them into some distribution types of brightness and contrast according to concentration of peak, pixel intensity etc. In order to obtain the appropriate threshold in actual image with various illumination, it is estimated as an important task. The number of peaks and intensities is considered in divided zone of histogram, as shown in
Figure 5. The intensity of each zone is scored as
Izone, while the peak of each zone is scored as
Pzone, as follow,
2.5. Proposing Machine Learning Method
Supervised Learning is a method of machine learning for inferring a function from training data, and supervised learners accurately guess predicted values for a given data from training data [
33]. The training data contain the characteristics of the input object in vector format, and the desired result is labeled for each vector. Supervised learning is divided into a predefined classification that predicts one of several possible class labels and a regression that extracts a continuous value from a given function [
34].
In order to predict brightness and contrast for better edge detection, we label the collected data using histograms and apply supervised learning. Types of classification methods that produce not continuous results including Support Vector Machine (SVM), K-Nearest Neighbor (KNN), Multilayer Perceptron (MLP), etc.
First, SVM is known as one of the most powerful classification tools [
35]. The general concept of SVM is to classify training samples by hyperplane in the space where the samples are mapped. Therefore, SVM only requires training samples close to the class boundary, so high-dimensional data can be processed using a small number of training samples [
36].
KNN is one of the most basic and simple classification methods. When there is little or no prior knowledge of data distribution, the KNN method is one of the first choices for classification. It is a nonparametric classification system that bypasses the probability density problem [
37].
MLP is the most common choice and corresponds to a functional model where the hidden unit is a sigmoid function [
38]. These are feed-forward networks where the input flows only in one direction to the output, and each neuron in the layer connects to all neurons in the successive layer, but there is no feedback for the neurons in the previous layer. As far as hidden layers and the number of units are concerned, you should choose a topology that provides optimal performance [
39]. We carry out machine learning as shown in
Figure 6.
2.6. Performance Evaluation
Mean square error (MSE) is the average of the square of the error and it calculates the variance of the data values at the same location between two images. It measures the average difference of pixels in the entire original ground truth image with the edge detection image. Higher MSE means there is a greater difference between the original image and the processed image.
The peak signal-to-noise ratio represents the maximum signal-to-noise ratio and peak signal-to-noise ratio (PSNR) is an objective measurement method to evaluate the degree of change in an image. PSNR is generally expressed in decibel (dB) scale and higher PSNR indicates higher quality [
40].
Furthermore, the Structural similarity index measure (SSIM) was not used in the measurement method. Because our method performs edge detection by adjusting the brightness and contrast of the original image. SSIM evaluates how similar the brightness, contrast, and structural differences are compared to the original image. So, it is not suitable for evaluating our image [
41].
We perform edge detection of the image applying the canny algorithm to the pre-processed image. Next, we measure the MSE and PSNR between each resulting edge detection image and the ground truth image.
2.7. Model Evaluation Method
Describes the metrics used to evaluate the classification performance of a model or pattern in machine learning.
As a performance evaluation index, we selected the following items. First, Precision is the ratio of the actual object edge among those classified as object edges and the ratio of those classified as object edges among those classified as object edges by the model was designated as the Recall value.
Lastly, the F1 score is the harmonic average of Precision and Recall. When the data label is unbalanced, it is possible to accurately evaluate the performance of the model and the performance can be evaluated with a single number.
3. Results
In the experiment, the most of testing set is categorized in type F, H, E, B therefore we compare F1 score of these types to test the performance of our method comparing original image without pre-processing with pre-processing in BIPED dataset. Not only the scores but also the edge detection result of the image is shown in
Figure 7. It can be seen from
Figure 7c that only Canny algorithm without pre-processing is too sensitive to noise. Compared with only Canny edge detection, our method maintains meaningful edge by overcoming the noise.
As shown in
Figure 8, the MSE was 0.168 and the PSNR was 55.991 dB. Standard deviation was 0.04 for MSE and 1.05 dB for PSNR and the difference in results between the images was small.
Table 2 shows the results of MSE and PSNR according to the edge detection method. It was confirmed that adjusting the brightness and contrast increases the function of edge detection according to the image characteristics through the PSNR value. Furthermore,
Table 2 lists the PSNR of the different methods. For the dataset used in each paper, “Rena”, “Baboon”, and “Pepper” were mainly used, and the number of pixel arrays that can affect the value of PSNR and the number of datasets used were entered.
As shown in
Figure 9, our method obtained the best F-measure values in BIPED dataset. It is proved that our method improve performance on F-measure from 0.235 to 0.823. It clearly illustrates the importance of preprocessing task in various illumination image and the performance can be enhanced through learning.
4. Discussion
The pre-processing method uses the basic information like brightness and contrast of the image, so you can simply select the characteristics of the data. In addition, if image pre-processing is performed using this method, ISP can find ROI more easily and faster than before. Furthermore, the phenomenon caused by not finding an object, such as flickering of AF seen when the image is bright or the boundary line is ambiguous, will also be reduced. Although testing was conducted with many image samples and data sets, there was a limitation in deriving various information because it was limited to the histogram type used in the data set. Therefore, afterwards, it is necessary to diversify and extract characteristics such as brightness and contrast by securing its own data set. The processing speed of pre-processing takes several minutes to the final step of receiving the image of the dataset, analyzing the histogram, applying the feature, and detecting the edge. In the case of processing speed, the speed can be sufficiently reduced by upgrading the graphic processor unit (GPU). It is necessary to run it on a real board and get the result.
Furthermore, the method we propose is to facilitate edge detection by using the basic information of the image as a pre-process to complement the ISP function of the CMOS image sensor when the brightness is strong or the contrast is low, the image itself appears hazy like a watercolor technique, it is possible to find the object necessary for AWB or AE at the ISP more clearly and easily using pre-processing we suggest. In addition, power consumption or noise can be reduced. In the case of hardware complexity, the method we used is image pre-processing for edge detection. Since the image was processed by the edge detection algorithm after receiving the existing image in the form of a file, it is necessary to consider proceeding the overall process of edge detection using the value input to the CMOS image sensor using a board equipped with an actual processor.
5. Conclusions
In this research, we a propose pre-processing method on light control in image with various illumination environments for optimized edge detection with high accuracy. Our method can improve the quality of image by adjusting brightness and contrast, which results in effective edge detection than implementation without light control. So, we see that our edge result achieves the best F-measure. It would be interesting to study further on detection of textures and roughness in images with varying illumination. In addition, the pre-processing we propose can respond more quickly and effectively to the perception of an object by detecting the edge of the image. In particular, it is used for ISP pre-processing so that it can recognize the boundary lines required for operation faster and more accurately, which improves the speed of data processing compared to the existing ISP. It will be useful for autonomous cars, medical information, aviation and defense industries, etc.
Author Contributions
Conceptualization, K.P. and M.C.; methodology, K.P.; software, M.C.; validation, M.C., K.P. and J.H.C.; formal analysis, M.C.; investigation, J.H.C.; resources, K.P.; data curation, K.P.; writing—original draft preparation, M.C.; writing—review and editing, J.H.C.; visualization, M.C.; supervision, J.H.C.; project administration, J.H.C.; funding acquisition, J.H.C. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by Institute of Korea Health Industry Development Institute (KHIDI), grant number HI19C1032 and The APC was funded by Ministry of Health and Welfare (MOHW).
Acknowledgments
This work was supported by Institute of Korea Health Industry Development Institute (KHIDI) grant funded by the Korea government (Ministry of Health and Welfare, MOHW) (No. HI19C1032, Development of autonomous defense-type security technology and management system for strengthening cloud-based CDM security).
Conflicts of Interest
The authors declare no conflict of interest.
References
- Ignatov, A.; Van Gool, L.; Timofte, R. Replacing mobile camera isp with a single deep learning model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 16–18 June 2020; pp. 536–537. [Google Scholar]
- Rafati, M.; Arabfard, M.; Rafati-Rahimzadeh, M. Comparison of different edge detections and noise reduction on ultrasound images of carotid and brachial arteries using a speckle reducing anisotropic diffusion filter. Iran. Red Crescent Med. J. 2014, 16, e14658. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Öztürk, S.; Akdemir, B. Comparison of edge detection algorithms for texture analysis on glass production. Procedia Soc. Behav. Sci. 2015, 195, 2675–2682. [Google Scholar] [CrossRef] [Green Version]
- Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 1627–1645. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Zitnick, C.L.; Dollár, P. Edge Boxes: Locating Object Proposals from Edges. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2014; pp. 391–405. [Google Scholar]
- Pal, N.R.; Pal, S.K. A review on image segmentation techniques. Pattern Recognit. 1993, 26, 1277–1294. [Google Scholar] [CrossRef]
- Singh, S.; Singh, R. Comparison of various edge detection techniques. In Proceedings of the IEEE 2015 2nd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 11–13 March 2015; pp. 393–396. [Google Scholar]
- Li, H.; Liao, X.; Li, C.; Huang, H.; Li, C. Edge detection of noisy images based on cellular neural networks. Commun. Nonlinear Sci. Numer. Simul. 1999, 16, 3746–3759. [Google Scholar] [CrossRef]
- Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 898–916. [Google Scholar] [CrossRef] [Green Version]
- Yang, J.; Price, B.; Cohen, S.; Lee, H.; Yang, M.-H. Object contour detection with a fully convolutional encoder-decoder network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 193–202. [Google Scholar]
- Liu, Y.; Cheng, M.-M.; Hu, X.; Wang, K.; Bai, X. Richer convolutional features for edge detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3000–3009. [Google Scholar]
- He, J.; Zhang, S.; Yang, M.; Shan, Y.; Huang, T. Bi-directional cascade network for perceptual edge detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–21 June 2019; pp. 3828–3837. [Google Scholar]
- Lee, K.; Kim, M.S.; Shim, P.; Han, I.; Lee, J.; Chun, J.; Cha, S. Technology advancement of laminate substrates for mobile, iot, and automotive applications. In Proceedings of the IEEE 2017 China Semiconductor Technology International Conference (CSTIC), Shanghai, China, 12–13 March 2017; pp. 1–4. [Google Scholar]
- Image Sensor Market. Available online: https://www.marketsandmarkets.com/Market-Reports/Image-Sensor-Semiconductor-Market-601.html?gclid=CjwKCAjwwab7BRBAEiwAapqpTDKqQhaxRMb7MA6f9d_mQXs4cJrjtZxg_LVMkER9m4eSUkmS_f3J_BoCvRcQAvD_BwE (accessed on 8 January 2020).
- Zhang, M.; Bermak, A. Cmos image sensor with on-chip image compression: A review and performance analysis. J. Sens. 2010. [Google Scholar] [CrossRef]
- Yahiaoui, L.; Horgan, J.; Deegan, B.; Yogamani, S.; Hughes, C.; Denny, P. Overview and empirical analysis of isp parameter tuning for visual perception in autonomous driving. J. Imaging 2019, 5, 78. [Google Scholar] [CrossRef] [Green Version]
- Wu, C.-T.; Isikdogan, L.F.; Rao, S.; Nayak, B.; Gerasimow, T.; Sutic, A.; Ain-kedem, L.; Michael, G. Visionisp: Repurposing the image signal processor for computer vision applications. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 4624–4628. [Google Scholar]
- Lee, J.-S.; Jung, Y.-Y.; Kim, B.-S.; Ko, S.-J. An advanced video camera system with robust af, ae, and awb control. IEEE Trans. Consum. Electron. 2001, 47, 694–699. [Google Scholar]
- Liang, J.; Qin, Y.; Hong, Z. An auto-exposure algorithm for detecting high contrast lighting conditions. In Proceedings of the IEEE 2007 7th International Conference on ASIC, Guilin, China, 22–25 October 2007; pp. 725–728. [Google Scholar]
- Nguyen, T.T.; Dai Pham, X.; Kim, D.; Jeon, J.W. Automatic exposure compensation for line detection applications. In Proceedings of the 2008 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems, Seoul, Korea, 20–22 August 2008; pp. 68–73. [Google Scholar]
- Ahmad, M.B.; Choi, T.-S. Local threshold and boolean function based edge detection. IEEE Trans. Consum. Electron. 1999, 45, 674–679. [Google Scholar] [CrossRef]
- Marr, D.; Hildreth, E. Theory of edge detection. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1980, 207, 187–217. [Google Scholar]
- Prewitt, J.M. Object enhancement and extraction. Pict. Process. Psychopictorics 1970, 10, 15–19. [Google Scholar]
- Sobel, M.E. Asymptotic confidence intervals for indirect effects in structural equation models. Sociol. Methodol. 1982, 13, 290–312. [Google Scholar] [CrossRef]
- Gao, W.; Zhang, X.; Yang, L.; Liu, H. An improved sobel edge detection. In Proceedings of the IEEE 2010 3rd International Conference on Computer Science and Information Technology, Chengdu, China, 9–11 July 2010; pp. 67–71. [Google Scholar]
- Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 6, 679–698. [Google Scholar] [CrossRef]
- Xuan, L.; Hong, Z. An improved canny edge detection algorithm. In Proceedings of the 2017 8th IEEE international conference on software engineering and service science (ICSESS), Beijing, China, 24–26 November 2017; pp. 275–278. [Google Scholar]
- Maini, R.; Aggarwal, H. Study and comparison of various image edge detection techniques. Int. J. Image Process. IJIP 2009, 3, 1–11. [Google Scholar]
- Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
- Mély, D.A.; Kim, J.; McGill, M.; Guo, Y.; Serre, T. A systematic comparison between visual cues for boundary detection. Vis. Res. 2016, 120, 93–107. [Google Scholar] [CrossRef]
- Poma, X.S.; Riba, E.; Sappa, A. Dense extreme inception network: Towards a robust cnn model for edge detection. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1923–1932. [Google Scholar]
- Gonzalez, R.C.; Woods, R.E.; Eddins, S.L. Digital Image Processing, 4th ed.; Pearson Education: London, UK, 2018; pp. 57–63. [Google Scholar]
- Russell, S.; Norvig, P. Artificial Intelligence: A Modern Approach, 2nd ed.; Prentice Hall: Upper Saddle River, NJ, USA, 2002. [Google Scholar]
- Mohri, M.; Rostamizadeh, A.; Talwalkar, A. Foundations of Machine Learning; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Cavallaro, G.; Riedel, M.; Richerzhagen, M.; Benediktsson, J.A.; Plaza, A. On understanding big data impacts in remotely sensed image classification using support vector machine methods. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2015, 8, 4634–4646. [Google Scholar] [CrossRef] [Green Version]
- Darasay, B. Nearest Neighbor Pattern Classification Techniques; IEEE Computer Society Press: Las Alamitos, LA, USA, 1991. [Google Scholar]
- Hsu, S.Y.; Masters, T.; Olson, M.; Tenorio, M.F.; Grogan, T. Comparative analysis of five neural network models. Remote Sens. Rev. 1992, 6, 319–329. [Google Scholar] [CrossRef]
- Del Frate, F.; Pacifici, F.; Schiavon, G.; Solimini, C. Use of neural networks for automatic classification from high-resolution images. IEEE Trans Geosci. Remote. Sens. 2007, 45, 800–809. [Google Scholar] [CrossRef]
- Poobathy, D.; Chezian, R. Manicka. Edge detection operators: Peak signal to noise ratio based comparison. IJ Image Graph. Signal Process. 2014, 10, 55–61. [Google Scholar]
- Pambrun, J.F.; Rita, N. Limitations of the SSIM quality metric in the context of diagnostic imaging. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015. [Google Scholar]
- Gaurav, K.; Ghanekar, U. Image steganography based on Canny edge detection, dilation operator and hybrid coding. J. Inf. Secur. Appl. 2018, 41, 41–51. [Google Scholar] [CrossRef]
- Ellinas, J.N. A robust wavelet-based watermarking algorithm using edge detection. World Acad. Sci. Eng. Technol. 2007, 291–296. [Google Scholar]
- Zhang, X.; Wang, S. Vulnerability of pixel-value differencing steganography to histogram analysis and modification for enhanced security. Pattern Recognit. Lett. 2004, 25, 331–339. [Google Scholar] [CrossRef]
- Al-Dmour, H.; Al-Ani, A. A steganography embedding method based on edge identification and XOR coding. Expert Syst. Appl. 2016, 46, 293–306. [Google Scholar] [CrossRef]
- Wu, D.-C.; Tsai, W.-H. A steganographic method for images by pixel-value differencing. Pattern Recognit. Lett. 2003, 24, 1613–1626. [Google Scholar] [CrossRef]
- Yang, C.H.; Weng, C.Y.; Wang, S.J.; Sun, H.M. Adaptive data hiding in edge areas of images with spatial LSB domain systems. IEEE Trans. Inf. Forensics Secur. 2008, 3, 488–497. [Google Scholar] [CrossRef]
- Bhardwaj, K.; Mann, P.S. Adaptive Neuro-Fuzzy Inference System (ANFIS) Based Edge Detection Technique. Int. J. Sci. Emerg. Technol. Latest Trends 2013, 8, 7–13. [Google Scholar]
- Singh, S.; Datar, A. Improved hash based approach for secure color image steganography using canny edge detection method. Int. J. Comput. Sci. Netw. Secur. (IJCSNS) 2015, 15, 92. [Google Scholar]
- Singla, K.; Kaur, S. A Hash Based Approach for secure image stegnograpgy using canny edge detection method. Int. J. Comput. Sci. Commun. 2012, 3, 156–157. [Google Scholar]
- Xu, J.; Wang, L.; Shi, Z. A switching weighted vector median filter based on edge detection. Signal Process. 2014, 98, 359–369. [Google Scholar] [CrossRef]
- Gambhir, D.; Rajpal, N. Fuzzy edge detector based blocking artifacts removal of DCT compressed images. In Proceedings of the IEEE 2013 International Conference on Circuits, Controls and Communications (CCUBE), Bengaluru, India, 27–28 December 2013; pp. 1–6. [Google Scholar]
- Kumar, S.; Saxena, R.; Singh, K. Fractional Fourier transform and fractional-order calculus-based image edge detection. Circuits Syst. Signal Process. 2017, 36, 1493–1513. [Google Scholar] [CrossRef]
- Ryu, Y.; Park, Y.; Kim, J.; Lee, S. Image edge detection using fuzzy c-means and three directions image shift method. IAENG Int. J. Comput. Sci. 2018, 45, 1–6. [Google Scholar]
- Topno, P.; Murmu, G. An Improved Edge Detection Method based on Median Filter. In Proceedings of the IEEE 2019 Devices for Integrated Circuit (DevIC), Kalyani, India, 23–24 March 2019; pp. 378–381. [Google Scholar]
- Anwar, S.; Raj, S. A neural network approach to edge detection using adaptive neuro-fuzzy inference system. In Proceedings of the IEEE 2014 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Noida, India, 24–27 September 2014; pp. 2432–2435. [Google Scholar]
- Srivastava, G.K.; Verma, R.; Mahrishi, R.; Rajesh, S. A novel wavelet edge detection algorithm for noisy images. In Proceedings of the IEEE 2009 International Conference on Ultra Modern Telecommunications & Workshops, St. Petersburg, Russia, 12–14 October 2009; pp. 1–8. [Google Scholar]
- Singh, H.; Kaur, T. Novel method for edge detection for gray scale images using VC++ environment. Int. J. Adv. Comput. Res. 2013, 3, 193. [Google Scholar]
- Shi, Q.; An, J.; Gagnon, K.K.; Cao, R.; Xie, H. Image Edge Detection Based on the Canny Edge and the Ant Colony Optimization Algorithm. In Proceedings of the IEEE 2019 12th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Suzhou, China, 19–21 October 2019; pp. 1–6. [Google Scholar]
- Ali, M.M.; Yannawar, P.; Gaikwad, A.T. Study of edge detection methods based on palmprint lines. In Proceedings of the IEEE 2016 International Conference on Electrical, Electronics, and Optimization Techniques (ICEEOT), Chennai, India, 3–5 March 2016; pp. 1344–1350. [Google Scholar]
| Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).