Next Article in Journal
A Threefold Approach for Enhancing Fuzzy Interpolative Reasoning: Case Study on Phishing Attack Detection Using Sparse Rule Bases
Next Article in Special Issue
Atrial Fibrillation Type Classification by a Convolutional Neural Network Using Contrast-Enhanced Computed Tomography Images
Previous Article in Journal
A Proposed Method of Automating Data Processing for Analysing Data Produced from Eye Tracking and Galvanic Skin Response
Previous Article in Special Issue
Quantum Image Compression: Fundamentals, Algorithms, and Advances
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modified Multiresolution Convolutional Neural Network for Quasi-Periodic Noise Reduction in Phase Shifting Profilometry for 3D Reconstruction

by
Osmar Antonio Espinosa-Bernal
*,
Jesús Carlos Pedraza-Ortega
*,
Marco Antonio Aceves-Fernandez
,
Juan Manuel Ramos-Arreguín
,
Saul Tovar-Arriaga
and
Efrén Gorrostieta-Hurtado
Facultad de Ingeniería, Universidad Autónoma de Querétaro, Querétaro 76010, Mexico
*
Authors to whom correspondence should be addressed.
Computers 2024, 13(11), 290; https://doi.org/10.3390/computers13110290
Submission received: 24 September 2024 / Revised: 31 October 2024 / Accepted: 2 November 2024 / Published: 8 November 2024
(This article belongs to the Special Issue Advanced Image Processing and Computer Vision)

Abstract

:
Fringe profilometry is a method that obtains the 3D information of objects by projecting a pattern of fringes. The three-step technique uses only three images to acquire the 3D information from an object, and many studies have been conducted to improve this technique. However, there is a problem that is inherent to this technique, and that is the quasi-periodic noise that appears due to this technique and considerably affects the final 3D object reconstructed. Many studies have been carried out to tackle this problem to obtain a 3D object close to the original one. The application of deep learning in many areas of research presents a great opportunity to to reduce or eliminate the quasi-periodic noise that affects images. Therefore, a model of convolutional neural network along with four different patterns of frequencies projected in the three-step technique is researched in this work. The inferences produced by models trained with different frequencies are compared with the original ones both qualitatively and quantitatively.

1. Introduction

The fringe projection is one method without contact that permits to measure heights from objects to generate 3D objects, and it is considered one of the most reliable for this aim [1,2,3].
The acquisition of 3D information is very essential in many areas, e.g., computer vision [4,5,6], industrial applications [7,8,9], optics [10,11], and biomedical applications [12,13,14], among others [15]. However, this method presents an inconvenience in the final 3D reconstruction due to the quasi-periodic noise [16,17,18,19,20] that is produced during the acquisition of images at the stage of phase unwrapping [21,22]. This stage of phase unwrapping recovers the 3D information from the image capture depending on the number of images. In this work, we apply the three-step technique [1,17], and therefore three images are required. This quasi-periodic or Moire noise, as is also known, has the particularity of affecting the shape of the 3D object [8,23,24,25], as it is shown in Figure 1, and it depends on the frequency of the pattern employed in the projection. This number of frequencies affects the way the noise appears in the images, as it is shown below in Figure 1 and Figure 2.
The reduction or elimination in periodic or quasi-periodic noise, known as Moire noise, began as soon as the first digital images could be obtained; however, it was not until it was analyzed in terms of how much noise was produced that research began on ways to attenuate or eliminate it from the images. Once the noise on the images was detected and analyzed, it was found that it is formed in a repetitive pattern and in different ways. Many studies have been conducted to reduce or eliminate such quasi-periodic noise, some processing the image in its spatial domain [18], others in its frequency domain [17]. In recent studies, thanks to the advances in artificial intelligence, specifically in the field of deep learning, images can be processed for different tasks, including image reduction or restoration. Convolutional neural networks are networks composed of neurons and are part of deep learning. They are composed of many layers of stacked neurons. These networks are designed to process an image by convolution, which is a technique that infers a pixel by calculating an average from information of neighboring pixels [26].
When all pixels of an image are completely processed, a complete image is produced by a model that is trained for this specific task. Many tasks are carried out with convolutional neural networks, such as classification [27,28], segmentation [29,30], restoration [31,32], object detection [33,34], among other tasks [35,36,37,38,39].
In this work, we propose a convolutional neural network to restore images affected by quasi-periodic noise in the process of 3D reconstruction by using the technique of fringe projection in three steps. The trained model will act as an image pre-processor by reducing the repetitive pattern present in the affected images, whose pattern appears like horizontal fringes that affect the surface of an object, improving the speed of this stage and obtaining an accurate 3D object. The convolutional neural network is based on the same architecture proposed by Sun [40], namely the multiresolution convolutional neural network, for the reduction in Moire patterns in digital images whose parameters will be described below. Section 3 will describe the results and Section 4 will present the conclusions.

2. Materials and Methods

The software Blender (https://www.blender.org/) emulates a 3-step fringe profilometry system and the 3D object models used for generating the database were acquired from platform Turbosquid and were free to use. Here, 75 different 3D models were used, and in Figure 3 some examples of these models are shown.
In the simulated system, a lamp is used to project the fringes over objects; then, pictures are acquired with four different patterns. For capturing the images, a camera with a focal length setting of 28 mm [41] was selected and the size of the captured images was 512 × 512 pixels. With the simulation system, a database of 1350 images with different objects at different positions was generated, but as four frequencies were applied to the pattern projected over the object, the total images were 5400. Each scene was composed of 16 different pictures plus 12 more that correspond to the references −3 for each different pattern projected shared by every scene. Figure 4 shows a complete set of all pictures of a single scene that conform to the generated database [17,41].
Figure 2 shows a single picture affected by four different patterns of quasi-periodic noise, and the process to obtain such images is shown in Figure 5.
All images were obtained using a laptop with NVIDIA GeForce RTX 3060 graphic card with 6 Gb of memory RAM, 16 Gb of memory RAM, and an I7-10750H processor @2.60 GHz. The images for training were 90% of 1050 images, for validation 10% of 1050 images, and 300 additional images were used for the test set. All this was only used to train a model with one single frequency, which was either 4, 8, 16, or 32. For the training of a model with multiple frequencies, were combined images with all four frequencies, adding up to a total of 4200 images, 90% for the training set, 10% for the validation set, and 1200 images affected with the four different patterns for the test set. Figure 6 shows some images with quasi-periodic noise at different frequencies and their respective targets generated by Blender.
The database of images generated with Blender includes three images of different 3D models with a pattern of 4 different frequencies and shifting 120° degrees. Applying a phase unwrapping algorithm (in this case, the PEARLS algorithm [21]) to obtain an absolute phase image, results in the generation of an image with noise known as quasi-periodic noise or Moire noise, as shown in Figure 2. Such noise is inherent in the technique of 3-step fringe profilometry to obtain the heights of an object from images using a single camera and affects the final 3D object reconstructed by altering its shape and losing 3D information.
When the stage to obtain the absolute phase of images pre-processed with the PEARLS algorithm is finished, the images are used to generate a database. The algorithm called PEARLS (Phase Estimation using Adaptive Regularization based on Local Smoothing) is described in the following pseudocode:
1.
Each pixel ( x , y ) in h H
(a)
The zero-order phase φ h ˜ x , y estimate is calculated;
(b)
Adaptive window size is applied to estimates φ h ˜ x , y to properly select a window size h + ( x , y ) ;
(c)
Compute first-order phase estimates with adaptive window size;
(d)
end.
2.
Unwrap the phase φ h + ˜ using one of the procedures developed for noise-free data.
For further information, see [21].
The database is then used as the source for training a CNN to learn to reduce or eliminate the quasi-periodic noise present in images that are affected by such noise. Finally, the trained model is used in the stage of pre-processing to generate a 3D object as a filter of noise. The process of the methodology to generate the images to train a model to reduce the noise in images is shown in Figure 5.
Once the database of images is obtained, a convolutional neural network based on Multiresolution-CNN proposed by Sun [40] is applied. The proposed model was modified to have 9 layers after the down-sampling and up-sampling operation with 3 × 3 kernel and 64 channels which were completely convolutional, contrary to the original one that had 5 layers in this stage of the architecture. In addition to this change, a layer as input and output of grayscale images or one channel was added as well. Figure 7 shows the architecture developed and implemented.
The original Multiresolution-CNN model was developed to reduce the Moire noise in color and white and black images. Therefore, the proposed model was modified to be trained with images that contain both quasi-periodic and Moire noise and reduce such noise. The trained model to reduce such noise was finally used as part of pre-processing images to generate a 3D object, improving the speed and the quality of the 3D objects generated. The novelty of this paper relies on the proposal of a modified multiresolution CNN in order to reduce the quasi-periodic noise on phase shifting profilometry at four different frequencies to generate a more reliable 3D reconstruction of an object.
The model was trained using a set of 1050 grayscale images with a size of 549 × 540 but adjusted to a size of 512 × 512 before feeding the model, and five experiments were carried out, one for every frequency present in the images. The last experiment was carried out using the set of images of every frequency gathered in one set of 4200 images. The projected frequency patterns were 4, 8, 16, and 32 fringes. Every experiment was performed using the optimizer Adam() [42,43] and the MSELoss() function [44] to calculate the training and validation loss. Internally, the algorithm took 10% of images randomly to be used as validation set in every training.
The fringe profilometry method allows for obtaining information on object heights through images. Therefore, a large number of images with a wide variety of shapes, surfaces, and contours are required to remove this specific noise. Although there are techniques to augment data and give them variety during the training of models for noise reduction and image restoration, this first approximation was carried out without data augmentation. This is performed in order to observe the results obtained and make the corresponding improvements. However, since this is a specific noise to be reduced, we leave some training techniques for future work. For now, we just add a large variety of objects to have a model trained with enough data to generalize to the greatest number of possible scenarios or objects to reduce or restore noise in images affected by quasi-periodic noise.
The selection of the neural network architecture is based on a previously published article, wherein a comparative study between three different architectures was carried out and the most appropriate neural network for this purpose was selected using performance criteria [23].

2.1. Optimizer and Loss Function

The optimizer Adam() has the advantage of requiring little memory, and it is computationally efficient and has an adaptive estimation to calculate moments of first and second order.
m t = β 1 m t 1 + 1 β 1 g t
v t = β 2 v t 1 + 1 β 2 g t 2 ,
g t evaluates the gradient in a timestep t, m t calculates the average of moving, v t is the squared gradient, and β 1 and β 2 calculate the decay rates for every moment estimates.
Equation (3) calculates the MSELoss() function
l ( x , y ) = L = l 1 , , l N T , l n = ( x n y n ) 2 ,
the batch-size is represented by N, x, and y which represent the dimensions that form a matrix of a given size with n elements [42,43].

2.2. IMMSE

The inverse mean square error (IMMSE) is a metric used to evaluate the quality of reconstructed images by comparing the original image with the generated image. The IMMSE formula is based on the calculation of the mean square error (MSE), but is applied in an inverse manner [45]. This equation is shown below
I M S S E = 1 m n i = 1 m i = j n ( I ( i , j ) K ( i , j ) ) 2
where I is the original image, K is the processed image, m number of rows, n number of cols, and K ( i , j ) is the value of the corresponding pixel in the reconstructed image. The IMMSE provides a measure of how similar the two images are, where higher values indicate better quality.

2.3. PSNR (Peak Signal-to-Noise Ratio)

Peak Signal-to-Noise Ratio (PSNR) is a widely used metric to assess the quality of compressed or reconstructed images. PSNR measures the ratio of the maximum power of a signal (the original image) to the noise that affects the quality of its representation (the reconstructed image) [45,46]. It is defined as follows:
P S N R ( f , g ) = 10 l o g 10 ( 255 2 / M S E ( f , g ) )
where
M S E ( f , g ) = 1 M N i = 1 M j = 1 N ( f i j g i , j ) 2
A higher PSNR indicates that the reconstructed image is more similar to the original, i.e., it has less noise. Typical PSNR values for high-quality images are in the range of 30–50 dB.

3. Results

The results were obtained by experimenting with different parameters and the parameters proposed by the author of the model, which were based on the model proposed in [40]. These parameters are summarized in Table 1.
In addition to the parameters shown in Table 1, 50 epochs were set and, if the obtained model had a better validation loss, it was saved as the best, but if a bad validation loss was obtained, the model was penalized. After training the model, the results obtained after every epoch were charted to show the evolution of the training and validation loss and were scaled for a better appreciation of the loss. The graphs of the evolution of every training are shown in Figure 8.
Figure 8 shows the training and validation loss of every model trained. The evolution of training and validation loss shows a constant decline, which is consistent and shows that the model is effectively “learning”, and it is also observed that the learning is performed rapidly and at the end the rate of learning is very low.
The time of learning and the training and validation loss are shown in Table 2.
According to the data obtained after performing the training of the models, these trainings took around one hour to complete, while the training with the set that contains all the images with the four frequencies lasted a bove five hours because its set contained more than 4000 images. The training and validation loss reached values equal to or below 0.1, indicating a constant learning by the trained models.

3.1. Inferences Obtained from Images Affected with Quasi-Periodic Noise of 4 Frequencies

The inferences obtained from images affected by quasi-periodic noise composed of four frequencies using all the trained models are shown in Figure 9, and the 3D reconstructions are shown in Figure 10.
The profiles obtained from these inferences, the ground-truth image, and the original image affected by quasi-periodic noise of four frequencies are compared and are charted in Figure 11. The heights are normalized from 0.0 to 1.0 and the x-axis represents pixels.
The error between the inferences made by the models trained and the ground-truth image is identified using the PSNR, SSIM, IMMSE, and the MSE Profile between the inference and the ground-truth image. The measures obtained for the images affected by quasi-periodic noise of four frequencies are summarized in Table 3.

3.2. Inferences Obtained from Images Affected with Quasi-Periodic Noise of 8 Frequencies

The inferences obtained from images affected by quasi-periodic noise composed of 8 frequencies using all the trained models are shown in Figure 12, and the 3D reconstructions are shown in Figure 13.
The profiles obtained from these inferences, the ground-truth image, and the original image affected by quasi-periodic noise of eight frequencies are compared and are charted in Figure 14. The heights are normalized from 0.0 to 1.0 and the x-axis represents pixels.
The error between the inferences made by the models trained and the ground-truth image is identified using the PSNR, SSIM, IMMSE, and the MSE Profile between the inference and the ground-truth image. The measures obtained in images affected by quasi-periodic noise of eight frequencies are summarized in Table 4.

3.3. Inferences Obtained from IMAGES Affected with Quasi-Periodic Noise of 16 Frequencies

The inferences obtained from images affected by quasi-periodic noise composed of 16 frequencies using all the trained models are shown in Figure 15, and the 3D reconstructions are shown in Figure 16.
The profiles obtained from these inferences, the ground-truth image, and the original image affected by quasi-periodic noise of 16 frequencies are compared and are charted in Figure 17. The heights are normalized from 0.0 to 1.0 and the x-axis represents pixels.
The error between the inferences made by the models trained and the ground-truth image is identified using the PSNR, SSIM, IMMSE, and the MSE Profile between the inference and the ground-truth image. The measures obtained in images affected by quasi-periodic noise of 16 frequencies are summarized in Table 5.

3.4. Inferences Obtained from Images Affected with Quasi-Periodic Noise of 32 Frequencies

The inferences obtained from images affected by quasi-periodic noise composed of 32 frequencies using all the trained models are shown in Figure 18, and the 3D reconstructions are shown in Figure 19.
The profiles obtained from these inferences, the ground-truth image, and the original image affected by quasi-periodic noise of 32 frequencies are compared and are charted in Figure 20. The heights are normalized from 0.0 to 1.0 and the x-axis represents pixels.
The error between the inferences made by the models trained and the ground-truth image is identified using the PSNR, SSIM, IMMSE, and the MSE Profile between the inference and the ground-truth image. The measures obtained in images affected by quasi-periodic noise of 32 frequencies are summarized in Table 6.

4. Discussion

The inferences obtained from each trained model with a different set of images affected with quasi-periodic noise of different frequencies show, in every case, a better performance when the model is trained with a set that contains images affected with different frequencies instead of using only a set of images with one frequency. Although an image affected with quasi-periodic noise of only four frequencies appears to show a better similarity with the ground-truth image, preserving better details of the object, it is difficult for the trained models to obtain a better inference in quantitative terms. This is observed in the metrics shown in the Table 3, Table 4, Table 5 and Table 6. In quantitative terms, in the four inferences performed with 4, 8, 16, and 32 fringes, the IMMSE value is reduced in each inference compared to the original model. In addition, it is observed that the inferences with model 5, which was trained with multiple frequencies, presented a better performance in the SSIM, PSNR, and MSE (Profile) metrics, compared to models 1 to 4.
At first glance at the images with quasi-periodic noise, it can be seen that those that are affected by a lower frequency of such noise lose fewer details of the 3D object. However, as the number of fringe frequencies in the projected patterns increases, this quasi-periodic noise decreases, but only in size. It therefore merges and blends with the details of the 3D object, making it almost impossible to determine what is noise and what is part of the 3D information. Another effect shown by using a low fringe frequency in the projected patterns is that the final height of the object inferred by the model better preserves the original height of the object. This is clearly appreciated in the images that compare the profiles at each frequency of the projected pattern analyzed, where the inferences represented as model 3 (fringe pattern with 16 frequencies) and model 4 (fringe pattern with 16 frequencies) are always lower in normalized height.
It is expected that training using images affected by noise of the same frequency may adequately restore images with similar noise present; however, it was found that performance improved when using images affected by other frequencies. Therefore, training was carried out wherein all the images affected by different frequencies were put together, achieving better results. Although better results were obtained by generalizing the training data more by combining affected images with different frequencies, another limitation was the number of images. By increasing the number of images in the database, it may be possible to further improve the results obtained.
Generating data using Blender allows us to obtain data in a way that is very similar to real data. On the one hand, extensive methodologies must be followed, such as calibration of cameras and objects that do not have restrictions of any kind, while synthetic data save us time and can be used freely. Since models tend to mimic real-world objects, it is possible to use them to represent even people or people’s faces roughly, but without the inconvenience of having to obtain them from real people. Furthermore, one can include images of objects captured from the real world, and carry out the 3D reconstruction process at the testing stage.
These results show the difficulty of eliminating the quasi-periodic noise that affects this particular fringe profilometry method for 3D reconstruction, even when trying with different frequencies. Trying different frequencies was found out that the speed of acquisition of image by fringe profilometry of 3-step, while less the frequency pattern projected is faster than with a high frequency. The next research will aim to improve the inferences obtained by either increasing the number of images in the training set or trying other models of convolutional neural networks or networks known as GAN.

5. Conclusions

The experiments performed using a set of images affected with quasi-periodic noise of four different frequencies show how these frequencies affect the 3D object reconstructed and the results obtained when an inference is generated after training a CNN model with these images. Quantitative results show better performance when the model is trained with a set of images that contains, in this case, a quasi-periodic noise pattern of four different frequencies showing that images affected with a higher frequency are the ones that obtain a better result and visually show greater similarity with the ground-truth image.
On the other hand, using a model trained to reduce noise in images obtained in PSP increases the speed of image pre-processing to obtain a 3D object. Trying different frequencies to produce images with different kinds of noise helps to create a high variety of such noise in datasets to train models of CNNs, generating good results both quantitatively and qualitatively.

Author Contributions

Conceptualization, O.A.E.-B., J.C.P.-O., M.A.A.-F., S.T.-A., J.M.R.-A., and E.G.-H.; methodology, O.A.E.-B.; software, O.A.E.-B.; validation, J.C.P.-O.; formal analysis, J.C.P.-O.; investigation, O.A.E.-B.; data curation, O.A.E.-B.; writing—original draft preparation, O.A.E.-B.; writing—review and editing, J.C.P.-O., M.A.A.-F., S.T.-A., J.M.R.-A., and E.G.-H.; supervision, J.C.P.-O., M.A.A.-F., S.T.-A., J.M.R.-A., and E.G.-H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

This work was supported in part by the Consejo Nacional de Humanidades, Ciencias y Tecnologías (CONAHCYT), México, in the Postgraduate Faculty of Engineering by the Universidad Autonoma de Querétaro, under Grant CVU 1099050. We also would like to thank FONDO PARA EL FORTALECIMIENTO DE LA INVESTIGACIÓN, VINCULACIÓN Y EXTENSIÓN (FONFIVE-UAQ 2024) for the support of this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
MSEMedia Square Error
PEARLSPhase Estimation using Adaptive Regularization based on Local Smothing

References

  1. Gorthi, S.S.; Rastogi, P. Fringe projection techniques: Whither we are? Opt. Lasers Eng. 2010, 48, 133–140. [Google Scholar] [CrossRef]
  2. Feng, S.; Zuo, C.; Zhang, L.; Tao, T.; Hu, Y.; Yin, W.; Qian, J.; Chen, Q. Calibration of fringe projection pro-filometry: A comparative review. Opt. Lasers Eng. 2021, 143, 106622. [Google Scholar] [CrossRef]
  3. Hu, Y.; Chen, Q.; Feng, S.; Zuo, C. Microscopic fringe projection profilometry: A review. Opt. Lasers Eng. 2020, 135, 106192. [Google Scholar] [CrossRef]
  4. Frank Chen, G.M.; Mumin, S. Overview of three-dimensional shape measurement using optical methods. Opt. Eng. 2000, 39, 10–21. [Google Scholar]
  5. Huang, L.; Idir, M.; Zuo, C.; Asundi, A. Review of phase measuring deflectometry. Opt. Lasers Eng. 2018, 107, 247–257. [Google Scholar] [CrossRef]
  6. Chen, C.; Gao, N.; Wang, X.; Zhang, Z.; Gao, F.; Jiang, X. Generic exponential fringe model for alleviating phase error in phase measuring profilometry. Opt. Lasers Eng. 2018, 110, 179–185. [Google Scholar] [CrossRef]
  7. Land, W.S., II; Zhang, B.; Ziegert, J.; Davies, A. In-situ metrology system for laser powder bed fusion additive process. Procedia Manuf. 2015, 1, 393–403. [Google Scholar] [CrossRef]
  8. Li, B.; Xu, Z.; Gao, F.; Cao, Y.; Dong, Q. 3D reconstruction of high reflective welding surface based on binocular structured light stereo vision. Machines 2022, 10, 159. [Google Scholar] [CrossRef]
  9. Sun, B.; Zheng, G.; Zhang, X.; Bai, L. Research on aero-engine blade surface detection based on three datum points integrating algorithm. AIP Adv. 2020, 10, 075305. [Google Scholar] [CrossRef]
  10. Qian, J.; Feng, S.; Tao, T.; Hu, Y.; Liu, K.; Wu, S.; Chen, Q.; Zuo, C. High-resolution real-time 360 3d model reconstruction of a handheld object with fringe projection profilometry. Opt. Lett. 2019, 44, 5751–5754. [Google Scholar] [CrossRef]
  11. Song, K.; Hu, S.; Wen, X.; Yan, Y. Fast 3D shape measurement using Fourier transform profilometry without phase unwrapping. Opt. Lasers Eng. 2016, 84, 74–81. [Google Scholar] [CrossRef]
  12. Jiang, C.; Jia, S.; Xu, Y.; Bao, Q.; Dong, J.; Lian, Q. The application of multi-frequency fringe projection profilometry on the measurement of biological tissues. Biomed. Mater. Eng. 2015, 26, S395–S403. [Google Scholar] [CrossRef] [PubMed]
  13. Chatterjee, A.; Dhanotia, J.; Bhatia, V.; Prakash, S. Non-destructive 3D profiling of orthopaedic titanium bone plate using fringe projection profilometry and Fourier transform analysis. In Proceedings of the 2017 6th International Conference on Computer Applications in Electrical Engineering-Recent Advances (CERA), Roorkee, India, 5–7 October 2017; pp. 389–392. [Google Scholar]
  14. Chatterjee, A.; Singh, P.; Bhatia, V.; Prakash, S. Ear biometrics recognition using laser biospeckled fringe projection profilometry. Opt. Laser Technol. 2019, 112, 368–378. [Google Scholar] [CrossRef]
  15. Xing, H.Z.; Zhang, Q.B.; Braithwaite, C.H.; Pan, B.; Zhao, J. High-speed photography and digital optical measurement techniques for geomaterials: Fundamentals and applications. Rock Mech. Rock Eng. 2017, 50, 1611–1659. [Google Scholar] [CrossRef]
  16. Aizenberg, I.N.; Butakoff, C. Frequency domain medianlike filter for periodic and quasi-periodic noise removal. Image Process. Algorithms Syst. 2002, 4667, 181–191. [Google Scholar]
  17. Espinosa-Bernal, O.A.; Pedraza-Ortega, J.C.; Aceves-Fernández, M.A.; Martínez-Suárez, V.M.; Tovar-Arriaga, S. Adaptive Based Frequency Domain Filter for Periodic Noise Reduction in Images Acquired by Projection Fringes. In International Congress of Telematics and Computing; Springer International Publishing: Cham, Switzerland, 2022; pp. 18–32. [Google Scholar]
  18. Aizenberg, I.; Butakoff, C. A windowed Gaussian notch filter for quasi-periodic noise removal. Image Vis. Comput. 2008, 26, 1347–1353. [Google Scholar] [CrossRef]
  19. López-Torres, C.V.; Salazar Colores, S.; Kells, K.; Pedraza-Ortega, J.C.; Ramos-Arreguin, J.M. Improving 3D reconstruction accuracy in wavelet transform profilometry by reducing shadow effects. IET Image Process. 2020, 14, 310–317. [Google Scholar] [CrossRef]
  20. Wang, J.; Yang, Y. Phase extraction accuracy comparison based on multi-frequency phase-shifting method in fringe projection profilometry. Measurement 2022, 199, 111525. [Google Scholar] [CrossRef]
  21. Bioucas-Dias, J.; Katkovnik, V.; Astola, J.; Egiazarian, K. Absolute phase estimation: Adaptive local denoising and global unwrapping. Appl. Opt. 2008, 47, 5358–5369. [Google Scholar] [CrossRef]
  22. Bioucas-Dias, J.M.; Valadao, G. Phase unwrapping via graph cuts. IEEE Trans. Image Process. 2007, 16, 698–709. [Google Scholar] [CrossRef]
  23. Espinosa-Bernal, O.A.; Pedraza-Ortega, J.C.; Aceves-Fernandez, M.A.; Martínez-Suárez, V.M.; Tovar-Arriaga, S.; Ramos-Arreguín, J.M.; Gorrostieta-Hurtado, E. Quasi/Periodic Noise Reduction in Images Using Modified Multiresolution-Convolutional Neural Networks for 3D Object Reconstructions and Comparison with Other Convolutional Neural Network Models. Computers 2024, 13, 145. [Google Scholar] [CrossRef]
  24. Qian, J.; Feng, S.; Tao, T.; Hu, Y.; Li, Y.; Chen, Q.; Zuo, C. Deep-learning-enabled geometric constraints and phase unwrapping for single-shot absolute 3D shape measurement. APL Photonics 2020, 5, 046105. [Google Scholar] [CrossRef]
  25. Alvarado Escoto, L.A.; Ortega, J.C.P.; Ramos Arreguin, J.M.; Gorrostieta Hurtado, E.; Tovar Arriaga, S. The effect of bilateral filtering in 3D reconstruction using PSP. In Telematics and Computing, Proceedings of the 9th International Congress, WITCOM 2020, Puerto Vallarta, Mexico, 2–6 November 2020; Proceedings 9; Springer International Publishing: Cham, Switzerland, 2020; pp. 268–280. [Google Scholar]
  26. Chollet, F. Deep Learning with Python; Manning Publications: Shelter Island, NY, USA, 2020. [Google Scholar]
  27. Dhiman, P.; Kaur, A.; Balasaraswathi, V.R.; Gulzar, Y.; Alwan, A.A.; Hamid, Y. Image acquisition, preprocessing and classification of citrus fruit diseases: A systematic literature review. Sustainability 2023, 15, 9643. [Google Scholar] [CrossRef]
  28. Alkhatib, M.Q.; Al-Saad, M.; Aburaed, N.; Almansoori, S.; Zabalza, J.; Marshall, S.; Al-Ahmad, H. Tri-CNN: A three branch model for hyperspectral image classification. Remote Sens. 2023, 15, 316. [Google Scholar] [CrossRef]
  29. Yuan, F.; Zhang, Z.; Fang, Z. An effective CNN and Transformer complementary network for medical image segmentation. Pattern Recognit. 2023, 136, 109228. [Google Scholar] [CrossRef]
  30. Nasreen, G.; Haneef, K.; Tamoor, M.; Irshad, A. A comparative study of state-of-the-art skin image segmentation techniques with CNN. Multimed. Tools Appl. 2023, 82, 10921–10942. [Google Scholar] [CrossRef]
  31. Ali, A.M.; Benjdira, B.; Koubaa, A.; El-Shafai, W.; Khan, Z.; Boulila, W. Vision transformers in image restoration: A survey. Sensors 2023, 23, 2385. [Google Scholar] [CrossRef]
  32. Wang, Q.; Li, Z.; Zhang, S.; Chi, N.; Dai, Q. A versatile Wavelet-Enhanced CNN-Transformer for improved fluorescence microscopy image restoration. Neural Netw. 2024, 170, 227–241. [Google Scholar] [CrossRef]
  33. Shah, A.; Shah, M.; Pandya, A.; Sushra, R.; Sushra, R.; Mehta, M.; Patel, K.; Patel, K. A comprehensive study on skin cancer detection using artificial neural network (ANN) and convolutional neural net-work (CNN). Clin. eHealth 2023, 6, 76–84. [Google Scholar] [CrossRef]
  34. Jakubec, M.; Lieskovská, E.; Bučko, B.; Zábovská, K. Comparison of CNN-based models for pothole detection in real-world adverse conditions: Overview and evaluation. Appl. Sci. 2023, 13, 5810. [Google Scholar] [CrossRef]
  35. Dash, A.; Ye, J.; Wang, G. A review of generative adversarial networks (GANs) and its applications in a wide variety of disciplines: From medical to remote sensing. IEEE Access 2024, 12, 18330–18357. [Google Scholar] [CrossRef]
  36. Chakraborty, T.; KS, U.R.; Naik, S.M.; Panja, M.; Manvitha, B. Ten years of generative adversarial nets (GANs): A survey of the state-of-the-art. Mach. Learn. Sci. Technol. 2024, 5, 011001. [Google Scholar] [CrossRef]
  37. Ahmad, Z.; Jaffri, Z.U.A.; Chen, M.; Bao, S. Understanding GANs: Fundamentals, variants, training challenges, applications, and open problems. Multimed. Tools Appl. 2024, 1–77. [Google Scholar] [CrossRef]
  38. Dunmore, A.; Jang-Jaccard, J.; Sabrina, F.; Kwak, J. A comprehensive survey of generative adver-sarial networks (GANs) in cybersecurity intrusion detection. IEEE Access 2023, 11, 76071–76094. [Google Scholar] [CrossRef]
  39. Chen, C.; Wu, Y.; Dai, Q.; Zhou, H.Y.; Xu, M.; Yang, S.; Han, X.; Yu, Y. A survey on graph neural networks and graph transformers in computer vision: A task-oriented perspective. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 10297–10318. [Google Scholar] [CrossRef]
  40. Sun, Y.; Yu, Y.; Wang, W. Moiré photo restoration using multiresolution convolutional neural networks. IEEE Trans. Image Process. 2018, 27, 4160–4172. [Google Scholar] [CrossRef]
  41. Martínez-Suárez, V.M.; Pedraza-Ortega, J.C.; Salazar-Colores, S.; Espinosa-Bernal, O.A.; Ra-mos-Arreguin, J.M. Environment emulation in 3d graphics software for fringe projection profilometry. In International Congress of Telematics and Computing; Springer International Publishing: Cham, Switzerland, 2022; pp. 122–138. [Google Scholar]
  42. Haji, S.H.; Abdulazeez, A.M.; Darrell, T. Comparison of optimization techniques based on gradient descent algorithm: A review. PalArch’s J. Archaeol. Egypt/Egyptol. 2021, 18, 2715–2743. [Google Scholar]
  43. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  44. Jiang, J.; Bao, S.; Shi, W.; Wei, Z. Improved traffic sign recognition algorithm based on YOLO v3 algorithm. J. Comput. Appl. 2020, 40, 2472. [Google Scholar]
  45. Martinez-Carranza, J.; Falaggis, K.; Kozacki, T. Fast and accurate phase-unwrapping algorithm based on the transport of intensity equation. Appl. Opt. 2017, 56, 7079–7088. [Google Scholar] [CrossRef]
  46. Hore, A.; Ziou, D. Image quality metrics: PSNR vs.SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
Figure 1. Three-dimensional reconstruction of an object, (a) affected by quasi-periodic noise, and (b) original object. The image shows the deformation of the surface caused by the noise present in images acquired by the fringe projection in three steps.
Figure 1. Three-dimensional reconstruction of an object, (a) affected by quasi-periodic noise, and (b) original object. The image shows the deformation of the surface caused by the noise present in images acquired by the fringe projection in three steps.
Computers 13 00290 g001
Figure 2. Images from the database with quasi-periodic noise at different frequencies: (a) quasi-periodic noise at 4 frequencies, (b) quasi-periodic noise at 8 frequencies, (c) quasi-periodic noise at 16 frequencies, (d) quasi-periodic noise at 32 frequencies.
Figure 2. Images from the database with quasi-periodic noise at different frequencies: (a) quasi-periodic noise at 4 frequencies, (b) quasi-periodic noise at 8 frequencies, (c) quasi-periodic noise at 16 frequencies, (d) quasi-periodic noise at 32 frequencies.
Computers 13 00290 g002aComputers 13 00290 g002b
Figure 3. Three-dimensional models acquired from platform Turbosquid.
Figure 3. Three-dimensional models acquired from platform Turbosquid.
Computers 13 00290 g003
Figure 4. Set of images obtained from a single scene with a 3D model. (a) Ground-truth, (b) original 3D model, (c) region of interest, (d) 3D model with background, (eg) images with object with 120° shifting pattern projected composed of 4 frequencies, (hj) reference images with a 4-frequency composite pattern, (km) images with object with 120° shifting pattern projected composed of 8 frequencies, (np) reference images with a 8-frequency composite pattern, (qs) images with object with 120° shifting pattern projected composed of 16 frequencies, (tv) reference images with a 16-frequency composite pattern, (wy) images with object with 120° shifting pattern projected composed of 32 frequencies, (z,aa,ab) reference images with a 32-frequency composite pattern.
Figure 4. Set of images obtained from a single scene with a 3D model. (a) Ground-truth, (b) original 3D model, (c) region of interest, (d) 3D model with background, (eg) images with object with 120° shifting pattern projected composed of 4 frequencies, (hj) reference images with a 4-frequency composite pattern, (km) images with object with 120° shifting pattern projected composed of 8 frequencies, (np) reference images with a 8-frequency composite pattern, (qs) images with object with 120° shifting pattern projected composed of 16 frequencies, (tv) reference images with a 16-frequency composite pattern, (wy) images with object with 120° shifting pattern projected composed of 32 frequencies, (z,aa,ab) reference images with a 32-frequency composite pattern.
Computers 13 00290 g004
Figure 5. The methodology used to generate a database of images with quasi-periodic noise.
Figure 5. The methodology used to generate a database of images with quasi-periodic noise.
Computers 13 00290 g005
Figure 6. Images from database created with Blender software: (a,c,e,g) images affected with quasi-periodic noise at different frequencies, (b,d,f,h) ground-truth image.
Figure 6. Images from database created with Blender software: (a,c,e,g) images affected with quasi-periodic noise at different frequencies, (b,d,f,h) ground-truth image.
Computers 13 00290 g006
Figure 7. The architecture of convolutional neural network model developed and implemented.
Figure 7. The architecture of convolutional neural network model developed and implemented.
Computers 13 00290 g007
Figure 8. Evolution of training and validation loss. Models train with noisy images affected by different frequencies due to different patterns projected. (a) Images with 4 frequencies, (b) images with 8 frequencies, (c) images with 16 frequencies (d) images with 32 frequencies, and (e) images with multifrequencies (4, 8, 16, and 32).
Figure 8. Evolution of training and validation loss. Models train with noisy images affected by different frequencies due to different patterns projected. (a) Images with 4 frequencies, (b) images with 8 frequencies, (c) images with 16 frequencies (d) images with 32 frequencies, and (e) images with multifrequencies (4, 8, 16, and 32).
Computers 13 00290 g008
Figure 9. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a four-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 9. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a four-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g009
Figure 10. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a four-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 10. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a four-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g010
Figure 11. Profile comparison of 3D objects.
Figure 11. Profile comparison of 3D objects.
Computers 13 00290 g011
Figure 12. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of an 8-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 12. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of an 8-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g012
Figure 13. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of an 8-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 13. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of an 8-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g013
Figure 14. Profile comparison of 3D objects.
Figure 14. Profile comparison of 3D objects.
Computers 13 00290 g014
Figure 15. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 16-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 15. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 16-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g015aComputers 13 00290 g015b
Figure 16. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 16-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 16. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 16-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g016
Figure 17. Profile comparison of 3D objects.
Figure 17. Profile comparison of 3D objects.
Computers 13 00290 g017
Figure 18. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 32-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 18. Two-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 32-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g018
Figure 19. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 32-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Figure 19. Three-dimensional representation of an object Cat. (a) Image with quasi-periodic noise produced by projection of a 32-frequency pattern, inference obtained with models trained with (b) four frequencies, (c) 8 frequencies, (d) 16 frequencies, (e) 32 frequencies, and (f) Multifrequencies. (g) ground-truth image, and (h) original object.
Computers 13 00290 g019
Figure 20. Profile comparison of 3D objects.
Figure 20. Profile comparison of 3D objects.
Computers 13 00290 g020
Table 1. Parameters used during network training for comparison, trained with images affected by quasi-periodic noise at four different patterns (4, 8, 16, and 32 frequencies), as seen in Figure 2.
Table 1. Parameters used during network training for comparison, trained with images affected by quasi-periodic noise at four different patterns (4, 8, 16, and 32 frequencies), as seen in Figure 2.
ParameterPattern 1 (Number of Fringes 4)Pattern 2 (Number of Fringes 8)Pattern 3 (Number of Fringes 16)Pattern 4 (Number of Fringes 32)Pattern 5 (Multifrequency Pattern)
Batch size44444
Initials weightsGaussian random (average = 0.0, standard deviation = 0.01)Gaussian random (average = 0.0, standard deviation = 0.01)Gaussian random (average = 0.0, standard deviation = 0.01)Gaussian random (average = 0.0, standard deviation = 0.01)Gaussian random (average = 0.0, standard deviation = 0.01)
Bias0.00.00.00.00.0
Learning rate0.0070.0070.0070.0070.007
OptimizerAdam()Adam()Adam()Adam()Adam()
Training lossMSELoss()MSELoss()MSELoss()MSELoss()MSELoss()
Validation lossMSELoss()MSELoss()MSELoss()MSELoss()MSELoss()
Test planing (train, val)90%, 10%90%, 10%90%, 10%90%, 10%90%, 10%
Images size (Width, Height) 512 × 512 pixels 512 × 512 pixels 512 × 512 pixels 512 × 512 pixels 512 × 512 pixels
Set train images10501050105010504200
Set validation images105105105105420
Set test images300300300300300
Table 2. Time employed to perform each training and training and validation loss reached during network training for comparison using images with four different patterns (4, 8, 16, and 32 frequencies), as seen in Figure 2.
Table 2. Time employed to perform each training and training and validation loss reached during network training for comparison using images with four different patterns (4, 8, 16, and 32 frequencies), as seen in Figure 2.
Pattern 1 (Number of Fringes 4)Pattern 2 (Number of Fringes 8)Pattern 3 (Number of Fringes 16)Pattern 4 (Number of Fringes 32)Pattern 5 (Multifrequency Pattern)
Training loss0.102750.119390.098010.088250.12041
Validation loss0.111870.103900.100420.097490.10443
Training time (HH:MM:SS)0:59:371:08:490:58:121:00:165:20:22
Table 3. Measures obtained with model trained with images affected by noise of four frequencies.
Table 3. Measures obtained with model trained with images affected by noise of four frequencies.
InferenceIMMSESSIMPSNRMSE (Profile)
10.0220.87164.6760.064
20.0170.87965.7670.048
30.0330.82862.9000.089
40.0460.79361.5470.124
50.0120.87367.2630.034
Table 4. Measures obtained with model trained with images affected by noise of 8 frequencies.
Table 4. Measures obtained with model trained with images affected by noise of 8 frequencies.
InferenceIMMSESSIMPSNRMSE (Profile)
10.0170.88265.8380.048
20.0120.88967.4880.031
30.0250.84664.2240.063
40.0360.81362.5610.095
50.0070.87869.6460.018
Table 5. Measures obtained with model trained with images affected by noise of 16 frequencies.
Table 5. Measures obtained with model trained with images affected by noise of 16 frequencies.
InferenceIMMSESSIMPSNRMSE (Profile)
10.0140.88666.5170.043
20.0090.90368.5490.025
30.0170.89765.7710.050
40.0280.87263.6090.082
50.0050.91471.4650.011
Table 6. Measures obtained with model trained with images affected by noise of 32 frequencies.
Table 6. Measures obtained with model trained with images affected by noise of 32 frequencies.
InferenceIMMSESSIMPSNRMSE (Profile)
10.0100.90568.3070.027
20.0050.92371.5430.011
30.0100.92268.0980.028
40.0190.90165.2730.054
50.0020.92775.1160.002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Espinosa-Bernal, O.A.; Pedraza-Ortega, J.C.; Aceves-Fernandez, M.A.; Ramos-Arreguín, J.M.; Tovar-Arriaga, S.; Gorrostieta-Hurtado, E. Modified Multiresolution Convolutional Neural Network for Quasi-Periodic Noise Reduction in Phase Shifting Profilometry for 3D Reconstruction. Computers 2024, 13, 290. https://doi.org/10.3390/computers13110290

AMA Style

Espinosa-Bernal OA, Pedraza-Ortega JC, Aceves-Fernandez MA, Ramos-Arreguín JM, Tovar-Arriaga S, Gorrostieta-Hurtado E. Modified Multiresolution Convolutional Neural Network for Quasi-Periodic Noise Reduction in Phase Shifting Profilometry for 3D Reconstruction. Computers. 2024; 13(11):290. https://doi.org/10.3390/computers13110290

Chicago/Turabian Style

Espinosa-Bernal, Osmar Antonio, Jesús Carlos Pedraza-Ortega, Marco Antonio Aceves-Fernandez, Juan Manuel Ramos-Arreguín, Saul Tovar-Arriaga, and Efrén Gorrostieta-Hurtado. 2024. "Modified Multiresolution Convolutional Neural Network for Quasi-Periodic Noise Reduction in Phase Shifting Profilometry for 3D Reconstruction" Computers 13, no. 11: 290. https://doi.org/10.3390/computers13110290

APA Style

Espinosa-Bernal, O. A., Pedraza-Ortega, J. C., Aceves-Fernandez, M. A., Ramos-Arreguín, J. M., Tovar-Arriaga, S., & Gorrostieta-Hurtado, E. (2024). Modified Multiresolution Convolutional Neural Network for Quasi-Periodic Noise Reduction in Phase Shifting Profilometry for 3D Reconstruction. Computers, 13(11), 290. https://doi.org/10.3390/computers13110290

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop