Next Article in Journal
Power Grids and Instrument Transformers up to 150 kHz: A Review of Literature and Standards
Next Article in Special Issue
Online Scene Semantic Understanding Based on Sparsely Correlated Network for AR
Previous Article in Journal
A Compensation Method for Full-Field-of-View Energy Nonuniformity in Dark-and-Weak-Target Simulators
Previous Article in Special Issue
Research on Three-Dimensional Reconstruction of Ribs Based on Point Cloud Adaptive Smoothing Denoising
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural Colour Correction for Indoor 3D Reconstruction Using RGB-D Data

by
Tiago Madeira
1,2,
Miguel Oliveira
1,3 and
Paulo Dias
1,2,*
1
Institute of Electronics and Informatics Engineering of Aveiro (IEETA), Intelligent System Associate Laboratory (LASI), University of Aveiro, 3810-193 Aveiro, Portugal
2
Department of Electronics, Telecommunications and Informatics (DETI), University of Aveiro, 3810-193 Aveiro, Portugal
3
Department of Mechanical Engineering (DEM), University of Aveiro, 3810-193 Aveiro, Portugal
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(13), 4141; https://doi.org/10.3390/s24134141
Submission received: 18 May 2024 / Revised: 17 June 2024 / Accepted: 25 June 2024 / Published: 26 June 2024
(This article belongs to the Special Issue 3D Reconstruction with RGB-D Cameras and Multi-sensors)

Abstract

:
With the rise in popularity of different human-centred applications using 3D reconstruction data, the problem of generating photo-realistic models has become an important task. In a multiview acquisition system, particularly for large indoor scenes, the acquisition conditions will differ along the environment, causing colour differences between captures and unappealing visual artefacts in the produced models. We propose a novel neural-based approach to colour correction for indoor 3D reconstruction. It is a lightweight and efficient approach that can be used to harmonize colour from sparse captures over complex indoor scenes. Our approach uses a fully connected deep neural network to learn an implicit representation of the colour in 3D space, while capturing camera-dependent effects. We then leverage this continuous function as reference data to estimate the required transformations to regenerate pixels in each capture. Experiments to evaluate the proposed method on several scenes of the MP3D dataset show that it outperforms other relevant state-of-the-art approaches.

1. Introduction

Three-dimensional (3D) reconstruction is the creation of 3D models from the captured shape and appearance of real objects. It is a field that has its roots in several areas within computer vision and graphics, and has gained high importance in others, such as architecture, robotics, autonomous driving, medicine, agriculture, and archaeology. With the rise in popularity of different human-centred applications requiring photo-realistic 3D reconstruction, such as virtual reality (VR) and augmented reality (AR) experiences, the problem of generating high-fidelity models has become an important task.
In a multiview acquisition system, particularly for large indoor scenes, the acquisition conditions, such as lighting, exposure, white balance, and other camera parameters, will differ along the environment, causing colour differences between captures and unappealing visual artefacts in the produced models. Different colour correction methodologies have been proposed for image stitching [1,2,3] algorithms. The global optimization strategy is widely used, and the colour correspondences are typically calculated based on the overlapping area between the images. When it comes to 3D reconstruction, especially through sparse RGB-D captures, this type of methodology cannot be directly applied, as the pose of the cameras will vary significantly in 6DOF within the environment. In this case, dense matching can be employed [4], but it is a very resource intensive process.
We propose a neural-based approach, using a multilayer perceptron (MLP) to learn an implicit representation of the colour in 3D space, while capturing camera-dependent effects. Then, to estimate the transformations required to regenerate pixels in each capture, one smaller MLP is trained using information provided by the larger MLP as reference colour. In the context of 3D reconstruction, even with a reduced number of pixel correspondences between images, a global optimization has scalability issues, as the combinations of images are exponential and processing of the captures required for an accurate reconstruction of an indoor scene can quickly become unfeasible. Conversely, our approach uses lightweight representations and is efficient in the integration process, using surface information derived from RGB-D data to create the continuous function of colour. Experiments to evaluate the proposed method on several scenes of the MP3D dataset show that it outperforms other relevant state-of-the-art approaches.
The remainder of this document is organized as follows: in Section 2, the related work is presented; in Section 3 the proposed methodology is described; in Section 4, the results are showcased and discussed; finally, Section 5 provides concluding remarks.

2. Related Work

In this section, the topic of colour correction is introduced, starting with pairwise techniques, also known as colour transfer, then going over colour consistency correction for image sets.

2.1. Pairwise Colour Correction

Depending on their foundational mapping principle, methodologies for colour correction of image pairs, or colour transfer, can be systematically categorized into two classes: nonparametric and parametric.
The majority of nonparametric methods are based on the direct mapping of all colour levels using a 2D joint image histogram (JIH). This is calculated from the known colour correspondences between the two images, often relying on overlapped regions. Pitié et al. [5] proposed a nonlinear iterative method, based on the one-dimensional probability density function transfer, mapping one N-dimensional distribution to another to create a colour mapping table. In [6], Su et al. used a gradient-aware decomposition model to separate the source image into several detail layers and base layers, performing colour transfer for each base layer using the previous methodology [5]. The base layers were then merged with boosted detail layers to generate the final output. De Marchi et al. [7] utilized a Vandermonde matrix to model the colour mapping function as a polynomial. Hwang et al. [8] combined moving least squares with the probabilistic modelling of the colour transfer in 3D colour space. Liu et al. [9] used unsupervised deep learning to separate the image into reflectance and shading components. Palette-based recolouring was then applied to the reflectance element, preventing colour overflow. Because the previously mentioned approaches are unable to fix local colour difference, local colour correction methods have been proposed. Wu et al. [10] used a content-based local colour transfer algorithm, integrating the spatial distribution of target colour style in the optimization process. Finlayson et al. [11] introduced a root-polynomial regression, invariant to scene irradiance and to camera exposure. Hwang et al. [12] improved upon their previous algorithm [8] by proposing bilateral weights, performing local colour transfer using spatial variation of colours. Niu et al. [13] perform a three-step correction: a coarse-grain colour correction for global colour matching, a fine-grain colour correction to improve both global and local colour consistency, and a guided filtering process to ensure structural consistency.
Unlike nonparametric methods, parametric methods describe the relationship between the colours of two images as a global transformation. The seminal work of Reinhard et al. [14] presented a global colour transfer using the mean and standard deviation. They employed the l α β colour space in an attempt to untangle the different colour channels. Xiao and Ma [15] attempted to preserve the image gradient by leveraging the histogram of the target image and the gradient map of the source image. Nguyen et al. [16] performed a white-balance step on both source and target images to remove colour casts caused by different illuminations. They then corrected the global brightness difference by applying Xiao and Ma’s method [15] along the aligned white axis. He et al. [17] used semantic correspondences between the source and target images, leveraging neural representations and iteratively optimizing a local linear model to satisfy both local and global constraints. Wu et al. [18] segmented the images into regions of salient and nonsalient pixels using a weighted attention map. They then employed Reinhard et al.’s method [14] to correct these regions in the three channels of the YUV colour space. Lee at al. [19] proposed a deep neural network that leverages colour histogram analogy for colour transfer, modulating the colour of the source image with abstract colour features of the reference image. Even though there is no straightforward approach to the application of pairwise colour correction methods to the problem of harmonizing the colour of multiple images, it is evident that they have facilitated the development of colour consistency correction for image sets [20].

2.2. Colour Consistency Correction

Colour consistency correction is fundamental in multiview imaging systems and for applications such as 3D reconstruction, image mosaicking, or panoramic stitching. There are two predominant kinds of methodologies: those based on path propagation and those employing a global optimization. Path propagation approaches entail the selection of a reference image, and the progressive transfer of colour information to other images through a selected propagation path. Pan et al. [21] determined this path based on the dimension of the image overlap and time of transfer, while employing a global to local policy to improve the efficacy of the algorithm. Chen et al. [22] defined both the propagation path and reference image by minimizing the normalized error using the shortest distance algorithm. Dal’Col et al. [23] used the reprojection of geometric data to calculate accurate mappings between images and applied this information both on the computation of the shortest propagation path and in the calculation of the JIH colour mapping function. In these approaches, cumulative errors in the colour propagation can cause significant degradation of the results, particularly if there are pairs of images with low or incorrect mappings between them. The lack of a clear standard for the selection of a reference image is also a reasoned downside.
Global optimization methods solve for the parameters of all images simultaneously, minimizing the colour difference in corresponding regions. Each image is assigned an independent colour mapping model, and the colour consistency correction problem is formalized as a global energy optimization problem. The optimal solution of the energy function provides the final parameters of the image models, which correct the original images. Brown and Lowe [1] used a global gain compensation to minimize the colour difference in the overlapped regions. This approach was proposed for image stitching and has been implemented in the OpenCV stitching model. Xiong and Pulli [24] performed colour correction in the YCbCr colour space, applying gamma correction for the luminance component and linear correction for the chrominance components. Xia et al. [2] parameterized the colour remapping curve as a transform model, expressing the constraints of colour consistency, contrast and gradient in a uniform energy function. Yu et al. [25] combined global and local optimization strategies. They formalized the colour difference elimination as a least squares optimization problem, and mitigated local colour differences in the overlap areas of the target images with the gamma transform algorithm. Xie et al. [26] used a path propagation method to obtain a first guess for their global optimization approach. Keeping the distance between the corrected colour and the initial solution as a constraint on the optimization process, they attempted to avoid solutions that drift to unnatural appearance in the images. HaCohen et al. [4] used dense matching between images as colour correspondences to apply a global energy cost function. Their work focused on interactive colour correction of photo collections by automatically propagating changes made by a user to other images. Moulon et al. [27] employed the virtual line descriptor filter to augment the colour correspondences produced using feature matching and estimate global transformations for each colour channel independently, minimizing the colour difference through histogram matching. This method has been applied in OpenMVG. Shen et al. [28] proposed to accurately calculate overlapping regions by reprojecting a generated 3D mesh to the images. Park et al. [29] presented a robust low-rank matrix factorization method to estimate white balance and gamma correction parameters for each input image. Yang et al. [30] employed a spline curve remapping function, minimizing the variance in the colour values of each observation of the sparse points generated by structure from motion. Li et al. [20] proposed to group images using a graph partition algorithm. They performed intragroup correction and intergroup correction in sequence in an attempt to maintain accuracy and efficiency in the colour correction of large-scale images.

3. Proposed Approach

We propose a novel approach to tackle the problem of colour consistency correction, not by calculating mapping functions between pairs of images, nor by global optimization of camera models. Inspired by recent advances in neural implicit functions, such as DeepSDF [31], Fourier features [32], and NeRF [33], we represent a continuous scene as a 6D vector-valued function whose input is a 3D point location ( x p , y p , z p ) and a 3D camera position ( x c , y c , z c ) , both under world coordinates, and whose output is a colour ( r , g , b ) . We approximate this continuous 6D scene representation with an MLP network,
E Θ : ( x p , y p , z p , x c , y c , z c ) ( r , g , b )
and optimize its weights ( Θ ) to map from each input 6D coordinate to its corresponding colour, see Figure 1. The loss is a simple MSE error between each pixel in each image and the colour predicted by the network for the corresponding 3D point as seen by the specific camera,
L = c C p P c C ^ ( p , c ) C ( p , c ) 2 2
where C is the set of cameras of the dataset, P c is the subset of 3D points that are seen by camera c , C ^ ( p , c ) is the estimated colour, and C ( p , c ) is the reference colour for the pixel corresponding to the projection of point p in camera c .
However, as discussed in [32], deep networks are biased towards learning lower-frequency functions, making it impossible to capture the high-frequency detail of the textures in 3D space. Therefore, we map our input to a higher dimensional space using the positional encoding:
γ ( v ) = v , sin 2 0 π v , cos 2 0 π v , , sin 2 L 1 π v , cos 2 L 1 π v
where v corresponds to each individual component of input ( x p , y p , z p , x c , y c , z c ) . The maximum frequency, L, was set to 10 for the point location, ( x p , y p , z p ) , and 4 for the camera location, ( x c , y c , z c ) .
We follow a similar architecture to DeepSDF [31] and NeRF [33], including a skip connection that concatenates the 3D point position to the fifth layer’s activation, as this seems to improve learning; see Figure 2. Additionally, by concatenating information on camera position only towards the end of the network, we encourage interpolation of the colour between cameras, capture shared colour, and complement information on 3D space that is not seen from that camera more effectively. Unlike radiance field models, which rely on consistent colour information across views to synthesize realistic colour, we use the camera information to capture differences between acquisition conditions, so that we may generate colour for any 3D point as if it were seen by a particular camera. In those approaches, volumetric rendering has become common, but since our goal is efficient colour harmonization for images used in texture mapping of 3D mesh models, we choose instead to use the geometric information provided by RGB-D captures, mapping colour directly to the surface coordinates. This way we avoid ray-sampling and all the computation required to find empty space, which for indoor reconstruction is most of the scene.
For each capture, we estimate a colour mapping function that will transform the RGB from the image to a set of colours consistent with the implicit representation of the colour in 3D space. This is achieved by training one smaller MLP per image, which will approximate the transformation required to regenerate the pixels for a particular capture,
D Θ : ( r , g , b , x , y ) ( r ^ , g ^ , b ^ )
where x , y correspond to the col and row of the pixel with colour r , g , b in any given capture, and r ^ , g ^ , b ^ represent the estimated values for corrected colour, see Figure 3. Since these networks are mapping from one texture to another, they do not require positional encoding to elevate input to a higher dimensionality, see Figure 4. This approach makes our method robust to decimation of the depth data used to train the implicit scene representation. When regenerating pixels, these smaller MLPs will recover texture detail that may have been lost in the continuous scene representation, requiring only enough information to estimate a colour transfer from each capture to a corrected version that matches the colour learned by the MLP in the sampled points. The loss function for one MLP, learning the colour transformation for a capture, is
L = i I C ^ ( i ) C ( p , c ) 2 2
where I is the set of pixels in a capture, C ^ ( i ) is the estimated colour for pixel i , and C ( p , c ) is the reference colour for the pixel, obtained from the continuous scene representation E Θ , by providing the corresponding 3D point p and a reference camera c .

4. Results

In order to quantitatively assess colour consistency within the image sets, we adopt the commonly used PSNR and CIEDE2000 metrics. We calculate the colour similarity between images using the pairs of corresponding pixels, which we compute robustly using geometric data. Since both PSNR and CIEDE2000 are pairwise image metrics, we carry out the evaluation over each image pair in the set. The displayed score values include the mean and standard deviation for all image pairs. Additionally, we present a weighted mean based on the number of pixel correspondences.
CIEDE2000 [34] was adopted as the most recent colour difference metric from the International Commission on Illumination (CIE). This metric measures colour dissimilarity, meaning that lower score values equate to more similar colour measurements. The CIEDE2000 metric can be formulated as follows:
CIEDE 2000 = Δ L 2 + Δ C a b 2 + Δ H a b 2 + R T · Δ C a b · Δ H a b
where Δ L , Δ C a b , and Δ H a b are the lightness, chroma, and hue differences, respectively, and R T is a hue rotation term.
PSNR [35] measures colour similarity, meaning higher score values equate to more similar colour measurements. The PSNR formula is given by
PSNR = 10 · log 10 L 2 MSE
where L is the largest possible value in the dynamic range of an image, and MSE is the mean squared error between the two images.
We showcase the results of our approach in three distinct scenes from the publicly available Matterport3D (MP3D) [36] dataset. We perform comparisons with several well-known state-of-the-art approaches, providing them with accurate and robust mappings between pixels, calculated by the reprojection of geometric data to each image. For [19], no pixel mappings are required, as colour histogram analogy is employed. In the case of pairwise approaches [7,11,14], we compute a graph of the image set and weigh the edges using the number of pixels mappings between captures. The reference image is decided by selecting the node with the largest centrality. We then compute the propagation paths by applying the shortest path algorithm, as suggested in [22,23].
Because of the nature of 3D reconstruction datasets, particularly those employing stationary sensors, captures can be sparse, with some image pairs presenting insufficient pixel correspondences to accurately calculate colour transfer. This can deteriorate results and cause excessive errors, particularly when applied through propagation paths. Furthermore, for optimization approaches [1,27], using every possible pair of images, regardless of the usefulness of their information, can quickly become unfeasible due to the exponential nature of combinations. Experimentally, we found that setting a small threshold relative to the resolution of the geometric data to avoid using image pairs with negligible information was beneficial for both the efficacy and efficiency of the tested colour correction approaches. For the results presented in this section, we discarded image pairs connected by less than 1% of the available geometric data.
Pairwise methods [7,11,14] have problems with accumulated error, despite the filtration process that was applied. Optimization-based approaches [1,27] seem to have trouble balancing a large number of images, particularly from such different perspectives and with large colour differences. The local colour discrepancies also make it impossible for the camera models to correctly compensate for the effects in the captures. Results for the colour balancing using histogram analogy [19] are somewhat underwhelming; this may be attributed to the fact that this approach is intended for more significant colour transfer and is suitable for problems involving style transfer between pairs of images rather than precise harmonization in a set. Our approach, leveraging implicit representation that takes into account the 3D space in which the textures will be applied, achieved the best results in all experiments; see Table 1, Table 2 and Table 3. As shown in Figure 5, Figure 6 and Figure 7, the proposed approach effectively harmonizes textures from different captures, significantly mitigating the appearance of texture seams. It should be noted that, while pixel regeneration is highly effective in creating consistent textures across the environment, it can sometimes result in a loss of high-frequency detail in certain areas. Several factors influence the final output, including the quality of registration between captures, the precision of sampling during training, the maximum frequency of positional encoding, the number of epochs, and other hyperparameters. These factors can lead to variations in texture detail and colour accuracy. Depending on the conditions of the input data, there is an inherent trade-off between the harmonization of the captures and the level of high-frequency information for particular regions.

5. Conclusions

In this paper, we propose a novel neural-based method for colour consistency correction. We leverage an MLP to learn an implicit representation of the environment colour in 3D space, while capturing camera-dependent effects. We then use this network to train smaller MLPs, approximating the required transformations to regenerate pixels in the original captures, producing more consistent colour. Our method is efficient and offers better scalability than global optimization approaches, which are the current most popular approach. Experiments show that the proposed solution outperforms other relevant state-of-the-art methods for colour correction.
The algorithm was designed within the scope of a 3D reconstruction pipeline, as a way to increase the quality of the generated models, making use of the available registered scanning information. However, this can limit its use in other applications. Also, since the proposed approach thrives on thorough depth information, poor-quality data may undermine the effectiveness of the algorithm. Nevertheless, as depth sensors improve and become more widespread, we can expect our method’s performance to enhance.
As future work, it would be interesting to explore anomaly detection techniques to identify points in 3D space with high variability in view-dependent colour. This approach could help pinpoint surfaces likely to exhibit more reflective properties. This information could be used to render the materials more accurately through physically-based rendering (PBR) and further enhance the colour correction results.

Author Contributions

Conceptualization, T.M., P.D., and M.O.; funding acquisition, P.D. and M.O.; methodology, T.M., P.D., and M.O.; software, T.M.; writing—original draft, T.M.; writing—review and editing, P.D. and M.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been funded by National Funds through the FCT - Foundation for Science and Technology, in the context of the Ph.D. scholarship 2020.07345.BD, with DOI 10.54499/2020.07345.BD, and project UIDB/00127/2020, supported by IEETA.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We thank everyone involved for their time and expertise.

Conflicts of Interest

The authors declare no conflicts of interest.

Correction Statement

This article has been republished with a minor correction to resolve spelling and grammatical errors. This change does not affect the scientific content of the article.

References

  1. Brown, M.; Lowe, D. Automatic Panoramic Image Stitching using Invariant Features. Int. J. Comput. Vis. 2007, 74, 59–73. [Google Scholar] [CrossRef]
  2. Xia, M.; Yao, J.; Xie, R.; Zhang, M.; Xiao, J. Color Consistency Correction Based on Remapping Optimization for Image Stitching. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2017; pp. 2977–2984. [Google Scholar] [CrossRef]
  3. Li, L.; Li, Y.; Xia, M.; Li, Y.; Yao, J.; Wang, B. Grid Model-Based Global Color Correction for Multiple Image Mosaicking. IEEE Geosci. Remote Sens. Lett. 2021, 18, 2006–2010. [Google Scholar] [CrossRef]
  4. HaCohen, Y.; Shechtman, E.; Goldman, D.B.; Lischinski, D. Optimizing color consistency in photo collections. ACM Trans. Graph. 2013, 32, 1–10. [Google Scholar] [CrossRef]
  5. Pitie, F.; Kokaram, A.; Dahyot, R. N-dimensional probability density function transfer and its application to color transfer. In Proceedings of the Tenth IEEE International Conference on Computer Vision (ICCV’05), Washington, DC, USA, 17–20 October 2005; Volume 2, pp. 1434–1439. [Google Scholar] [CrossRef]
  6. Su, Z.; Deng, D.; Yang, X.; Luo, X. Color transfer based on multiscale gradient-aware decomposition and color distribution mapping. In Proceedings of the 20th ACM International Conference on Multimedia, Nara, Japan, 29 October–2 November 2012; pp. 753–756. [Google Scholar] [CrossRef]
  7. De Marchi, S. Polynomials arising in factoring generalized Vandermonde determinants: An algorithm for computing their coefficients. Math. Comput. Model. 2001, 34, 271–281. [Google Scholar] [CrossRef]
  8. Hwang, Y.; Lee, J.Y.; Kweon, I.S.; Kim, S.J. Color Transfer Using Probabilistic Moving Least Squares. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3342–3349. [Google Scholar] [CrossRef]
  9. Liu, X.; Zhu, L.; Xu, S.; Du, S. Palette-Based Recoloring of Natural Images Under Different Illumination. In Proceedings of the 2021 IEEE 6th International Conference on Computer and Communication Systems (ICCCS), Chengdu, China, 23–26 April 2021; pp. 347–351. [Google Scholar] [CrossRef]
  10. Wu, F.; Dong, W.; Kong, Y.; Mei, X.; Paul, J.C.; Zhang, X. Content-based colour transfer. Comput. Graph. Forum 2013, 32, 190–203. [Google Scholar] [CrossRef]
  11. Finlayson, G.D.; Mackiewicz, M.; Hurlbert, A. Color Correction Using Root-Polynomial Regression. IEEE Trans. Image Process. 2015, 24, 1460–1470. [Google Scholar] [CrossRef]
  12. Hwang, Y.; Lee, J.Y.; Kweon, I.S.; Kim, S.J. Probabilistic moving least squares with spatial constraints for nonlinear color transfer between images. Comput. Vis. Image Underst. 2019, 180, 1–12. [Google Scholar] [CrossRef]
  13. Niu, Y.; Zheng, X.; Zhao, T.; Chen, J. Visually Consistent Color Correction for Stereoscopic Images and Videos. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 697–710. [Google Scholar] [CrossRef]
  14. Reinhard, E.; Adhikhmin, M.; Gooch, B.; Shirley, P. Color transfer between images. IEEE Comput. Graph. Appl. 2001, 21, 34–41. [Google Scholar] [CrossRef]
  15. Xiao, X.; Ma, L. Gradient-Preserving Color Transfer. Comput. Graph. Forum 2009, 28, 1879–1886. [Google Scholar] [CrossRef]
  16. Nguyen, R.M.; Kim, S.J.; Brown, M.S. Illuminant aware gamut-based color transfer. Comput. Graph. Forum 2014, 33, 319–328. [Google Scholar] [CrossRef]
  17. He, M.; Liao, J.; Chen, D.; Yuan, L.; Sander, P.V. Progressive Color Transfer With Dense Semantic Correspondences. ACM Trans. Graph. 2019, 38, 1–18. [Google Scholar] [CrossRef]
  18. Wu, Z.; Xue, R. Color Transfer With Salient Features Mapping via Attention Maps Between Images. IEEE Access 2020, 8, 104884–104892. [Google Scholar] [CrossRef]
  19. Lee, J.; Son, H.; Lee, G.; Lee, J.; Cho, S.; Lee, S. Deep color transfer using histogram analogy. Vis. Comput. 2020, 36, 2129–2143. [Google Scholar] [CrossRef]
  20. Li, Y.; Li, Y.; Yao, J.; Gong, Y.; Li, L. Global Color Consistency Correction for Large-Scale Images in 3-D Reconstruction. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3074–3088. [Google Scholar] [CrossRef]
  21. Li, Y.; Fang, C.; Yang, J.; Wang, Z.; Lu, X.; Yang, M.H. Universal style transfer via feature transforms. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 385–395. [Google Scholar]
  22. Chen, C.; Chen, Z.; Li, M.; Liu, Y.; Cheng, L.; Ren, Y. Parallel relative radiometric normalisation for remote sensing image mosaics. Comput. Geosci. 2014, 73, 28–36. [Google Scholar] [CrossRef]
  23. Dal’Col, L.; Coelho, D.; Madeira, T.; Dias, P.; Oliveira, M. A Sequential Color Correction Approach for Texture Mapping of 3D Meshes. Sensors 2023, 23, 607. [Google Scholar] [CrossRef]
  24. Xiong, Y.; Pulli, K. Color matching of image sequences with combined gamma and linear corrections. In Proceedings of the International Conference on Multimedia, Firenze Italy, 25–29 October 2010; pp. 261–270. [Google Scholar]
  25. Yu, L.; Zhang, Y.; Sun, M.; Zhou, X.; Liu, C. An auto-adapting global-to-local color balancing method for optical imagery mosaic. ISPRS J. Photogramm. Remote Sens. 2017, 132, 1–19. [Google Scholar] [CrossRef]
  26. Xie, R.; Xia, M.; Yao, J.; Li, L. Guided color consistency optimization for image mosaicking. ISPRS J. Photogramm. Remote Sens. 2018, 135, 43–59. [Google Scholar] [CrossRef]
  27. Moulon, P.; Duisit, B.; Monasse, P. Global multiple-view color consistency. In Proceedings of the Conference on Visual Media Production, London, UK, 30 November–1 December 2013. [Google Scholar]
  28. Shen, T.; Wang, J.; Fang, T.; Zhu, S.; Quan, L. Color Correction for Image-Based Modeling in the Large. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016. [Google Scholar]
  29. Park, J.; Tai, Y.W.; Sinha, S.N.; Kweon, I.S. Efficient and Robust Color Consistency for Community Photo Collections. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 430–438. [Google Scholar] [CrossRef]
  30. Yang, J.; Liu, L.; Xu, J.; Wang, Y.; Deng, F. Efficient global color correction for large-scale multiple-view images in three-dimensional reconstruction. ISPRS J. Photogramm. Remote Sens. 2021, 173, 209–220. [Google Scholar] [CrossRef]
  31. Park, J.J.; Florence, P.; Straub, J.; Newcombe, R.; Lovegrove, S. Deepsdf: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 165–174. [Google Scholar]
  32. Tancik, M.; Srinivasan, P.P.; Mildenhall, B.; Fridovich-Keil, S.; Raghavan, N.; Singhal, U.; Ramamoorthi, R.; Barron, J.T.; Ng, R. Fourier features let networks learn high frequency functions in low dimensional domains. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Online, 6–12 December 2020. [Google Scholar]
  33. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. Nerf: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 2021, 65, 99–106. [Google Scholar] [CrossRef]
  34. Sharma, G.; Wu, W.; Dalal, E.N. The CIEDE2000 color-difference formula: Implementation notes, supplementary test data, and mathematical observations. Color Res. Appl. 2005, 30, 21–30. [Google Scholar] [CrossRef]
  35. Xu, W.; Mulligan, J. Performance evaluation of color correction approaches for automatic multi-view image and video stitching. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 263–270. [Google Scholar] [CrossRef]
  36. Chang, A.; Dai, A.; Funkhouser, T.; Halber, M.; Niessner, M.; Savva, M.; Song, S.; Zeng, A.; Zhang, Y. Matterport3D: Learning from RGB-D Data in Indoor Environments. In Proceedings of the International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017. [Google Scholar]
Figure 1. Overview of the training of our continuous scene representation using an MLP to approximate the colour in 3D space. Blue arrow represents a discrete sampling of the implicit function.
Figure 1. Overview of the training of our continuous scene representation using an MLP to approximate the colour in 3D space. Blue arrow represents a discrete sampling of the implicit function.
Sensors 24 04141 g001
Figure 2. Visualization of our neural network architecture for implicit scene representation. Input vectors are shown in green. The number inside each block signifies the vector’s dimension. All layers are standard fully-connected layers, blue arrows indicate layers with ReLU activations, dashed black arrows indicate layers with sigmoid activation, and “+” denotes vector concatenation.
Figure 2. Visualization of our neural network architecture for implicit scene representation. Input vectors are shown in green. The number inside each block signifies the vector’s dimension. All layers are standard fully-connected layers, blue arrows indicate layers with ReLU activations, dashed black arrows indicate layers with sigmoid activation, and “+” denotes vector concatenation.
Sensors 24 04141 g002
Figure 3. Overview of the training of our colour mapping function approximators, using the continuous scene representation as reference. The MLP previously trained to estimate colour in 3D space is used to provide ground truth for each of the smaller MLPs that will hold the approximation to regenerate pixels in the captures. Blue arrow represents the final output of the system.
Figure 3. Overview of the training of our colour mapping function approximators, using the continuous scene representation as reference. The MLP previously trained to estimate colour in 3D space is used to provide ground truth for each of the smaller MLPs that will hold the approximation to regenerate pixels in the captures. Blue arrow represents the final output of the system.
Sensors 24 04141 g003
Figure 4. Visualization of our neural network architecture for individual capture colour transformation. Input vectors are shown in green. The number inside each block signifies the vector’s dimension. All layers are standard fully-connected layers, blue arrows indicate layers with ReLU activations, dashed black arrows indicate layers with sigmoid activation.
Figure 4. Visualization of our neural network architecture for individual capture colour transformation. Input vectors are shown in green. The number inside each block signifies the vector’s dimension. All layers are standard fully-connected layers, blue arrows indicate layers with ReLU activations, dashed black arrows indicate layers with sigmoid activation.
Sensors 24 04141 g004
Figure 5. Qualitative comparison of our method with relevant state-of-the-art in scene 2azQ1b91cZZ_7 of the MP3D dataset. We observe that our method presents a more visually appealing mesh, with significantly reduced texture seams [1,7,11,14,19,27].
Figure 5. Qualitative comparison of our method with relevant state-of-the-art in scene 2azQ1b91cZZ_7 of the MP3D dataset. We observe that our method presents a more visually appealing mesh, with significantly reduced texture seams [1,7,11,14,19,27].
Sensors 24 04141 g005
Figure 6. Qualitative comparison of our method with relevant state-of-the-art in scene VLzqgDo317F_19 of the MP3D dataset. We observe that our method presents a more visually appealing mesh, with significantly reduced texture seams [1,7,11,14,19,27].
Figure 6. Qualitative comparison of our method with relevant state-of-the-art in scene VLzqgDo317F_19 of the MP3D dataset. We observe that our method presents a more visually appealing mesh, with significantly reduced texture seams [1,7,11,14,19,27].
Sensors 24 04141 g006
Figure 7. Qualitative comparison of our method with relevant state-of-the-art in scene aayBHfsNo7d_1 of the MP3D dataset. We observe that our method presents a more visually appealing mesh, with significantly reduced texture seams [1,7,11,14,19,27].
Figure 7. Qualitative comparison of our method with relevant state-of-the-art in scene aayBHfsNo7d_1 of the MP3D dataset. We observe that our method presents a more visually appealing mesh, with significantly reduced texture seams [1,7,11,14,19,27].
Sensors 24 04141 g007
Table 1. Evaluation of colour consistency correction in scene 2azQ1b91cZZ_7 of the MP3D dataset. We observe that our method improves in all metrics, highlighted in bold.
Table 1. Evaluation of colour consistency correction in scene 2azQ1b91cZZ_7 of the MP3D dataset. We observe that our method improves in all metrics, highlighted in bold.
   CIEDE2000    PSNR
Method μ μ w σ μ μ w σ
Original images10.8710.712.6618.4418.362.51
Reinhard et al. [14]10.5210.332.3918.3018.292.34
De Marchi et al. [7]8.969.144.0520.2019.924.27
Brown and Lowe [1]10.3810.152.2419.1019.072.15
Moulon et al. [27]6.496.411.3723.7323.712.19
Finlayson et al. [11]8.778.842.9320.1219.963.00
Lee et al. [19]11.3911.172.6617.9417.992.30
Ours3.363.450.7826.7726.542.66
Table 2. Evaluation of colour consistency correction in scene VLzqgDo317F_19 of the MP3D dataset. We observe that our method improves in all metrics, highlighted in bold.
Table 2. Evaluation of colour consistency correction in scene VLzqgDo317F_19 of the MP3D dataset. We observe that our method improves in all metrics, highlighted in bold.
   CIEDE2000    PSNR
Method μ μ w σ μ μ w σ
Original images15.2015.194.8515.5515.562.29
Reinhard et al. [14]10.9911.002.3717.8717.851.81
De Marchi et al. [7]12.4112.444.7917.5317.513.33
Brown and Lowe [1]13.5113.533.9716.7116.731.89
Moulon et al. [27]7.247.221.5221.5021.551.80
Finlayson et al. [11]11.1911.223.4617.8117.772.36
Lee et al. [19]12.1112.102.0917.3817.381.34
Ours3.593.571.2326.8326.862.79
Table 3. Evaluation of colour consistency correction in scene aayBHfsNo7d_1 of the MP3D dataset. We observe that our method improves in all metrics, highlighted in bold.
Table 3. Evaluation of colour consistency correction in scene aayBHfsNo7d_1 of the MP3D dataset. We observe that our method improves in all metrics, highlighted in bold.
   CIEDE2000    PSNR
Method μ μ w σ μ μ w σ
Original images9.749.564.3620.7420.624.85
Reinhard et al. [14]11.0710.873.3018.0017.993.16
De Marchi et al. [7]8.438.515.5822.0621.585.51
Brown and Lowe [1]8.218.363.3022.7022.215.16
Moulon et al. [27]7.327.743.6424.5723.826.34
Finlayson et al. [11]7.687.795.4822.8622.315.96
Lee et al. [19]13.0012.673.3816.4416.562.33
Ours2.492.551.2630.6030.046.38
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Madeira, T.; Oliveira, M.; Dias, P. Neural Colour Correction for Indoor 3D Reconstruction Using RGB-D Data. Sensors 2024, 24, 4141. https://doi.org/10.3390/s24134141

AMA Style

Madeira T, Oliveira M, Dias P. Neural Colour Correction for Indoor 3D Reconstruction Using RGB-D Data. Sensors. 2024; 24(13):4141. https://doi.org/10.3390/s24134141

Chicago/Turabian Style

Madeira, Tiago, Miguel Oliveira, and Paulo Dias. 2024. "Neural Colour Correction for Indoor 3D Reconstruction Using RGB-D Data" Sensors 24, no. 13: 4141. https://doi.org/10.3390/s24134141

APA Style

Madeira, T., Oliveira, M., & Dias, P. (2024). Neural Colour Correction for Indoor 3D Reconstruction Using RGB-D Data. Sensors, 24(13), 4141. https://doi.org/10.3390/s24134141

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop