Asymmetric Stereo High Dynamic Range Imaging with Smartphone Cameras
Abstract
:1. Introduction
- Difficulties with aligning images from asymmetric lenses at different exposures.
- Difficulties with temporal instability when capturing multiple images at different exposures from a single lens for HDR.
- Aligns differently exposed images from different lenses, each with different properties.
- Fuses the aligned LDR images into a single HDR image.
- Is accurate, robust and temporally stable, using only two heterogeneous images as inputs.
2. Existing Works
2.1. Multi-Modal Image Fusion
2.2. Depth Estimation
2.3. Optical Flow
2.4. Transformer Methods
2.5. Deep Learning
2.6. Mixed-Focal-Length Methods
2.7. Contributions of This Work
- No implementation exists for aligning differently exposed images from asymmetric lenses.
- No implementation exists for taking under-/over-exposed LDR images from asymmetric lenses and fusing these into an HDR image.
- Existing implementations for generating HDR images from multiple LDR images struggle with occlusions and temporal stability.
- A pipeline capable of taking differently exposed LDR images from asymmetric lenses and aligning them in preparation for HDR fusion.
- For the first time, showing a pipeline capable of fusing these aligned LDR images into a single HDR image, capable of gracefully dealing with occlusions and temporal instability.
3. Methodology
- Exposure Equalisation Module—where input images will be aligned to a similar exposure level.
- Warp-Alignment Module—where equalised images are warped and aligned to a similar plane.
- HDR Fusion Module—where aligned images are fused to produce an HDR result.
3.1. Solution Implementation
3.1.1. Exposure Equalisation
3.1.2. Warp Alignment
- Conduct flow-warping on the under-exposed ultra-wide image using the over-exposed wide-angle image as reference.
- Conduct a confidence map-informed blending operation to minimise flow-warping errors.
- Detect affine shape/keypoint pairs between the ultra-wide image and the homography-warped ultra-wide image.
- Use keypoints to warp the ultra-wide image via a homography transform.
- Repeat steps 1–2 with the ultra-wide image replaced with the affine-warped ultra-wide image.
3.1.3. HDR Fusion
3.2. Gathering Results
3.3. Comparing to Existing Solutions
- MEF with an unwarped ultra-wide image and a wide-angle image.
- MEF with two misaligned wide-angle images (to simulate motion).
- MEF with two aligned wide-angle images.
3.4. Evaluation Metrics
3.5. Degradation Measurements
3.5.1. High-Dynamic-Range Visual-Difference-Predictor 3
3.5.2. NoR-VDP-Net
3.6. Qualitative Testing
4. Results and Discussion
4.1. Warping
4.1.1. Affine Warp
4.1.2. Flow-Warp
4.2. HDR Generation
4.2.1. Ultra-Wide/Wide Fusion
4.2.2. Misaligned Fusion
4.2.3. Aligned Fusion
4.3. Performance
5. Conclusions and Future Work
5.1. Conclusions
5.2. Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Han, X.; Khan, I.R.; Rahardja, S. High Dynamic Range Image Tone Mapping: Literature review and performance benchmark. Digit. Signal Process. 2023, 137, 104015. [Google Scholar] [CrossRef]
- Park, W.J.; Ji, S.W.; Kang, S.J.; Jung, S.W.; Ko, S.J. Stereo vision-based high dynamic range imaging using differently-exposed image pair. Sensors 2017, 17, 1473. [Google Scholar] [CrossRef] [PubMed]
- Chari, P.; Vadathya, A.K.; Mitra, K. Optimal HDR and depth from dual cameras. arXiv 2020, arXiv:2003.05907. [Google Scholar]
- Wang, T.; Xie, J.; Sun, W.; Yan, Q.; Chen, Q. Dual-camera super-resolution with aligned attention modules. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 10–17 October 2021; pp. 2001–2010. [Google Scholar]
- Lee, J.; Lee, M.; Cho, S.; Lee, S. Reference-based video super-resolution using multi-camera video triplets. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17824–17833. [Google Scholar]
- Luo, X.; Peng, J.; Zhao, W.; Xian, K.; Lu, H.; Cao, Z. Point-and-Shoot All-in-Focus Photo Synthesis from Smartphone Camera Pair. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 2088–2101. [Google Scholar] [CrossRef]
- Teed, Z.; Deng, J. Raft: Recurrent all-pairs field transforms for optical flow. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part II 16. Springer: Cham, Switzerland, 2020; pp. 402–419. [Google Scholar]
- Truong, P.; Danelljan, M.; Van Gool, L.; Timofte, R. Learning accurate dense correspondences and when to trust them. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 5714–5724. [Google Scholar]
- Xu, H.; Ma, J.; Yuan, J.; Le, Z.; Liu, W. RFNet: Unsupervised Network for Mutually Reinforcing Multi-modal Image Registration and Fusion. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 19647–19656. [Google Scholar] [CrossRef]
- Peng, F.; Zhang, M.; Lai, S.; Tan, H.; Yan, S. Deep HDR reconstruction of dynamic scenes. In Proceedings of the 2018 IEEE 3rd International Conference on Image, Vision and Computing (ICIVC), Chongqing, China, 27–29 June 2018; pp. 347–351. [Google Scholar]
- Prabhakar, K.R.; Arora, R.; Swaminathan, A.; Singh, K.P.; Babu, R.V. A fast, scalable, and reliable deghosting method for extreme exposure fusion. In Proceedings of the 2019 IEEE International Conference on Computational Photography (ICCP), Tokyo, Japan, 15–17 May 2019; pp. 1–8. [Google Scholar]
- Qu, L.; Liu, S.; Wang, M.; Song, Z. TransMEF: A Transformer-Based Multi-Exposure Image Fusion Framework using Self-Supervised Multi-Task Learning. arXiv 2021, arXiv:2112.01030. [Google Scholar] [CrossRef]
- Eilertsen, G.; Kronander, J.; Denes, G.; Mantiuk, R.K.; Unger, J. HDR image reconstruction from a single exposure using deep CNNs. ACM Trans. Graph. (TOG) 2017, 36, 1–15. [Google Scholar] [CrossRef]
- Zhang, Q.; Nie, Y.; Zheng, W.-S. Dual Illumination Estimation for Robust Exposure Correction. arXiv 2019, arXiv:1910.13688. [Google Scholar] [CrossRef]
- Chen, Y.; Yu, M.; Chen, K.; Jiang, G.; Song, Y.; Peng, Z.; Chen, F. New stereo high dynamic range imaging method using generative adversarial networks. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 3502–3506. [Google Scholar]
- Dong, X.; Hu, X.; Li, W.; Wang, X.; Wang, Y. MIEHDR CNN: Main image enhancement based ghost-free high dynamic range imaging using dual-lens systems. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 1264–1272. [Google Scholar]
- Mishkin, D.; Radenovic, F.; Matas, J. Repeatability is not enough: Learning affine regions via discriminability. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 284–300. [Google Scholar]
- Derpanis, K.G. Overview of the RANSAC Algorithm. Image Rochester NY 2010, 4, 2–3. [Google Scholar]
- Reinhard, E.; Adhikhmin, M.; Gooch, B.; Shirley, P. Color transfer between images. IEEE Comput. Graph. Appl. 2001, 21, 34–41. [Google Scholar] [CrossRef]
- Hahne, C.; Aggoun, A. PlenoptiCam v1.0: A Light-Field Imaging Framework. IEEE Trans. Image Process. 2021, 30, 6757–6771. [Google Scholar] [CrossRef] [PubMed]
- Xu, H. FusedY2RGB. 2020. Available online: https://github.com/hanna-xu/utils/blob/master/fusedY2RGB/fusedY2RGB.m (accessed on 26 August 2024).
- Billa, A. ProCamera—Capture RAW Photos. 2023. Available online: https://www.procamera-app.com/en/testimonials/ (accessed on 26 August 2024).
- Li, H.; Ma, K.; Yong, H.; Zhang, L. Fast multi-scale structural patch decomposition for multi-exposure image fusion. IEEE Trans. Image Process. 2020, 29, 5805–5816. [Google Scholar] [CrossRef] [PubMed]
- Hayat, N.; Imran, M. Ghost-free multi exposure image fusion technique using dense SIFT descriptor and guided filter. J. Vis. Commun. Image Represent. 2019, 62, 295–308. [Google Scholar] [CrossRef]
- Karakaya, D.; Ulucan, O.; Turkan, M. PAS-MEF: Multi-exposure image fusion based on principal component analysis, adaptive well-exposedness and saliency map. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Virtual, 7–13 May 2022; pp. 2345–2349. [Google Scholar]
- Mertens, T.; Kautz, J.; Van Reeth, F. Exposure fusion: A simple and practical alternative to high dynamic range photography. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2009; Volume 28, pp. 161–171. [Google Scholar]
- Mantiuk, R.K.; Hammou, D.; Hanji, P. HDR-VDP-3: A multi-metric for predicting image differences, quality and contrast distortions in high dynamic range and regular content. arXiv 2023, arXiv:2304.13625. [Google Scholar]
- Banterle, F.; Artusi, A.; Moreo, A.; Carrara, F. Nor-Vdpnet: A No-Reference High Dynamic Range Quality Metric Trained on Hdr-Vdp 2. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Virtual, 25–28 October 2020. [Google Scholar]
Statistic | Flow-Warp | Warp Alignment | Fusion | Total Time |
---|---|---|---|---|
Mean | 8.491 | 12.54 | 18.16 | 35.89 |
Standard Deviation | 1.634 | 1.660 | 2.599 | 3.761 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Russell, F.; Midgley, W.J.B. Asymmetric Stereo High Dynamic Range Imaging with Smartphone Cameras. Sensors 2024, 24, 5876. https://doi.org/10.3390/s24185876
Russell F, Midgley WJB. Asymmetric Stereo High Dynamic Range Imaging with Smartphone Cameras. Sensors. 2024; 24(18):5876. https://doi.org/10.3390/s24185876
Chicago/Turabian StyleRussell, Finn, and William J. B. Midgley. 2024. "Asymmetric Stereo High Dynamic Range Imaging with Smartphone Cameras" Sensors 24, no. 18: 5876. https://doi.org/10.3390/s24185876
APA StyleRussell, F., & Midgley, W. J. B. (2024). Asymmetric Stereo High Dynamic Range Imaging with Smartphone Cameras. Sensors, 24(18), 5876. https://doi.org/10.3390/s24185876