Deep Learning Tone-Mapping and Demosaicing for Automotive Vision Systems
Abstract
:1. Introduction
2. Related Work
2.1. Tone Mapping
2.2. HDR Imaging and Object Detection
2.3. Demosaicing
2.4. Novelty of the Performed Research
2.5. ExpandNet
3. Proposed Algorithms
- -
- The devised architectures of the proposed CNNs, based on the architecture of the ExpandNet [68] network;
- -
- The concepts of the training and inference for the proposed CNNs;
- -
- The details of the training and the validation data sets;
- -
- The details of the augmentation procedures as part of the training methodology for the proposed CNNs.
3.1. On Simulating the Effects from the HDR Image Acquisition Processes
- -
- The acquired HDR training data (by applying the ExpandNet algorithm for inverse tone-mapping on an SDR data set consisting of traffic scenes annotated for traffic-related objects);
- -
- The existence of the CFA (by creating mosaics, i.e., Bayer CFA, of the input training images to the CNNs);
- -
- The presence of photon noise (by the application of noise with a Poisson distribution on the simulated mosaiced data);
- -
- The AD conversion process (by applying quantization with a predefined bit-depth parameter higher than 8 bits per channel);
- -
- The internal camera processing (by applying demosaicing with the use of a classical state-of-the-art algorithm [56]).
3.2. Proposed CNN for TM
3.2.1. Architecture
3.2.2. Training and Inference
3.3. Proposed CNN for Joint DM and TM
3.3.1. Architecture
3.3.2. Training and Inference
3.4. Training and Validation Data Sets
- -
- Is annotated for 10 object categories: car, traffic sign, traffic light, person, truck, bus, bike, rider, motor and train;
- -
- Consist of scenes with diverse weather conditions: rain, snow, sunlight, cloudy weather;
- -
- Consists of scenes captured at different times of day: nighttime, daytime, dawn/dusk;
- -
- Consists of scenes in various locales: city streets, residential areas, highways.
3.5. Training Methodology
- -
- Apply data augmentation (contrast and color augmentation) to account for different scenes;
- -
- Apply Bayer CFA mosaicing;
- -
- Simulate photon noise;
- -
- Simulate quantization;
- -
- For the proposed CNN for TM, perform demosaicing with a classical SOTA algorithm [56] to simulate the internal camera processing. This procedure is not applied during training of the proposed CNN for joint DM and TM.
3.5.1. Pre-Processing
3.5.2. Data Augmentation
Contrast Augmentation Procedures
Color Temperature Augmentation Procedures
3.5.3. CFA Mosaicing
3.5.4. Noise Application
3.5.5. Quantization
3.5.6. Demosaicing
3.5.7. Importance of the Applied Pipeline of Sequential Procedures Mosaicing-Noise Application–Quantization-Demosaicing
- -
- Mosaicing;
- -
- Application of noise with Poisson distribution;
- -
- Quantization;
- -
- Demosaicing.
- -
- For the proposed CNN for TM, we approach the problem of tone-mapping and noise and artifact suppression in a joint optimal way;
- -
- For the proposed CNN for joint DM and TM, we approach the problem of demosaicing, tone-mapping and noise and artifact suppression in a joint optimal way.
4. Evaluation
4.1. Test Data Set
- -
- From different times of day: nighttime, dawn/dusk, daytime;
- -
- With different lighting conditions: very dark scenes, very bright scenes and scenes with strong lights in the background;
- -
- From traffic jams and from lower frequency traffic;
- -
- Abundant with VRUs and objects crucial for road safety.
- -
- 1477 objects belonging to the class “person”;
- -
- 998 objects belonging to the class “traffic sign”;
- -
- 357 objects belonging to the class “colored traffic light”.
- -
- The time of day;
- -
- Whether there is sun glare or strong lights present in the scene;
- -
- Whether the captured scene is too bright, too dark, or the lighting conditions are normal;
- -
- Whether or not there is occlusion of the objects or the VRUs of interest;
- -
- Whether or not there is motion blur (caused by a fast change in the motion of the platform on which the cameras are positioned or because the VRUs are moving too fast) in the content visible to the annotator;
- -
- Existence of strong lights or sun glare in the scene;
- -
- It is a night scene;
- -
- Most of the relevant objects are occluded and hence hard to detect;
- -
- The scene is either too dark or too bright;
- -
- There is motion blur.
4.2. Methods for Comparison
4.3. Metrics and Methodology for Evaluation
4.3.1. Quantitative Analysis
- ∗
- The object detection evaluation is performed with using the following object detectors:
- -
- Yolo v3 [1] for the classes “person”, “stop sign” and “traffic light” (only detection on colored traffic lights is evaluated, as being the most relevant in tone-mapping and for road safety).
- -
- ∗
- The object detection performance is analyzed for each of the three classes: “person” (P), “colored traffic light” (TL) and “traffic sign” (TS). For the class “traffic sign” (TS), we consider the three subcategories of traffic signs in combination with the “stop sign”.
- ∗
- From the available metrics for object detection, we use:
- -
- F score: because it is a metric that combines precision and recall, while it penalizes the “missed detections” (False Negatives, i.e., FN) more than the false positives (FP) and unlike the symmetric F score, it gives more weight to the recall than to the precision;
- -
- True Positive Rate (TPR or recall), as a metric for correct detections;
- -
- False Negative Rate (FNR), as a metric for missed detections;
- -
- False Positives Per Image (FPPI): a metric that is calculated as the average number of false positives (FP) over all images in the test data set.
The F score, where , is shown in the following equations. In general, the F score, as a function of the number of True Positives (TP), number of False Positives (FP) and number of False Negatives (FN), is given by:When , F becomes F score:As a function of precision and recall, the F score in general can be calculated in the following manner: - ∗
- The object detection evaluation is performed in two ways:
- -
- -
- By applying the best performance object detection threshold (found from the F score vs. FPPI curves for the best performing algorithm for each object class) on the content obtained with each of the TM algorithms, as well as on the SDR content. Then, for the specific best performance object detection threshold for each object class, we measure the F score:
- -
- On the complete test data set;
- -
- On the split test data set in two categories: “difficult” and “easy” traffic scenes.
- ∗
- Additionally, to compare the complexity of the neural networks, for each of the used CNNs, we measure the number of parameters and the number of multiplication and addition (multiply-accumulate) operations on floating-point numbers (MACs).
4.3.2. Qualitative Analysis
- -
- Contrast;
- -
- Color appearance;
- -
- Tone-mapping of strong lights;
- -
- Presence of noise and artifacts.
- -
- Appearance of disturbing demosaicing artifacts;
- -
- Sharpness of the details;
- is analyzed by visual observation on parts of the content with sharp edges and/or repetitive structure.
5. Results
5.1. Results from the Quantitative Analysis
5.2. Results from the Qualitative Analysis
5.2.1. Qualitative Analysis of the Object Detection Results
- -
- The sequentially combined, SOTA demosaicing and TM algorithms;
- -
- The sequentially combined, SOTA demosaicing and the proposed CNN for TM;
- -
- The proposed CNN for joint DM and TM.
5.2.2. Qualitative Analysis of the Tone-Mapping and Demosaicing Results
- -
- In consistently obtaining overall high global and local contrast;
- -
- In obtaining high fidelity in color tone-mapping especially for traffic signs and traffic lights;
- -
- In performing localized tone-mapping of the strong lights (without producing “halo” artifacts in the surroundings);
- -
- In performing noise suppression;
- -
- In obtaining robustness to demosaicing and noise artifacts.
6. Conclusions and Future Work
- -
- Incorporating demosaicing in a neural network devised for tone-mapping. Based on the proposed CNN model for TM, we devise a CNN that performs joint demosaicing and tone-mapping of HDR content with noise suppression and good robustness on artifacts occurrence. The proposed CNN for joint DM and TM also has light-weight architecture and only slightly higher computational cost than the proposed CNN for TM.
- -
- Extensive evaluation with respect to ADS object detection accuracy of reconstructed (demosaiced and tone-mapped) content obtained with SOTA demosaicing and tone-mapping algorithms and the proposed CNNs. There is also a discussion and analysis on the quality of the reconstructed content by addressing the main aspects of high quality tone-mapping and demosaicing.
- -
- The content obtained with the proposed CNN models shows similar or distinctively better ADS object detection performance compared to the content obtained with the SOTA tone-mapping and demosaicing algorithms and also compared to the SDR content. Specifically, with the proposed CNNs, we succeed in improving the detectability of traffic-related objects and pedestrians over two existing fundamental cases, that of sequential demosaicing and tone-mapping of HDR data and only using the SDR content.
- -
- With the obtained similar computational cost between the proposed CNNs and the very similar results from the quantitative and the qualitative analysis, we confirm our second hypothesis.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
HDR | High dynamic range |
SDR | Standard dynamic range |
ADS | Automated driving systems |
AVS | Automotive vision systems |
CNN | Convolutional neural network |
TM | tone-mapping |
CNN for joint DM and TM | Convolutional neural network for joint demosaicing and tone-mapping |
CNN for TM | Convolutional neural network for tone-mapping |
DL HDR TM | Deep-learning high dynamic range tone-mapping |
LDR | Low dynamic range |
SOTA | state-of-the-art |
CFA | Color filter array |
AD | Analog to digital |
MACs | Multiply-accumulate operations |
PSNR | Peak Signal-to-Noise Ratio |
SSIM | Structural Similarity Index |
HVS | Human visual system |
FP | False Positives |
TP | True Positives |
TN | True Negatives |
FPPI | False Positives per Image |
References
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Shopovska, I.; Stojkovic, A.; Aelterman, J.; Van Hamme, D.; Philips, W. High-Dynamic-Range Tone Mapping in Intelligent Automotive Systems. Sensors 2023, 23, 5767. [Google Scholar] [CrossRef]
- Eilertsen, G.; Mantiuk, R.K.; Unger, J. A comparative review of tone-mapping algorithms for high dynamic range video. Comput. Graph. Forum 2017, 36, 565–592. [Google Scholar] [CrossRef]
- Eilertsen, G. The High Dynamic Range Imaging Pipeline; Linköping University Electronic Press: Linköping, Switzerland, 2018; Volume 1939. [Google Scholar]
- Banterle, F.; Artusi, A.; Debattista, K.; Chalmers, A. Advanced High Dynamic Range Imaging; CRC Press: Boca Raton, FL, USA; Taylor and Francis Group: Abingdon, UK, 2018. [Google Scholar]
- Darmont, A. High Dynamic Range Imaging: Sensors and Architectures, 2nd ed.; SPIE Press: Bellingham, WA, USA, 2019; Volume PM298. [Google Scholar]
- Dufaux, F.; Le Callet, P.; Mantiuk, R.; Mrak, M. High Dynamic Range Video: From Acquisition, to Display and Applications; Academic Press: London, UK, 2016. [Google Scholar]
- Reinhard, E.; Heidrich, W.; Debevec, P.; Pattanaik, S.; Ward, G.; Myszkowski, K. High Dynamic Range Imaging: Acquisition, Display, and Image-Based Lighting; Morgan Kaufmann: Burlington, MA, USA, 2010. [Google Scholar]
- Parraga, C.A.; Otazu, X. Which tone-mapping operator is the best? A comparative study of perceptual quality. JOSA A 2018, 35, 626–638. [Google Scholar]
- Ward, G. A contrast-based scalefactor for luminance display. Graph. Gems 1994, 4, 415–421. [Google Scholar]
- Tumblin, J.; Rushmeier, H. Tone reproduction for realistic images. IEEE Comput. Graph. Appl. 1993, 13, 42–48. [Google Scholar] [CrossRef]
- Drago, F.; Myszkowski, K.; Annen, T.; Chiba, N. Adaptive logarithmic mapping for displaying high contrast scenes. Comput. Graph. Forum 2003, 22, 419–426. [Google Scholar] [CrossRef]
- Reinhard, E.; Stark, M.; Shirley, P.; Ferwerda, J. Photographic tone reproduction for digital images. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, San Antonio, TX, USA, 23–26 July 2002; pp. 267–276. [Google Scholar]
- Reinhard, E.; Devlin, K. Dynamic range reduction inspired by photoreceptor physiology. IEEE Trans. Vis. Comput. Graph. 2005, 11, 13–24. [Google Scholar] [CrossRef]
- Ferwerda, J.A.; Pattanaik, S.N.; Shirley, P.; Greenberg, D.P. A model of visual adaptation for realistic image synthesis. In Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA, 4–9 August 1996; pp. 249–258. [Google Scholar]
- Ramsey, S.D.; Johnson, J.T., III; Hansen, C. Adaptive temporal tone mapping. In Proceedings of the 7th IASTED International Conference on Computer Graphics and Imaging, Kauai, HI, USA, 16–18 August 2004; pp. 124–128.
- Van Hateren, J.H. Encoding of high dynamic range video with a model of human cones. ACM Trans. Graph. (TOG) 2006, 25, 1380–1399. [Google Scholar] [CrossRef]
- Irawan, P.; Ferwerda, J.A.; Marschner, S.R. Perceptually Based Tone Mapping of High Dynamic Range Image Streams. In Proceedings of the Eurographics Symposium on Rendering, Konstanz, Germany, 29 June–1 July 2005; pp. 231–242. [Google Scholar]
- Mantiuk, R.; Daly, S.; Kerofsky, L. Display adaptive tone mapping. In Proceedings of the SIGGRAPH ’08: Special Interest Group on Computer Graphics and Interactive Techniques Conference, Los Angeles, CA, USA, 11–15 August 2008; pp. 1–10. [Google Scholar]
- Durand, F.; Dorsey, J. Fast bilateral filtering for the display of high-dynamic-range images. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, San Antonio, TX, USA, 23–26 July 2002; pp. 257–266. [Google Scholar]
- Fattal, R.; Lischinski, D.; Werman, M. Gradient domain high dynamic range compression. In Proceedings of the 29th Annual Conference on Computer Graphics and Interactive Techniques, San Antonio, TX, USA, 23–26 July 2002; pp. 249–256. [Google Scholar]
- Kuang, J.; Johnson, G.M.; Fairchild, M.D. iCAM06: A refined image appearance model for HDR image rendering. J. Vis. Commun. Image Represent. 2007, 18, 406–414. [Google Scholar] [CrossRef]
- Meylan, L.; Alleysson, D.; Süsstrunk, S. Model of retinal local adaptation for the tone mapping of color filter array images. JOSA A 2007, 24, 2807–2816. [Google Scholar] [CrossRef]
- Rahman, Z.U.; Jobson, D.J.; Woodell, G.A. Multiscale retinex for color rendition and dynamic range compression. In Proceedings of the Third IEEE International Conference on Image Processing (ICIP’96), Lausanne, Switzerland, 16–19 September 1996; Volume 2847, pp. 183–191. [Google Scholar]
- Pattanaik, S.N.; Ferwerda, J.A.; Fairchild, M.D.; Greenberg, D.P. A multiscale model of adaptation and spatial vision for realistic image display. In Proceedings of the 25th Annual Conference on Computer Graphics and Interactive Techniques, Orlando, FL, USA, 19–24 July 1998; pp. 287–298. [Google Scholar]
- Tumblin, J.; Turk, G. LCIS: A boundary hierarchy for detail-preserving contrast reduction. In Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 8–13 August 1999; pp. 83–90. [Google Scholar]
- Pattanaik, S.; Yee, H. Adaptive gain control for high dynamic range image display. In Proceedings of the 18th Spring Conference on Computer Graphics, Budmerice, Slovakia, 24–27 April 2002; pp. 83–87. [Google Scholar]
- Farbman, Z.; Fattal, R.; Lischinski, D.; Szeliski, R. Edge-preserving decompositions for multi-scale tone and detail manipulation. ACM Trans. Graph. (TOG) 2008, 27, 1–10. [Google Scholar] [CrossRef]
- Fattal, R. Edge-avoiding wavelets and their applications. ACM Trans. Graph. (TOG) 2009, 28, 1–10. [Google Scholar]
- Larson, G.W.; Rushmeier, H.; Piatko, C. A visibility matching tone reproduction operator for high dynamic range scenes. IEEE Trans. Vis. Comput. Graph. 1997, 3, 291–306. [Google Scholar] [CrossRef]
- Eilertsen, G.; Mantiuk, R.K.; Unger, J. Real-time noise-aware tone mapping. ACM Trans. Graph. (TOG) 2015, 34, 1–15. [Google Scholar] [CrossRef]
- Rana, A.; Singh, P.; Valenzise, G.; Dufaux, F.; Komodakis, N.; Smolic, A. Deep tone mapping operator for high dynamic range images. IEEE Trans. Image Process. 2019, 29, 1285–1298. [Google Scholar] [CrossRef] [PubMed]
- Le, C.; Yan, J.; Fang, Y.; Ma, K. Perceptually Optimized Deep High-Dynamic-Range Image Tone Mapping. arXiv 2021, arXiv:2109.00180. [Google Scholar]
- Vinker, Y.; Huberman-Spiegelglas, I.; Fattal, R. Unpaired Learning for High Dynamic Range Image Tone Mapping. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 14657–14666. [Google Scholar]
- Panetta, K.; Kezebou, L.; Oludare, V.; Agaian, S.; Xia, Z. Tmo-net: A parameter-free tone mapping operator using generative adversarial network, and performance benchmarking on large scale hdr dataset. IEEE Access 2021, 9, 39500–39517. [Google Scholar] [CrossRef]
- Wang, C.; Chen, B.; Seidel, H.P.; Myszkowski, K.; Serrano, A. Learning a self-supervised tone mapping operator via feature contrast masking loss. Comput. Graph. Forum 2022, 41, 71–84. [Google Scholar] [CrossRef]
- Zhang, J.; Wang, Y.; Tohidypour, H.; Pourazad, M.T.; Nasiopoulos, P. A Generative Adversarial Network Based Tone Mapping Operator for 4K HDR Images. In Proceedings of the 2023 International Conference on Computing, Networking and Communications (ICNC), Honolulu, HI, USA, 20–22 February 2023; pp. 473–477. [Google Scholar]
- Hu, L.; Chen, H.; Allebach, J.P. Joint Multi-Scale Tone Mapping and Denoising for HDR Image Enhancement. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2022; pp. 729–738. [Google Scholar]
- Mukherjee, R.; Melo, M.; Filipe, V.; Chalmers, A.; Bessa, M. Backward compatible object detection using HDR image content. IEEE Access 2020, 8, 142736–142746. [Google Scholar] [CrossRef]
- Mukherjee, R.; Bessa, M.; Melo-Pinto, P.; Chalmers, A. Object detection under challenging lighting conditions using high dynamic range imagery. IEEE Access 2021, 9, 77771–77783. [Google Scholar] [CrossRef]
- Onzon, E.; Mannan, F.; Heide, F. Neural auto-exposure for high-dynamic range object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7710–7720. [Google Scholar]
- Kocdemir, I.H.; Koz, A.; Akyuz, A.O.; Chalmers, A.; Alatan, A.; Kalkan, S. Tmo-Det: Deep Tone-Mapping Optimized with and for Object Detection. Pattern Recognit. Lett. 2023, 172, 230–236. [Google Scholar] [CrossRef]
- Bayer, B.E. Color Imaging Array. U.S. Patent 3,971,065, 20 July 1976. [Google Scholar]
- Hirakawa, K.; Wolfe, P.J. Spatio-Spectral Color Filter Array Design for Optimal Image Recovery. IEEE Trans. Image Process. 2008, 17, 1876–1890. [Google Scholar] [CrossRef]
- Quad-Bayer Camera Sensors For Better Photos. Available online: https://www.ubergizmo.com/articles/quad-bayer-camera-sensor/ (accessed on 24 August 2023).
- Sony Releases Stacked CMOS Image Sensor for Smartphones with Industry’s Highest 48 Effective Megapixels. Available online: https://www.sony.net/SonyInfo/News/Press/201807/18-060E/index.html (accessed on 24 August 2023).
- Lukac, R.; Plataniotis, K. Color filter arrays: Design and performance analysis. IEEE Trans. Consum. Electron. 2005, 51, 1260–1267. [Google Scholar] [CrossRef]
- Condat, L. A new random color filter array with good spectral properties. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 1613–1616. [Google Scholar]
- Gunturk, B.; Glotzbach, J.; Altunbasak, Y.; Schafer, R.; Mersereau, R. Demosaicking: Color filter array interpolation. IEEE Signal Process. Mag. 2005, 22, 44–54. [Google Scholar] [CrossRef]
- Li, X.; Gunturk, B.; Zhang, L. Image demosaicing: A systematic survey. In Proceedings of the Visual Communications and Image Processing 2008, San Jose, CA, USA, 29–31 January 2008; Volume 6822, pp. 489–503. [Google Scholar]
- Menon, D.; Calvagno, G. Color image demosaicking: An overview. Signal Process. Image Commun. 2011, 26, 518–533. [Google Scholar] [CrossRef]
- Wang, Y.; Cao, R.; Guan, Y.; Liu, T.; Yu, Z. A deep survey in the Applications of demosaicking. In Proceedings of the 2021 3rd International Academic Exchange Conference on Science and Technology Innovation (IAECST), Guangzhou, China, 10–12 December 2021; pp. 596–602. [Google Scholar]
- Kwan, C.; Larkin, J. Demosaicing of bayer and CFA 2.0 patterns for low lighting images. Electronics 2019, 8, 1444. [Google Scholar] [CrossRef]
- Alleysson, D.; Susstrunk, S.; Herault, J. Linear demosaicing inspired by the human visual system. IEEE Trans. Image Process. 2005, 14, 439–449. [Google Scholar] [CrossRef]
- Aelterman, J.; Goossens, B.; De Vylder, J.; Pižurica, A.; Philips, W. Computationally efficient locally adaptive demosaicing of color filter array images using the dual-tree complex wavelet packet transform. PLoS ONE 2013, 8, e61846. [Google Scholar] [CrossRef]
- Menon, D.; Andriani, S.; Calvagno, G. Demosaicing With Directional Filtering and a posteriori Decision. IEEE Trans. Image Process. 2007, 16, 132–141. [Google Scholar] [CrossRef]
- Monno, Y.; Kiku, D.; Tanaka, M.; Okutomi, M. Adaptive Residual Interpolation for Color and Multispectral Image Demosaicking. Sensors 2017, 17, 2787. [Google Scholar] [CrossRef]
- Monno, Y.; Kikuchi, S.; Tanaka, M.; Okutomi, M. A Practical One-Shot Multispectral Imaging System Using a Single Image Sensor. IEEE Trans. Image Process. 2015, 24, 3048–3059. [Google Scholar] [CrossRef] [PubMed]
- Zhang, C.; Li, Y.; Wang, J.; Hao, P. Universal Demosaicking of Color Filter Arrays. IEEE Trans. Image Process. 2016, 25, 5173–5186. [Google Scholar] [CrossRef] [PubMed]
- Cui, K.; Jin, Z.; Steinbach, E. Color Image Demosaicking Using a 3-Stage Convolutional Neural Network Structure. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 2177–2181. [Google Scholar] [CrossRef]
- Stojkovic, A.; Shopovska, I.; Luong, H.; Aelterman, J.; Jovanov, L.; Philips, W. The Effect of the Color Filter Array Layout Choice on State-of-the-Art Demosaicing. Sensors 2019, 19, 3215. [Google Scholar] [CrossRef] [PubMed]
- Park, Y.; Lee, S.; Jeong, B.; Yoon, J. Joint demosaicing and denoising based on a variational deep image prior neural network. Sensors 2020, 20, 2970. [Google Scholar] [CrossRef] [PubMed]
- Xu, Y.; Liu, Z.; Wu, X.; Chen, W.; Wen, C.; Li, Z. Deep joint demosaicing and high dynamic range imaging within a single shot. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4255–4270. [Google Scholar] [CrossRef]
- Guo, S.; Liang, Z.; Zhang, L. Joint denoising and demosaicking with green channel prior for real-world burst images. IEEE Trans. Image Process. 2021, 30, 6930–6942. [Google Scholar] [CrossRef]
- Zhang, T.; Fu, Y.; Li, C. Deep Spatial Adaptive Network for Real Image Demosaicing. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 3326–3334. [Google Scholar]
- Gharbi, M.; Chaurasia, G.; Paris, S.; Durand, F. Deep Joint Demosaicking and Denoising. ACM Trans. Graph. 2016, 35, 1–12. [Google Scholar] [CrossRef]
- Kokkinos, F.; Lefkimmiatis, S. Iterative Residual Network for Deep Joint Image Demosaicking and Denoising. arXiv 2018, arXiv:1807.06403. [Google Scholar]
- Marnerides, D.; Bashford-Rogers, T.; Hatchett, J.; Debattista, K. ExpandNet: A Deep Convolutional Neural Network for High Dynamic Range Expansion from Low Dynamic Range Content. Comput. Graph. Forum 2018, 37, 37–49. [Google Scholar] [CrossRef]
- Git Repository for the ExpandNet Algorithm. Available online: https://github.com/dmarnerides/hdr-expandnet (accessed on 24 August 2023).
- Yu, F.; Chen, H.; Wang, X.; Xian, W.; Chen, Y.; Liu, F.; Madhavan, V.; Darrell, T. Bdd100k: A diverse driving dataset for heterogeneous multitask learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2636–2645. [Google Scholar]
- BDD100K: A Large-scale Diverse Driving Video Database. Available online: https://bair.berkeley.edu/blog/2018/05/30/bdd/ (accessed on 24 August 2023).
- Git Repository for the Implementation of “Berkeley Driving Dataset 100K. BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning”. Available online: https://github.com/bdd100k/bdd100k (accessed on 24 August 2023).
- Saruchi, S. Adaptive sigmoid function to enhance low contrast images. Int. J. Comput. Appl. 2012, 55, 45–49. [Google Scholar] [CrossRef]
- Helland, T. Converting Temperature (Kelvin) to RGB: An Overview. 2012. Available online: https://tannerhelland.com/2012/09/18/convert-temperature-rgb-algorithm-code.html (accessed on 24 August 2023).
- Hirakawa, K.; Parks, T.W. Joint demosaicing and denoising. IEEE Trans. Image Process. 2006, 15, 2146–2157. [Google Scholar] [CrossRef]
- Liu, L.; Jia, X.; Liu, J.; Tian, Q. Joint demosaicing and denoising with self guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2240–2249. [Google Scholar]
- Git Repository for the “Unpaired Learning for High Dynamic Range Image Tone Mapping” Algorithm. Available online: https://github.com/yael-vinker/unpaired_hdr_tmo/ (accessed on 24 August 2023).
- Arcos-Garcia, A.; Alvarez-Garcia, J.A.; Soria-Morillo, L.M. Evaluation of deep neural networks for traffic sign detection systems. Neurocomputing 2018, 316, 332–344. [Google Scholar] [CrossRef]
- Git Repository for “Evaluation of Deep Neural Networks for Traffic Sign Detection Systems”. Available online: https://github.com/aarcosg/traffic-sign-detection (accessed on 24 August 2023).
Neural Network | Number of Parameters | MACs in G (Input Image Size: 1920 × 1080 × 3) |
---|---|---|
Proposed CNN for TM | 340,227 | 32,535 G |
Proposed CNN for joint DM and TM | 340,587 | 32,668 G |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Stojkovic, A.; Aelterman, J.; Van Hamme, D.; Shopovska, I.; Philips, W. Deep Learning Tone-Mapping and Demosaicing for Automotive Vision Systems. Sensors 2023, 23, 8507. https://doi.org/10.3390/s23208507
Stojkovic A, Aelterman J, Van Hamme D, Shopovska I, Philips W. Deep Learning Tone-Mapping and Demosaicing for Automotive Vision Systems. Sensors. 2023; 23(20):8507. https://doi.org/10.3390/s23208507
Chicago/Turabian StyleStojkovic, Ana, Jan Aelterman, David Van Hamme, Ivana Shopovska, and Wilfried Philips. 2023. "Deep Learning Tone-Mapping and Demosaicing for Automotive Vision Systems" Sensors 23, no. 20: 8507. https://doi.org/10.3390/s23208507
APA StyleStojkovic, A., Aelterman, J., Van Hamme, D., Shopovska, I., & Philips, W. (2023). Deep Learning Tone-Mapping and Demosaicing for Automotive Vision Systems. Sensors, 23(20), 8507. https://doi.org/10.3390/s23208507