Enhancing Visual Perception in Immersive VR and AR Environments: AI-Driven Color and Clarity Adjustments Under Dynamic Lighting Conditions
Abstract
:1. Introduction
- We develop a real-time AI model based on a modified U-Net architecture tailored for immersive media, capable of dynamic chromatic adjustments. Our approach is comprehensively evaluated against traditional image processing techniques and state-of-the-art color constancy algorithms.
- We conduct extensive experiments using subjective user evaluations and objective eye-tracking data to assess performance. The results show up to 41% improvement in color accuracy under low-light conditions, significantly outperforming traditional methods.
- We demonstrate the practical benefits of our AI-based system in enhancing image quality and reducing visual discomfort in VR/AR environments.
- Section 2: Related Work reviews research on image quality in immersive environments and advancements in AI-based image processing.
- Section 3: Methodology outlines the AI model architecture, experimental setup, and data collection process.
- Section 4: Experimental Results presents and analyzes user study data, comparing AI-enhanced and non-enhanced images under varying lighting conditions.
- Section 5: Discussion interprets the findings, emphasizing the improvements introduced by the AI-based system and discussing implications for VR/AR applications.
2. Related Work
2.1. Challenges and Techniques in Image Quality for VR/AR Environments
2.2. Limitations of Traditional Image Processing Techniques and Proposed AI Solutions
2.3. Research Gap and Our Contribution
3. Methodology
- Training settings and hyperparameters:
3.1. Dataset Preparation and Augmentation
3.2. Training and Evaluation Procedures
3.3. Experimental Setup and Data Collection
3.4. Evaluation Metrics
- Quantitative metrics: Mean absolute error (MAE) and structural similarity index (SSIM) were used to measure the model’s performance in terms of color accuracy and image clarity.
- Qualitative metrics: Subjective user ratings of image clarity, color accuracy, and visual appeal were collected to assess the user experience in immersive environments.
3.5. Ethical Considerations and Data Collection
- Color accuracy: How would you rate the color accuracy of the images you viewed on a scale from 1 to 5 (1 being poor, 5 being excellent)?
- Image clarity: How clear did the images appear to you on a scale from 1 to 5 (1 being not clear at all, 5 being very clear)?
- Overall visual appeal: How visually appealing did you find the images on a scale from 1 to 5?
- Fixation duration perception: Did you find yourself looking at certain parts of the images for an extended period of time? (Yes/No)
- Saccade exploration perception: Did you feel like you needed to move your eyes around a lot to explore the image? (1 being not at all, 5 being very much so)
- Discomfort or eye strain: Did you experience any discomfort or eye strain while viewing the images? (Yes/No)
- Engagement level: How engaged did you feel while interacting with the images? (1 being not engaged at all, 5 being fully engaged)
- Impact of lighting conditions: Did the lighting conditions (low, medium, high light) affect your viewing experience? (Yes/No). If yes, please specify how it affected your experience.
- Preference for AI-enhanced images: Do you prefer the AI-enhanced images over the non-enhanced images? (Yes/No)
- Recommendation: Would you recommend this type of immersive visual experience to others? (Yes/No)
4. Experimental Results
4.1. Visual Comparisons and Subjective Results
4.2. Detailed Analysis of Color Accuracy and Image Clarity
4.3. Objective Metrics: Fixation Duration and Saccade Patterns
4.4. Statistical and Computational Performance
4.5. Conclusions
5. Discussion and Implications
5.1. Interpretation of Findings
5.2. Comparison with Related Work and Practical Implications
5.3. Theoretical Contributions and Ethical Considerations
5.4. Limitations and Future Directions
5.5. Conclusion and Recommendations for Practitioners
- Adopt AI-based solutions: AI-driven chromatic adjustments can enhance user engagement and satisfaction in immersive environments.
- Optimize for real-time performance: Ensure that AI models are optimized for real-time processing to maintain user immersion without latency issues.
- User-centered design: Engage users in the testing process to refine visual enhancements and address issues such as over-saturation.
- Prioritize accessibility: Ensure that AI-enhanced systems consider the needs of users with visual impairments, enabling more inclusive immersive experiences.
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Lee, J.; Lee, S.; Kim, Y.C.; Kim, S.; Hong, S. Augmented virtual reality and 360 spatial visualization for supporting user-engaged design. J. Comput. Des. Eng. 2023, 10, 1047–1059. [Google Scholar] [CrossRef]
- Garner, J.J.; D’Zmura, M. Impact of low visual contrast on posture, motion sickness, and performance in VR. In Proceedings of the SPIE Defense + Commercial Sensing, Online, 12–17 April 2021; p. 117590A. [Google Scholar] [CrossRef]
- Chan, T.; Wang, Y.; So, R.H.; Jia, J. Predicting Subjective Discomfort Associated with Lens Distortion in VR Headsets During Vestibulo-Ocular Response to VR Scenes. IEEE Trans. Vis. Comput. Graph. 2022, 29, 3656–3669. [Google Scholar] [CrossRef] [PubMed]
- Ganj, A.; Zhao, Y.; Su, H.; Guo, T. Mobile AR Depth Estimation: Challenges & Prospects—Extended Version. arXiv 2023, arXiv:2310.14437. [Google Scholar] [CrossRef]
- Shekhar, S.; Reimann, M.; Semmo, A.; Pasewaldt, S.; Dollner, J.; Trapp, M. Low-light Image and Video Enhancement via Selective Manipulation of Chromaticity. arXiv 2022, arXiv:2203.04889. [Google Scholar] [CrossRef]
- Forsyth, D. A novel algorithm for color constancy. Int. J. Comput. Vis. 1990, 5, 5–35. [Google Scholar] [CrossRef]
- Zamir, S.W.; Arora, A.; Khan, S.H.; Hayat, M.; Khan, F.; Yang, M.H.; Shao, L. Learning Enriched Features for Fast Image Restoration and Enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1934–1948. [Google Scholar] [CrossRef]
- Bianco, S.; Cusano, C.; Schettini, R. Single and Multiple Illuminant Estimation Using Convolutional Neural Networks. IEEE Trans. Image Process. 2015, 26, 4347–4362. [Google Scholar] [CrossRef]
- Lecca, M. STAR: A Segmentation-Based Approximation of Point-Based Sampling Milano Retinex for Color Image Enhancement. IEEE Trans. Image Process. 2018, 27, 5802–5812. [Google Scholar] [CrossRef]
- Kim, H.; Lee, I.K. Studying the Effects of Congruence of Auditory and Visual Stimuli on Virtual Reality Experiences. IEEE Trans. Vis. Comput. Graph. 2022, 28, 2080–2090. [Google Scholar] [CrossRef]
- Yang, E.; Park, S.; Ryu, J. The effects of physical fidelity and task repetition on perceived task load and performance in the virtual reality-based training simulation. Br. J. Educ. Technol. 2023, 55, 1507–1527. [Google Scholar] [CrossRef]
- Abd-Alhamid, F.; Kent, M.; Bennett, C.R.; Calautit, J.; Wu, Y. Developing an Innovative Method for Visual Perception Evaluation in a Physical-Based Virtual Environment. Build. Environ. 2019, 162, 106278. [Google Scholar] [CrossRef]
- Ang, S.Y.; Ng, G. A Study on the Effect of the Real-world Backgrounds using Colour Blending Technique on Optical See-through AR User Interface Design. In Proceedings of the 2020 6th International Conference on Interactive Digital Media (ICIDM), Bandung, Indonesia, 14–15 December 2020; pp. 1–4. [Google Scholar] [CrossRef]
- Matsui, T.; Ikehara, M. Low-Light Image Enhancement Using a Simple Network Structure. IEEE Access 2023, 11, 65507–65516. [Google Scholar] [CrossRef]
- Shahidi, F. Breast Cancer Histopathology Image Super-Resolution Using Wide-Attention GAN with Improved Wasserstein Gradient Penalty and Perceptual Loss. IEEE Access 2021, 9, 32795–32809. [Google Scholar] [CrossRef]
- Quan, Y.; Chen, Y.; Shao, Y.; Teng, H.; Xu, Y.; Ji, H. Image denoising using complex-valued deep CNN. Pattern Recognit. 2021, 111, 107639. [Google Scholar] [CrossRef]
- Gharbi, M.; Chen, J.; Barron, J.; Hasinoff, S.; Durand, F. Deep bilateral learning for real-time image enhancement. ACM Trans. Graph. (Tog) 2017, 36, 118. [Google Scholar] [CrossRef]
- Ravendran, A.; Bryson, M.; Dansereau, D. Burst Imaging for Light-Constrained Structure-From-Motion. IEEE Robot. Autom. Lett. 2021, 7, 1040–1047. [Google Scholar] [CrossRef]
- Sun, Y.; Zhao, Z.; Jiang, D.; Tong, X.; Tao, B.; Jiang, G.; Kong, J.; Yun, J.; Liu, Y.; Liu, X.; et al. Low-Illumination Image Enhancement Algorithm Based on Improved Multi-Scale Retinex and ABC Algorithm Optimization. Front. Bioeng. Biotechnol. 2022, 10, 865820. [Google Scholar] [CrossRef]
- Li, J.; Li, J.; Fang, F.; Li, F.; Zhang, G. Luminance-Aware Pyramid Network for Low-Light Image Enhancement. IEEE Trans. Multimed. 2021, 23, 3153–3165. [Google Scholar] [CrossRef]
- Liu, C.; Wang, L.; Li, Z.; Quan, S.; Xu, Y. Real-Time Lighting Estimation for Augmented Reality via Differentiable Screen-Space Rendering. IEEE Trans. Vis. Comput. Graph. 2022, 29, 2132–2145. [Google Scholar] [CrossRef]
- Ibtehaz, N.; Rahman, M.S. MultiResUNet: Rethinking the U-Net Architecture for Multimodal Biomedical Image Segmentation. Neural Netw. Off. J. Int. Neural Netw. Soc. 2019, 121, 74–87. [Google Scholar] [CrossRef] [PubMed]
- Li, C.; Guo, C.; Han, L.; Jiang, J.; Cheng, M.M.; Gu, J.; Loy, C.C. Low-Light Image and Video Enhancement Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9396–9416. [Google Scholar] [CrossRef] [PubMed]
- Zhang, R.; Zhu, J.Y.; Isola, P.; Geng, X.; Lin, A.S.; Yu, T.; Efros, A.A. Real-time user-guided image colorization with learned deep priors. ACM Trans. Graph. (TOG) 2017, 36, 119. [Google Scholar] [CrossRef]
- McMahan, R.P.; Bowman, D.; Zielinski, D.J.; Brady, R. Evaluating Display Fidelity and Interaction Fidelity in a Virtual Reality Game. IEEE Trans. Vis. Comput. Graph. 2012, 18, 626–633. [Google Scholar] [CrossRef] [PubMed]
- Nie, G.; Duh, H.; Liu, Y.; Wang, Y. Analysis on Mitigation of Visually Induced Motion Sickness by Applying Dynamical Blurring on a User’s Retina. IEEE Trans. Vis. Comput. Graph. 2020, 26, 2535–2545. [Google Scholar] [CrossRef]
- Rallabandi, S.; Kakodkar, I.G.S.; Avuku, O. Ethical U se of AI in Social Media. In Proceedings of the 2023 International Workshop on Intelligent Systems (IWIS), Ulsan, Republic of Korea, 9–11 August 2023; pp. 1–9. [Google Scholar] [CrossRef]
- Bhagat, S.; Joshi, P.; Agarwal, A.; Gupta, S. Accessibility evaluation of major assistive mobile applications available for the visually impaired. ITU J. Future Evol. Technol. 2023, 4, 631–643. [Google Scholar] [CrossRef]
Metric | AI-Enhanced | HE | GC | Non-Enhanced |
---|---|---|---|---|
Color Accuracy | ||||
Image Clarity | ||||
Overall Visual Appeal |
Lighting Condition | AI-Enhanced | HE | GC | Non-Enhanced |
---|---|---|---|---|
Low Light (200 lux) | ||||
Medium Light (500 lux) | ||||
High Light (1000 lux) |
Metric | AI-Enhanced | HE | GC | Non-Enhanced |
---|---|---|---|---|
Saccade amplitude (degrees) | ||||
Saccade frequency (per image) |
Method | Processing Time (ms) |
---|---|
Our AI Model | |
Histogram Equalization | |
Gamma Correction | |
CCC | |
STAR |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Abbasi, M.; Váz, P.; Silva, J.; Martins, P. Enhancing Visual Perception in Immersive VR and AR Environments: AI-Driven Color and Clarity Adjustments Under Dynamic Lighting Conditions. Technologies 2024, 12, 216. https://doi.org/10.3390/technologies12110216
Abbasi M, Váz P, Silva J, Martins P. Enhancing Visual Perception in Immersive VR and AR Environments: AI-Driven Color and Clarity Adjustments Under Dynamic Lighting Conditions. Technologies. 2024; 12(11):216. https://doi.org/10.3390/technologies12110216
Chicago/Turabian StyleAbbasi, Maryam, Paulo Váz, José Silva, and Pedro Martins. 2024. "Enhancing Visual Perception in Immersive VR and AR Environments: AI-Driven Color and Clarity Adjustments Under Dynamic Lighting Conditions" Technologies 12, no. 11: 216. https://doi.org/10.3390/technologies12110216
APA StyleAbbasi, M., Váz, P., Silva, J., & Martins, P. (2024). Enhancing Visual Perception in Immersive VR and AR Environments: AI-Driven Color and Clarity Adjustments Under Dynamic Lighting Conditions. Technologies, 12(11), 216. https://doi.org/10.3390/technologies12110216