sensors-logo

Journal Browser

Journal Browser

Advances in AI-Based Processing of Image and Video Data Acquired by Various Sensors

A special issue of Sensors (ISSN 1424-8220). This special issue belongs to the section "Sensing and Imaging".

Deadline for manuscript submissions: 25 January 2025 | Viewed by 7873

Special Issue Editors


E-Mail Website
Guest Editor
School of Electrical and Computer Engineering, Ben Gurion University of the Negev, Be’er-Sheva 84105001, Israel
Interests: image and video processing/compression; deep learning in various emerging applications in computer vision
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
School of Electrical and Computer Engineering , Ben Gurion University of the Negev, Be’er-Sheva 84105001, Israel
Interests: image and video correction and analysis
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

Different types of imaging sensors produce a large variety of images, covering different electromagnetic regimes and various characteristics of the imaged objects. The field of artificial intelligence (AI) has made significant strides in recent years, particularly in the areas of image and video processing. With the proliferation of high-resolution imaging sensors and the ever-increasing amounts of visual data being generated by them, AI-based techniques have become indispensable for efficient and accurate image and video processing and analysis.

This Special Issue aims to bring together researchers and practitioners from academia and industry to showcase the latest advancements in AI-based image- and video-processing technology. We welcome original research papers, review articles, and short communications on topics including but not limited to:

  • Deep-learning techniques for image and video processing
  • Computer vision and pattern recognition
  • Object detection and tracking in video streams
  • Semantic segmentation and image classification
  • Image and video restoration and enhancement
  • Generative models for image and video synthesis
  • Multimodal data fusion and analysis
  • Applications of AI-based image and video processing in various domains
  • Super-resolution techniques based on AI
  • Image and video compression based on AI
  • Medical image processing with AI

We encourage submissions that showcase novel techniques and applications, as well as contributions that demonstrate the practicality and scalability of AI-based image- and video-processing solutions. All papers will undergo a rigorous peer-review process to ensure high quality and relevance to the theme of this Special Issue.

Dr. Ofer Hadar
Dr. Yitzhak Yitzhaky
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Sensors is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • DL in computer vision
  • AI in computer vision
  • image and video processing with AI

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (7 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

16 pages, 495 KiB  
Article
Reduction of Vision-Based Models for Fall Detection
by Asier Garmendia-Orbegozo, Miguel Angel Anton and Jose David Nuñez-Gonzalez
Sensors 2024, 24(22), 7256; https://doi.org/10.3390/s24227256 - 13 Nov 2024
Viewed by 280
Abstract
Due to the limitations that falls have on humans, early detection of these becomes essential to avoid further damage. In many applications, various technologies are used to acquire accurate information from individuals such as wearable sensors, environmental sensors or cameras, but all of [...] Read more.
Due to the limitations that falls have on humans, early detection of these becomes essential to avoid further damage. In many applications, various technologies are used to acquire accurate information from individuals such as wearable sensors, environmental sensors or cameras, but all of these require high computational resources in many cases, delaying the response of the entire system. The complexity of the models used to process the input data and detect these activities makes them almost impossible to complete on devices with limited resources, which are the ones that could offer an immediate response avoiding unnecessary communications between sensors and centralized computing centers. In this work, we chose to reduce the models to detect falls using images as input data. We proceeded to use image sequences as video frames, using data from two open source datasets, and we applied the Sparse Low Rank Method to reduce certain layers of the Convolutional Neural Networks that were the backbone of the models. Additionally, we chose to replace a convolutional block with Long Short Term Memory to consider the latest updates of these data sequences. The results showed that performance was maintained decently while significantly reducing the parameter size of the resulting models. Full article
Show Figures

Figure 1

19 pages, 5196 KiB  
Article
Bilateral Defect Cutting Strategy for Sawn Timber Based on Artificial Intelligence Defect Detection Model
by Chenlong Fan, Zilong Zhuang, Ying Liu, Yutu Yang, Haiyan Zhou and Xu Wang
Sensors 2024, 24(20), 6697; https://doi.org/10.3390/s24206697 - 18 Oct 2024
Viewed by 563
Abstract
Solid wood is renowned as a superior material for construction and furniture applications. However, characteristics such as dead knots, live knots, piths, and cracks are easily formed during timber’s growth and processing stages. These features and defects significantly undermine the mechanical characteristics of [...] Read more.
Solid wood is renowned as a superior material for construction and furniture applications. However, characteristics such as dead knots, live knots, piths, and cracks are easily formed during timber’s growth and processing stages. These features and defects significantly undermine the mechanical characteristics of sawn timber, rendering it unsuitable for specific applications. This study introduces BDCS-YOLO (Bilateral Defect Cutting Strategy based on You Only Look Once), an artificial intelligence bilateral sawing strategy to advance the automation of timber processing. Grounded on a dual-sided image acquisition platform, BDCS-YOLO achieves a commendable mean average feature detection precision of 0.94 when evaluated on a meticulously curated dataset comprising 450 images. Furthermore, a dual-side processing optimization module is deployed to enhance the accuracy of defect detection bounding boxes and establish refined processing coordinates. This innovative approach yields a notable 12.3% increase in the volume yield of sawn timber compared to present production, signifying a substantial leap toward efficiently utilizing solid wood resources in the lumber processing industry. Full article
Show Figures

Figure 1

20 pages, 56573 KiB  
Article
Asymmetric Stereo High Dynamic Range Imaging with Smartphone Cameras
by Finn Russell and William J. B. Midgley
Sensors 2024, 24(18), 5876; https://doi.org/10.3390/s24185876 - 10 Sep 2024
Viewed by 875
Abstract
Stereo high dynamic range imaging (SHDRI) offers a more temporally stable solution to high dynamic range (HDR) imaging from low dynamic range input images compared to bracketing and removes the loss of accuracy that single-image HDR solutions offer. However, few solutions currently exist [...] Read more.
Stereo high dynamic range imaging (SHDRI) offers a more temporally stable solution to high dynamic range (HDR) imaging from low dynamic range input images compared to bracketing and removes the loss of accuracy that single-image HDR solutions offer. However, few solutions currently exist that take advantage of the different (asymmetric) lenses, commonly found on modern smartphones, to achieve SHDRI. This paper presents a method that achieves single-shot asymmetric HDR fusion via a reference-based deep learning approach. Results demonstrate a system that is more robust to aperture and image signal processing pipeline differences than existing solutions. Full article
Show Figures

Figure 1

14 pages, 2815 KiB  
Article
Classification of Fashion Models’ Walking Styles Using Publicly Available Data, Pose Detection Technology, and Multivariate Analysis: From Past to Current Trendy Walking Styles
by Yoshiyuki Kobayashi, Sakiko Saito and Tatsuya Murahori
Sensors 2024, 24(12), 3865; https://doi.org/10.3390/s24123865 - 14 Jun 2024
Cited by 1 | Viewed by 987
Abstract
Understanding past and current trends is crucial in the fashion industry to forecast future market demands. This study quantifies and reports the characteristics of the trendy walking styles of fashion models during real-world runway performances using three cutting-edge technologies: (a) publicly available video [...] Read more.
Understanding past and current trends is crucial in the fashion industry to forecast future market demands. This study quantifies and reports the characteristics of the trendy walking styles of fashion models during real-world runway performances using three cutting-edge technologies: (a) publicly available video resources, (b) human pose detection technology, and (c) multivariate human-movement analysis techniques. The skeletal coordinates of the whole body during one gait cycle, extracted from publicly available video resources of 69 fashion models, underwent principal component analysis to reduce the dimensionality of the data. Then, hierarchical cluster analysis was used to classify the data. The results revealed that (1) the gaits of the fashion models analyzed in this study could be classified into five clusters, (2) there were significant differences in the median years in which the shows were held between the clusters, and (3) reconstructed stick-figure animations representing the walking styles of each cluster indicate that an exaggerated leg-crossing gait has become less common over recent years. Accordingly, we concluded that the level of leg crossing while walking is one of the major changes in trendy walking styles, from the past to the present, directed by the world’s leading brands. Full article
Show Figures

Figure 1

23 pages, 10188 KiB  
Article
Optimized OTSU Segmentation Algorithm-Based Temperature Feature Extraction Method for Infrared Images of Electrical Equipment
by Xueli Liu, Zhanlong Zhang, Yuefeng Hao, Hui Zhao and Yu Yang
Sensors 2024, 24(4), 1126; https://doi.org/10.3390/s24041126 - 8 Feb 2024
Cited by 4 | Viewed by 1325
Abstract
Infrared image processing is an effective method for diagnosing faults in electrical equipment, in which target device segmentation and temperature feature extraction are key steps. Target device segmentation separates the device to be diagnosed from the image, while temperature feature extraction analyzes whether [...] Read more.
Infrared image processing is an effective method for diagnosing faults in electrical equipment, in which target device segmentation and temperature feature extraction are key steps. Target device segmentation separates the device to be diagnosed from the image, while temperature feature extraction analyzes whether the device is overheating and has potential faults. However, the segmentation of infrared images of electrical equipment is slow due to issues such as high computational complexity, and the temperature information extracted lacks accuracy due to the insufficient consideration of the non-linear relationship between the image grayscale and temperature. Therefore, in this study, we propose an optimized maximum between-class variance thresholding method (OTSU) segmentation algorithm based on the Gray Wolf Optimization (GWO) algorithm, which accelerates the segmentation speed by optimizing the threshold determination process using OTSU. The experimental results show that compared to the non-optimized method, the optimized segmentation method increases the threshold calculation time by more than 83.99% while maintaining similar segmentation results. Based on this, to address the issue of insufficient accuracy in temperature feature extraction, we propose a temperature value extraction method for infrared images based on the K-nearest neighbor (KNN) algorithm. The experimental results demonstrate that compared to traditional linear methods, this method achieves a 73.68% improvement in the maximum residual absolute value of the extracted temperature values and a 78.95% improvement in the average residual absolute value. Full article
Show Figures

Figure 1

15 pages, 3288 KiB  
Article
Offshore Oil Spill Detection Based on CNN, DBSCAN, and Hyperspectral Imaging
by Ce Zhan, Kai Bai, Binrui Tu and Wanxing Zhang
Sensors 2024, 24(2), 411; https://doi.org/10.3390/s24020411 - 10 Jan 2024
Cited by 1 | Viewed by 1804
Abstract
Offshore oil spills have the potential to inflict substantial ecological damage, underscoring the critical importance of timely offshore oil spill detection and remediation. At present, offshore oil spill detection typically combines hyperspectral imaging with deep learning techniques. While these methodologies have made significant [...] Read more.
Offshore oil spills have the potential to inflict substantial ecological damage, underscoring the critical importance of timely offshore oil spill detection and remediation. At present, offshore oil spill detection typically combines hyperspectral imaging with deep learning techniques. While these methodologies have made significant advancements, they prove inadequate in scenarios requiring real-time detection due to limited model detection speeds. To address this challenge, a method for detecting oil spill areas is introduced, combining convolutional neural networks (CNNs) with the DBSCAN clustering algorithm. This method aims to enhance the efficiency of oil spill area detection in real-time scenarios, providing a potential solution to the limitations posed by the intricate structures of existing models. The proposed method includes a pre-feature selection process applied to the spectral data, followed by pixel classification using a convolutional neural network (CNN) model. Subsequently, the DBSCAN algorithm is employed to segment oil spill areas from the classification results. To validate our proposed method, we simulate an offshore oil spill environment in the laboratory, utilizing a hyperspectral sensing device to collect data and create a dataset. We then compare our method with three other models—DRSNet, CNN-Visual Transformer, and GCN—conducting a comprehensive analysis to evaluate the advantages and limitations of each model. Full article
Show Figures

Figure 1

10 pages, 4888 KiB  
Communication
Highway Visibility Estimation in Foggy Weather via Multi-Scale Fusion Network
by Pengfei Xiao, Zhendong Zhang, Xiaochun Luo, Jiaqing Sun, Xuecheng Zhou, Xixi Yang and Liang Huang
Sensors 2023, 23(24), 9739; https://doi.org/10.3390/s23249739 - 10 Dec 2023
Cited by 1 | Viewed by 1472
Abstract
Poor visibility has a significant impact on road safety and can even lead to traffic accidents. The traditional means of visibility monitoring no longer meet the current needs in terms of temporal and spatial accuracy. In this work, we propose a novel deep [...] Read more.
Poor visibility has a significant impact on road safety and can even lead to traffic accidents. The traditional means of visibility monitoring no longer meet the current needs in terms of temporal and spatial accuracy. In this work, we propose a novel deep network architecture for estimating the visibility directly from highway surveillance images. Specifically, we employ several image feature extraction methods to extract detailed structural, spectral, and scene depth features from the images. Next, we design a multi-scale fusion network to adaptively extract and fuse vital features for the purpose of estimating visibility. Furthermore, we create a real-scene dataset for model learning and performance evaluation. Our experiments demonstrate the superiority of our proposed method to the existing methods. Full article
Show Figures

Figure 1

Back to TopTop