Advanced Machine Vision with Mathematics

A special issue of Mathematics (ISSN 2227-7390). This special issue belongs to the section "Mathematics and Computer Science".

Deadline for manuscript submissions: 20 July 2025 | Viewed by 5538

Special Issue Editor


E-Mail Website
Guest Editor
Biomedical, Industrial and Systems Engineering Department, Gannon University, Erie, PA 16541, USA
Interests: smart manufacturing; machine learning; computer vision; simulation; scheduling
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

Machine vision technologies have undergone significant advancements in recent years, fueling their widespread applications across various industries. Leveraging computer vision, artificial intelligence, and hardware innovations, these technologies are transforming automation, inspection, and analysis processes. In manufacturing, machine vision is used for quality control, defect detection, and product traceability, enhancing production efficiency and reducing errors. In agriculture, it aids in crop monitoring, pest control, and yield optimization. In healthcare, it supports medical image analysis and diagnosis, improving patient care. The automotive sector relies on machine vision for autonomous vehicles, enabling navigation and obstacle detection. Retail benefits from facial recognition and inventory management, while security employs it for surveillance and access control. Environmental monitoring, robotics, and augmented reality also harness machine vision. With the ever-evolving technology landscape, the potential applications of machine vision continue to expand, offering enhanced precision, efficiency, and innovation across numerous domains.
How to further apply mathematical models, algorithms and techniques in machine vision is an essential problem worthy of study. This special issue aims to invite and publish recent research studies on the latest advances in various intersections of machine vision technologies and applied mathematics, and their recent applications in various industries. Topics include, but are not limited to, the following:

  • Image classification;
  • object recognition and detection;
  • image segmentation;
  • feature extraction;
  • image registration;
  • object tracking;
  • 3D vision;
  • generative models;
  • biometrics;
  • deep learning for machine vision;
  • visual slam;
  • medical imaging;
  • other applications of machine vision.

Dr. Longfei Zhou
Guest Editor

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Mathematics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • image classification
  • object recognition and detection
  • image segmentation
  • feature extraction
  • image registration
  • object tracking
  • 3D vision
  • generative models
  • biometrics
  • deep learning for machine vision
  • visual slam
  • medical imaging

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (5 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

17 pages, 29239 KiB  
Article
Three-Dimensional Vehicle Detection and Pose Estimation in Monocular Images for Smart Infrastructures
by Javier Borau Bernad, Álvaro Ramajo-Ballester and José María Armingol Moreno
Mathematics 2024, 12(13), 2027; https://doi.org/10.3390/math12132027 - 29 Jun 2024
Viewed by 894
Abstract
Over the last decades, the idea of smart cities has evolved from a visionary concept of the future into a concrete reality. However, the vision of smart cities has not been fully realized within our society, partly due to the challenges encountered in [...] Read more.
Over the last decades, the idea of smart cities has evolved from a visionary concept of the future into a concrete reality. However, the vision of smart cities has not been fully realized within our society, partly due to the challenges encountered in contemporary data collection systems. Despite these obstacles, advancements in deep learning and computer vision have propelled the development of highly accurate detection algorithms capable of obtaining 3D data from image sources. Nevertheless, this approach has predominantly centered on data extraction from a vehicle’s perspective, bypassing the advantages of using infrastructure-mounted cameras for performing 3D pose estimation of vehicles in urban environments. This paper focuses on leveraging 3D pose estimation from this alternative perspective, benefiting from the enhanced field of view that infrastructure-based cameras provide, avoiding occlusions, and obtaining more information from the objects’ sizes, leading to better results and more accurate predictions compared to models trained on a vehicle’s viewpoint. Therefore, this research proposes a new path for exploration, supporting the integration of monocular infrastructure-based data collection systems into smart city development. Full article
(This article belongs to the Special Issue Advanced Machine Vision with Mathematics)
Show Figures

Figure 1

31 pages, 7137 KiB  
Article
Distributed Batch Learning of Growing Neural Gas for Quick and Efficient Clustering
by Chyan Zheng Siow, Azhar Aulia Saputra, Takenori Obo and Naoyuki Kubota
Mathematics 2024, 12(12), 1909; https://doi.org/10.3390/math12121909 - 20 Jun 2024
Cited by 1 | Viewed by 859
Abstract
Growing neural gas (GNG) has been widely used in topological mapping, clustering and unsupervised tasks. It starts from two random nodes and grows until it forms a topological network covering all data. The time required for growth depends on the total amount of [...] Read more.
Growing neural gas (GNG) has been widely used in topological mapping, clustering and unsupervised tasks. It starts from two random nodes and grows until it forms a topological network covering all data. The time required for growth depends on the total amount of data and the current network nodes. To accelerate growth, we introduce a novel distributed batch processing method to extract the rough distribution called Distributed Batch Learning Growing Neural Gas (DBL-GNG). First, instead of using a for loop in standard GNG, we adopt a batch learning approach to accelerate learning. To do this, we replace most of the standard equations with matrix calculations. Next, instead of starting with two random nodes, we start with multiple nodes in different distribution areas. Furthermore, we also propose to add multiple nodes to the network instead of adding them one by one. Finally, we introduce an edge cutting method to reduce unimportant links between nodes to obtain a better cluster network. We demonstrate DBL-GNG on multiple benchmark datasets. From the results, DBL-GNG performs faster than other GNG methods by at least 10 times. We also demonstrate the scalability of DBL-GNG by implementing a multi-scale batch learning process in it, named MS-DBL-GNG, which successfully obtains fast convergence results. In addition, we also demonstrate the dynamic data adaptation of DBL-GNG to 3D point cloud data. It is capable of processing and mapping topological nodes on point cloud objects in real time. Full article
(This article belongs to the Special Issue Advanced Machine Vision with Mathematics)
Show Figures

Figure 1

17 pages, 2679 KiB  
Article
Enhanced Unmanned Aerial Vehicle Localization in Dynamic Environments Using Monocular Simultaneous Localization and Mapping and Object Tracking
by Youssef El Gaouti, Fouad Khenfri, Mehdi Mcharek and Cherif Larouci
Mathematics 2024, 12(11), 1619; https://doi.org/10.3390/math12111619 - 22 May 2024
Cited by 1 | Viewed by 998
Abstract
This work proposes an innovative approach to enhance the localization of unmanned aerial vehicles (UAVs) in dynamic environments. The methodology integrates a sophisticated object-tracking algorithm to augment the established simultaneous localization and mapping (ORB-SLAM) framework, utilizing only a monocular camera setup. Moving objects [...] Read more.
This work proposes an innovative approach to enhance the localization of unmanned aerial vehicles (UAVs) in dynamic environments. The methodology integrates a sophisticated object-tracking algorithm to augment the established simultaneous localization and mapping (ORB-SLAM) framework, utilizing only a monocular camera setup. Moving objects are detected by harnessing the power of YOLOv4, and a specialized Kalman filter is employed for tracking. The algorithm is integrated into the ORB-SLAM framework to improve UAV pose estimation by correcting the impact of moving elements and effectively removing features connected to dynamic elements from the ORB-SLAM process. Finally, the results obtained are recorded using the TUM RGB-D dataset. The results demonstrate that the proposed algorithm can effectively enhance the accuracy of pose estimation and exhibits high accuracy and robustness in real dynamic scenes. Full article
(This article belongs to the Special Issue Advanced Machine Vision with Mathematics)
Show Figures

Figure 1

16 pages, 59324 KiB  
Article
A New Biorthogonal Spline Wavelet-Based K-Layer Network for Underwater Image Enhancement
by Dujuan Zhou, Zhanchuan Cai and Dan He
Mathematics 2024, 12(9), 1366; https://doi.org/10.3390/math12091366 - 30 Apr 2024
Viewed by 857
Abstract
Wavelet decomposition is pivotal for underwater image processing, known for its ability to analyse multi-scale image features in the frequency and spatial domains. In this paper, we propose a new biorthogonal cubic special spline wavelet (BCS-SW), based on the Cohen–Daubechies–Feauveau (CDF) wavelet construction [...] Read more.
Wavelet decomposition is pivotal for underwater image processing, known for its ability to analyse multi-scale image features in the frequency and spatial domains. In this paper, we propose a new biorthogonal cubic special spline wavelet (BCS-SW), based on the Cohen–Daubechies–Feauveau (CDF) wavelet construction method and the cubic special spline algorithm. BCS-SW has better properties in compact support, symmetry, and frequency domain characteristics. In addition, we propose a K-layer network (KLN) based on the BCS-SW for underwater image enhancement. The KLN performs a K-layer wavelet decomposition on underwater images to extract various frequency domain features at multiple frequencies, and each decomposition layer has a convolution layer corresponding to its spatial size. This design ensures that the KLN can understand the spatial and frequency domain features of the image at the same time, providing richer features for reconstructing the enhanced image. The experimental results show that the proposed BCS-SW and KLN algorithm has better image enhancement effect than some existing algorithms. Full article
(This article belongs to the Special Issue Advanced Machine Vision with Mathematics)
Show Figures

Figure 1

24 pages, 21453 KiB  
Article
U-Net-Based Learning Using Enhanced Lane Detection with Directional Lane Attention Maps for Various Driving Environments
by Seung-Hwan Lee and Sung-Hak Lee
Mathematics 2024, 12(8), 1206; https://doi.org/10.3390/math12081206 - 17 Apr 2024
Cited by 3 | Viewed by 1405
Abstract
Recent advancements in optical and electronic sensor technologies, coupled with the proliferation of computing devices (such as GPUs), have enabled real-time autonomous driving systems to become a reality. Hence, research in algorithmic advancements for advanced driver assistance systems (ADASs) is rapidly expanding, with [...] Read more.
Recent advancements in optical and electronic sensor technologies, coupled with the proliferation of computing devices (such as GPUs), have enabled real-time autonomous driving systems to become a reality. Hence, research in algorithmic advancements for advanced driver assistance systems (ADASs) is rapidly expanding, with a primary focus on enhancing robust lane detection capabilities to ensure safe navigation. Given the widespread adoption of cameras on the market, lane detection relies heavily on image data. Recently, CNN-based methods have attracted attention due to their effective performance in lane detection tasks. However, with the expansion of the global market, the endeavor to achieve reliable lane detection has encountered challenges presented by diverse environmental conditions and road scenarios. This paper presents an approach that focuses on detecting lanes in road areas traversed by vehicles equipped with cameras. In the proposed method, a U-Net based framework is employed for training, and additional lane-related information is integrated into a four-channel input data format that considers lane characteristics. The fourth channel serves as the edge attention map (E-attention map), helping the modules achieve more specialized learning regarding the lane. Additionally, the proposition of an approach to assign weights to the loss function during training enhances the stability and speed of the learning process, enabling robust lane detection. Through ablation experiments, the optimization of each parameter and the efficiency of the proposed method are demonstrated. Also, the comparative analysis with existing CNN-based lane detection algorithms shows that the proposed training method demonstrates superior performance. Full article
(This article belongs to the Special Issue Advanced Machine Vision with Mathematics)
Show Figures

Figure 1

Back to TopTop