remotesensing-logo

Journal Browser

Journal Browser

Advanced Artificial Intelligence and Deep Learning for Remote Sensing II

A special issue of Remote Sensing (ISSN 2072-4292). This special issue belongs to the section "Remote Sensing Image Processing".

Deadline for manuscript submissions: 15 January 2025 | Viewed by 11667

Special Issue Editors


E-Mail Website
Guest Editor
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
Interests: remote sensing image processing; object detection and tracking; scene understanding
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
College of Microelectronics and Communication Engineering, Chongqing University, Chongqing 401331, China
Interests: radar signal detection; target detection and recognition; radar system

Special Issue Information

Dear Colleagues,

Remote sensing is a fundamental tool for looking at the world from afar. The development of artificial intelligence(AI) and deep learning (DL) applications has paved the way for new research opportunities in various fields such as remote sensing, which uses Earth observation, disaster warning, and environmental monitoring. In recent years, with the continuous development of remote sensing technologies, especially the continuous emergence of different detection sensors and new detection systems, and the continuous accumulation of historical data and samples, it is possible to use AI and DL for big data training, and the field has become a research hotspot.

This Special Issue aims to report the latest advances and trends concerning the advanced AI and DL techniques applied to remote sensing data processing issues. Papers of both theoretical and applicative nature, as well as contributions regarding new AI and DL techniques for the remote sensing research community, are welcome. For this Special Issue, we invite experts and scholars in the field to contribute to the latest research progress of AI and DL in the fields of Earth observation, disaster warning, surface multi-temporal changes, environmental remote sensing, optical remote sensing and different sensor detection and imaging, so as to further promote the technological progress in this field.

The topic includes but is not limited to:

  • Object detection in high-resolution remote sensing imagery.
  • SAR object detection, scene classification.
  • Targets oriented multi temporal change detection.
  • Infrared target detection and recognition.
  • LiDAR point cloud data processing and scenes reconstruction.
  • UAV remote sensing and scenes perception.
  • Big data mining in remote sensing.
  • Interpretable deep learning in remote sensing.

Prof. Dr. Zhenming Peng
Prof. Dr. Zhengzhou Li
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Remote Sensing is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • objects detection
  • artificial intelligence
  • deep learning
  • scene reconstruction
  • scene perception
  • data mining
  • change detection
  • object recognition

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Related Special Issue

Published Papers (9 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Other

19 pages, 7749 KiB  
Article
Generative Simplex Mapping: Non-Linear Endmember Extraction and Spectral Unmixing for Hyperspectral Imagery
by John Waczak and David J. Lary
Remote Sens. 2024, 16(22), 4316; https://doi.org/10.3390/rs16224316 - 19 Nov 2024
Viewed by 440
Abstract
We introduce a new model for non-linear endmember extraction and spectral unmixing of hyperspectral imagery called Generative Simplex Mapping (GSM). The model represents endmember mixing using a latent space of points sampled within a (n1)-simplex corresponding to n [...] Read more.
We introduce a new model for non-linear endmember extraction and spectral unmixing of hyperspectral imagery called Generative Simplex Mapping (GSM). The model represents endmember mixing using a latent space of points sampled within a (n1)-simplex corresponding to n unique sources. Barycentric coordinates within this simplex are naturally interpreted as relative endmember abundances satisfying both the abundance sum-to-one and abundance non-negativity constraints. Points in this latent space are mapped to reflectance spectra via a flexible function combining linear and non-linear mixing. Due to the probabilistic formulation of the GSM, spectral variability is also estimated by a precision parameter describing the distribution of observed spectra. Model parameters are determined using a generalized expectation-maximization algorithm, which guarantees non-negativity for extracted endmembers. We first compare the GSM against three varieties of non-negative matrix factorization (NMF) on a synthetic data set of linearly mixed spectra from the USGS spectral database. Here, the GSM performed favorably for both endmember accuracy and abundance estimation with all non-linear contributions driven to zero by the fitting procedure. In a second experiment, we apply the GTM to model non-linear mixing in real hyperspectral imagery captured over a pond in North Texas. The model accurately identified spectral signatures corresponding to near-shore algae, water, and rhodamine tracer dye introduced into the pond to simulate water contamination by a localized source. Abundance maps generated using the GSM accurately track the evolution of the dye plume as it mixes into the surrounding water. Full article
Show Figures

Figure 1

13 pages, 10253 KiB  
Article
Combining KAN with CNN: KonvNeXt’s Performance in Remote Sensing and Patent Insights
by Minjong Cheon and Changbae Mun
Remote Sens. 2024, 16(18), 3417; https://doi.org/10.3390/rs16183417 - 14 Sep 2024
Viewed by 1332
Abstract
Rapid advancements in satellite technology have led to a significant increase in high-resolution remote sensing (RS) images, necessitating the use of advanced processing methods. Additionally, patent analysis revealed a substantial increase in deep learning and machine learning applications in remote sensing, highlighting the [...] Read more.
Rapid advancements in satellite technology have led to a significant increase in high-resolution remote sensing (RS) images, necessitating the use of advanced processing methods. Additionally, patent analysis revealed a substantial increase in deep learning and machine learning applications in remote sensing, highlighting the growing importance of these technologies. Therefore, this paper introduces the Kolmogorov-Arnold Network (KAN) model to remote sensing to enhance efficiency and performance in RS applications. We conducted several experiments to validate KAN’s applicability, starting with the EuroSAT dataset, where we combined the KAN layer with multiple pre-trained CNN models. Optimal performance was achieved using ConvNeXt, leading to the development of the KonvNeXt model. KonvNeXt was evaluated on the Optimal-31, AID, and Merced datasets for validation and achieved accuracies of 90.59%, 94.1%, and 98.1%, respectively. The model also showed fast processing speed, with the Optimal-31 and Merced datasets completed in 107.63 s each, while the bigger and more complicated AID dataset took 545.91 s. This result is meaningful since it achieved faster speeds and comparable accuracy compared to the existing study, which utilized VIT and proved KonvNeXt’s applicability for remote sensing classification tasks. Furthermore, we investigated the model’s interpretability by utilizing Occlusion Sensitivity, and by displaying the influential regions, we validated its potential use in a variety of domains, including medical imaging and weather forecasting. This paper is meaningful in that it is the first to use KAN in remote sensing classification, proving its adaptability and efficiency. Full article
Show Figures

Figure 1

18 pages, 18089 KiB  
Communication
High-Resolution PM10 Estimation Using Satellite Data and Model-Agnostic Meta-Learning
by Yue Yang, Jan Cermak, Xu Chen, Yunping Chen and Xi Hou
Remote Sens. 2024, 16(13), 2498; https://doi.org/10.3390/rs16132498 - 8 Jul 2024
Viewed by 1005
Abstract
Characterizing the spatial distribution of particles smaller than 10 μm (PM10) is of great importance for air quality management yet is very challenging because of the sparseness of air quality monitoring stations. In this study, we use a model-agnostic meta-learning-trained artificial [...] Read more.
Characterizing the spatial distribution of particles smaller than 10 μm (PM10) is of great importance for air quality management yet is very challenging because of the sparseness of air quality monitoring stations. In this study, we use a model-agnostic meta-learning-trained artificial neural network (MAML-ANN) to estimate the concentrations of PM10 at 60 m × 60 m spatial resolution by combining satellite-derived aerosol optical depth (AOD) with meteorological data. The network is designed to regress from the predictors at a specific time to the ground-level PM10 concentration. We utilize the ANN model to capture the time-specific nonlinearity among aerosols, meteorological conditions, and PM10, and apply MAML to enable the model to learn the nonlinearity across time from only a small number of data samples. MAML is also employed to transfer the knowledge learned from coarse spatial resolution to high spatial resolution. The MAML-ANN model is shown to accurately estimate high-resolution PM10 in Beijing, with coefficient of determination of 0.75. MAML improves the PM10 estimation performance of the ANN model compared with the baseline using pre-trained initial weights. Thus, MAML-ANN has the potential to estimate particulate matter estimation at high spatial resolution over other data-sparse, heavily polluted, and small regions. Full article
Show Figures

Figure 1

20 pages, 22183 KiB  
Article
FEFN: Feature Enhancement Feedforward Network for Lightweight Object Detection in Remote Sensing Images
by Jing Wu, Rixiang Ni, Zhenhua Chen, Feng Huang and Liqiong Chen
Remote Sens. 2024, 16(13), 2398; https://doi.org/10.3390/rs16132398 - 29 Jun 2024
Cited by 1 | Viewed by 1220
Abstract
Object detection in remote sensing images has become a crucial component of computer vision. It has been employed in multiple domains, including military surveillance, maritime rescue, and military operations. However, the high density of small objects in remote sensing images makes it challenging [...] Read more.
Object detection in remote sensing images has become a crucial component of computer vision. It has been employed in multiple domains, including military surveillance, maritime rescue, and military operations. However, the high density of small objects in remote sensing images makes it challenging for existing networks to accurately distinguish objects from shallow image features. These factors contribute to many object detection networks that produce missed detections and false alarms, particularly for densely arranged objects and small objects. To address the above problems, this paper proposes a feature enhancement feedforward network (FEFN), based on a lightweight channel feedforward module (LCFM) and a feature enhancement module (FEM). First, the FEFN captures shallow spatial information in images through a lightweight channel feedforward module that can extract the edge information of small objects such as ships. Next, it enhances the feature interaction and representation by utilizing a feature enhancement module that can achieve more accurate detection results for densely arranged objects and small objects. Finally, comparative experiments on two publicly challenging remote sensing datasets demonstrate the effectiveness of the proposed method. Full article
Show Figures

Figure 1

21 pages, 29397 KiB  
Article
TFCD-Net: Target and False Alarm Collaborative Detection Network for Infrared Imagery
by Siying Cao, Zhi Li, Jiakun Deng, Yi’an Huang and Zhenming Peng
Remote Sens. 2024, 16(10), 1758; https://doi.org/10.3390/rs16101758 - 15 May 2024
Viewed by 926
Abstract
Infrared small target detection (ISTD) plays a crucial role in both civilian and military applications. Detecting small targets against dense cluttered backgrounds remains a challenging task, requiring the collaboration of false alarm source elimination and target detection. Existing approaches mainly focus on modeling [...] Read more.
Infrared small target detection (ISTD) plays a crucial role in both civilian and military applications. Detecting small targets against dense cluttered backgrounds remains a challenging task, requiring the collaboration of false alarm source elimination and target detection. Existing approaches mainly focus on modeling targets while often overlooking false alarm sources. To address this limitation, we propose a Target and False Alarm Collaborative Detection Network to leverage the information provided by false alarm sources and the background. Firstly, we introduce a False Alarm Source Estimation Block (FEB) that estimates potential interferences present in the background by extracting features at multiple scales and using gradual upsampling for feature fusion. Subsequently, we propose a framework that employs multiple FEBs to eliminate false alarm sources across different scales. Finally, a Target Segmentation Block (TSB) is introduced to accurately segment the targets and produce the final detection result. Experiments conducted on public datasets show that our model achieves the highest and second-highest scores for the IoU, Pd, and AUC and the lowest Fa among the DNN methods. These results demonstrate that our model accurately segments targets while effectively extracting false alarm sources, which can be used for further studies. Full article
Show Figures

Figure 1

22 pages, 3618 KiB  
Article
An Integrated Detection and Multi-Object Tracking Pipeline for Satellite Video Analysis of Maritime and Aerial Objects
by Zhijuan Su, Gang Wan, Wenhua Zhang, Ningbo Guo, Yitian Wu, Jia Liu, Dianwei Cong, Yutong Jia and Zhanji Wei
Remote Sens. 2024, 16(4), 724; https://doi.org/10.3390/rs16040724 - 19 Feb 2024
Cited by 2 | Viewed by 1446
Abstract
Optical remote sensing videos, as a new source of remote sensing data that has emerged in recent years, have significant potential in remote sensing applications, especially national defense. In this paper, a tracking pipeline named TDNet (tracking while detecting based on a neural [...] Read more.
Optical remote sensing videos, as a new source of remote sensing data that has emerged in recent years, have significant potential in remote sensing applications, especially national defense. In this paper, a tracking pipeline named TDNet (tracking while detecting based on a neural network) is proposed for optical remote sensing videos based on a correlation filter and deep neural networks. The pipeline is used to simultaneously track ships and planes in videos. There are many target tracking methods for general video data, but they suffer some difficulties in remote sensing videos with low resolution and those influenced by weather conditions. The tracked targets are usually misty. Therefore, in TDNet, we propose a new multi-target tracking method called MT-KCF and a detecting-assisted tracking (i.e., DAT) module to improve tracking accuracy and precision. Meanwhile, we also design a new target recognition (i.e., NTR) module to recognise newly emerged targets. In order to verify the performance of TDNet, we compare our method with several state-of-the-art tracking methods on optical video remote sensing data sets acquired from the Jilin No. 1 satellite. The experimental results demonstrate the effectiveness and the state-of-the-art performance of the proposed method. The proposed method can achieve more than 90% performance in terms of precision for single-target tracking tasks and more than 85% performance in terms of MOTA for multi-object tracking tasks. Full article
Show Figures

Figure 1

24 pages, 13566 KiB  
Article
EL-NAS: Efficient Lightweight Attention Cross-Domain Architecture Search for Hyperspectral Image Classification
by Jianing Wang, Jinyu Hu, Yichen Liu, Zheng Hua, Shengjia Hao and Yuqiong Yao
Remote Sens. 2023, 15(19), 4688; https://doi.org/10.3390/rs15194688 - 25 Sep 2023
Cited by 3 | Viewed by 1504
Abstract
Deep learning (DL) algorithms have demonstrated important breakthroughs for hyperspectral image (HSI) classification. Despite the remarkable success of DL, the burden of a manually designed DL structure with increased depth and size aroused the difficulty for the application in the mobile and embedded [...] Read more.
Deep learning (DL) algorithms have demonstrated important breakthroughs for hyperspectral image (HSI) classification. Despite the remarkable success of DL, the burden of a manually designed DL structure with increased depth and size aroused the difficulty for the application in the mobile and embedded devices in a real application. To tackle this issue, in this paper, we proposed an efficient lightweight attention network architecture search algorithm (EL-NAS) for realizing an efficient automatic design of a lightweight DL structure as well as improving the classification performance of HSI. First, aimed at realizing an efficient search procedure, we construct EL-NAS based on a differentiable network architecture search (NAS), which can greatly accelerate the convergence of the over-parameter supernet in a gradient descent manner. Second, in order to realize lightweight search results with high accuracy, a lightweight attention module search space is designed for EL-NAS. Finally, further for alleviating the problem of higher validation accuracy and worse classification performance, the edge decision strategy is exploited to perform edge decisions through the entropy of distribution estimated over non-skip operations to avoid further performance collapse caused by numerous skip operations. To verify the effectiveness of EL-NAS, we conducted experiments on several real-world hyperspectral images. The results demonstrate that the proposed EL-NAS indicates a more efficient search procedure with smaller parameter sizes and high accuracy performance for HSI classification, even under data-independent and sensor-independent scenarios. Full article
Show Figures

Graphical abstract

Other

Jump to: Research

16 pages, 8285 KiB  
Technical Note
A Feature-Driven Inception Dilated Network for Infrared Image Super-Resolution Reconstruction
by Jiaxin Huang, Huicong Wang, Yuhan Li and Shijian Liu
Remote Sens. 2024, 16(21), 4033; https://doi.org/10.3390/rs16214033 - 30 Oct 2024
Viewed by 422
Abstract
Image super-resolution (SR) algorithms based on deep learning yield good visual performances on visible images. Due to the blurred edges and low contrast of infrared (IR) images, methods transferred directly from visible images to IR images have a poor performance and ignore the [...] Read more.
Image super-resolution (SR) algorithms based on deep learning yield good visual performances on visible images. Due to the blurred edges and low contrast of infrared (IR) images, methods transferred directly from visible images to IR images have a poor performance and ignore the demands of downstream detection tasks. Therefore, an Inception Dilated Super-Resolution (IDSR) network with multiple branches is proposed. A dilated convolutional branch captures high-frequency information to reconstruct edge details, while a non-local operation branch captures long-range dependencies between any two positions to maintain the global structure. Furthermore, deformable convolution is utilized to fuse features extracted from different branches, enabling adaptation to targets of various shapes. To enhance the detection performance of low-resolution (LR) images, we crop the images into patches based on target labels before feeding them to the network. This allows the network to focus on learning the reconstruction of the target areas only, reducing the interference of background areas in the target areas’ reconstruction. Additionally, a feature-driven module is cascaded at the end of the IDSR network to guide the high-resolution (HR) image reconstruction with feature prior information from a detection backbone. This method has been tested on the FLIR Thermal Dataset and the M3FD Dataset and compared with five mainstream SR algorithms. The final results demonstrate that our method effectively maintains image texture details. More importantly, our method achieves 80.55% mAP, outperforming other methods on FLIR Dataset detection accuracy, and with 74.7% mAP outperforms other methods on M3FD Dataset detection accuracy. Full article
Show Figures

Figure 1

15 pages, 10515 KiB  
Technical Note
A DeturNet-Based Method for Recovering Images Degraded by Atmospheric Turbulence
by Xiangxi Li, Xingling Liu, Weilong Wei, Xing Zhong, Haotong Ma and Junqiu Chu
Remote Sens. 2023, 15(20), 5071; https://doi.org/10.3390/rs15205071 - 23 Oct 2023
Cited by 3 | Viewed by 1491
Abstract
Atmospheric turbulence is one of the main issues causing image blurring, dithering, and other degradation problems when detecting targets over long distances. Due to the randomness of turbulence, degraded images are hard to restore directly using traditional methods. With the rapid development of [...] Read more.
Atmospheric turbulence is one of the main issues causing image blurring, dithering, and other degradation problems when detecting targets over long distances. Due to the randomness of turbulence, degraded images are hard to restore directly using traditional methods. With the rapid development of deep learning, blurred images can be restored correctly and directly by establishing a nonlinear mapping relationship between the degraded and initial objects based on neural networks. These data-driven end-to-end neural networks offer advantages in turbulence image reconstruction due to their real-time properties and simplified optical systems. In this paper, inspired by the connection between the turbulence phase diagram characteristics and the attentional mechanisms for neural networks, we propose a new deep neural network called DeturNet to enhance the network’s performance and improve the quality of image reconstruction results. DeturNet employs global information aggregation operations and amplifies notable cross-dimensional reception regions, thereby contributing to the recovery of turbulence-degraded images. Full article
Show Figures

Figure 1

Back to TopTop