Next Article in Journal
Investigation of Device- and Circuit-Level Reliability of Inverse-Mode Silicon-Germanium Heterojunction Bipolar Transistors
Previous Article in Journal
Gearbox Fault Diagnosis Based on Adaptive Variational Mode Decomposition–Stationary Wavelet Transform and Ensemble Refined Composite Multiscale Fluctuation Dispersion Entropy
Previous Article in Special Issue
A Low-Measurement-Cost-Based Multi-Strategy Hyperspectral Image Classification Scheme
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Hyperspectral Images of Explosive Fragments Based on Spatial–Spectral Combination

by
Donge Zhao
1,2,*,†,
Peiyun Yu
2,†,
Feng Guo
2,
Xuefeng Yang
2,
Yayun Ma
2,
Changli Wang
3,
Kang Li
3,
Wenbo Chu
4 and
Bin Zhang
2
1
State Key Laboratory of Dynamic Measurement Technology, North University of China, Taiyuan 030051, China
2
School of Information and Communication Engineering, North University of China, Taiyuan 030051, China
3
Northwest Institute of Nuclear Technology, Xi’an 710024, China
4
College of Mechanical and Electrical Engineering, North University of China, Taiyuan 030051, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(22), 7131; https://doi.org/10.3390/s24227131
Submission received: 24 September 2024 / Revised: 27 October 2024 / Accepted: 4 November 2024 / Published: 6 November 2024

Abstract

:
The identification and recovery of explosive fragments can provide a reference for the evaluation of explosive power and the design of explosion-proof measures. At present, fragment detection usually uses a few bands in the visible light or infrared bands for imaging, without fully utilizing multi-band spectral information. Hyperspectral imaging has high spectral resolution and can provide multidimensional reference information for the fragments to be classified. Therefore, this article proposed a spatial–spectral joint method for explosive fragment classification by combining hyperspectral imaging technology. In a laboratory environment, this article collected hyperspectral images of explosion fragments scattered in simulated scenes. In order to extract effective features from redundant spectral information and improve classification accuracy, this paper adopted a classification framework based on deep learning. This framework used a convolutional neural network–bidirectional long short-term memory network (CNN-BiLSTM) as the spectral information classification model and a U-shaped network (U-Net) as the spatial segmentation model. The experimental results showed that the overall accuracy exceeds 95.2%. The analysis results indicated that the method of spatial–spectral combination can accurately identify explosive fragment targets. It validated the feasibility of using hyperspectral imaging for explosive fragment classification in laboratory environments. Due to the complex environment of the actual explosion site, this study still needs to be validated in outdoor environments. Our next step is to use airborne hyperspectral imaging to identify explosive fragments in outdoor environments.

1. Introduction

The performance parameters of fragments primarily involve the number [1], material composition, and velocity. Identifying explosive fragments is crucial for assessing explosive power and designing effective explosion-proof measures. Currently, research primarily focuses on evaluating target damage effectiveness [2] and implementing optical imaging techniques [3]. The dispersion [4], shape [5], average quality [6], and other characteristics of explosive fragments hold significant importance in performance evaluations. However, existing fragment detection methods only provide one-dimensional information about fragments, such as mass, spatial details, and spectral information, typically captured using a single band, such as visible or infrared. These methods overlook the varying degrees of light absorption at different wavelengths by fragments and background materials, failing to extract the multidimensional information of the fragments. Consequently, incorporating multidimensional information for fragment analysis is essential for enhancing their structure and performance.
Hyperspectral imaging technology combines the advantages of traditional imaging technology and spectral technology [7]. It can obtain the spatial information and the spectral information of continuous bands in the images [8]. Hyperspectral images can provide reference images for spatial positions and provide multi-dimensional information for target classification. In the realm of object detection, hyperspectral imaging technology offers substantial advantages in the enhancement of classification accuracy [9]. Over the course of several decades, this technology has found applications in civilian sectors, including agricultural production [10], landscape characterization [11], environmental monitoring [12], and food ingredient identification [13]. In military domains, hyperspectral detection technology has been successfully employed in the recognition of maritime and shallow-water targets [14], battlefield camouflage [15], and land targets [16], showcasing its extensive range of potential applications.
The small size and irregular shape of the generated fragments pose challenges for conventional imaging methods in accurately identifying ground fragments. The fragments and background materials have different molecular structures, leading to variations in the reflection characteristics of light across different wavelengths. On the other hand, hyperspectral images provide a contiguous spectral signature for each pixel, reflecting the physical and chemical properties of objects [17]. Therefore, the method of hyperspectral imaging can capture more detailed information under different spectra [18], making it highly suitable for distinguishing between different fragments and background materials. The recognition of iron fragments was studied to verify the feasibility of hyperspectral detection methods for iron fragment recognition based on traditional methods [19].
In comparison to traditional classification methods, deep learning offers the advantage of automated feature extraction and selection within the model, eliminating the need for manual feature construction [20]. At present, various methods of deep learning have been proposed for the classification of hyperspectral images [18,21]. Compared with other deep-learning methods, convolutional neural networks (CNN) have extensive applications in hyperspectral image classification due to their excellent spectral feature representation. Hu first used the CNN in hyperspectral image classification with only a one-dimensional (1D) convolution kernel and focused on the spectral features of hyperspectral imagery [22]. Despite CNNs having achieved promising results in most cases [23,24], they still encounter the following limitations:
(1)
CNN is a vector-based approach that treats the input of hyperspectral images as a collection of pixel vectors [25]. The equal treatment of all channels can lead to a CNN being incredibly insensitive to the spectral sequence information [26]. The results are that the feature maps cannot effectively represent the spectral sequence information from hyperspectral images;
(2)
A 1D CNN can only consider the spectral information of hyperspectral images and ignores the spatial relationships between pixels.
However, the spectral information of the sample points in hyperspectral images is typical sequence data. The change in reflectivity intensity of adjacent bands usually leads to a change in the overall trend of reflectivity. While CNNs are known for their automatic feature-extraction capabilities [27], recurrent neural networks (RNNs) are more adept at handling sequential data [28]. Specifically, bidirectional long short-term memory networks (BiLSTM), a type of RNN, are well-suited for processing time series data and effectively capturing the relationships between the current data and their preceding or subsequent data points [29]. Taking advantage of both CNNs and RNNs, we adopted a combined model called a convolutional neural network–bidirectional long short-term memory network (CNN-BiLSTM) for spectral information classification. The CNN-BiLSTM model offers robust feature-extraction capabilities while simultaneously considering the spectral reflectance sequence’s front and back band information for each sample point [30]. This fusion of CNN and BiLSTM allows for comprehensive analysis and improves the classification performance of hyperspectral data.
Nowadays, the application of deep-learning technology in the image segment has attracted extensive attention. Ronneberger et al. proposed a U-shaped network (U-Net) at the MICCAI conference in 2015 [31]. U-Net has distinct advantages, including its ability to achieve high accuracy with less training data and its fast training speed [32]. To extract the spatial relationships between pixels from hyperspectral images, this study trained the semantic segmentation model based on U-Net. The model trained with U-Net can restrain the irrelevant regions in an input image and highlight the striking features suitable for specific tasks [33]. It is conducive to eradicating the inevitability of applying backgrounds of cascading CNNs.
On this basis, the study developed a spatial–spectral combination algorithm wherein the majority category of pixels within a region determines the category assigned to the entire region. On the basis of U-Net for image segmentation, the CNN-BiLSTM model is used to recognize fragments and background pixel by pixel. The category with the highest number of pixels in each region represents the category of that region, accurately identifying the target fragments from the background. The method increases the precision of the model and retains the spatial data aimed at 2D segmentation. By combining both the spatial and spectral information, the method effectively leverages the strengths of the respective techniques.
The main technical contributions of this work are summarized as follows:
(1)
Creation of datasets: We collected hyperspectral images containing five types of explosive fragments and five common background materials. We preprocessed these data, including black-and-white correction and region of interest (ROI) extraction. These datasets include both spectral and spatial information, providing valuable resources for further analysis;
(2)
Spectral information classification model: To classify the spectral information accurately, we adopted the CNN-BiLSTM model. This novel approach combines CNN and BiLSTM to achieve high classification accuracy, even with limited training samples;
(3)
Spatial–spectral combination algorithm: We designed a spatial–spectral combination algorithm that successfully classifies targets in hyperspectral images. By combining U-Net with CNN-BiLSTM, we effectively merge both spatial and spectral information, enhancing the overall classification performance;
(4)
Experimental verification: To validate the effectiveness of our hyperspectral imaging detection methods in the identification of explosive fragments, we conducted classification experiments on the collected datasets. The results demonstrate the feasibility and potential of our approach.
This study used hyperspectral imaging technology for fragment classification based on the differences in explosion fragments and background spectral reflectance information. This research could effectively distinguish between fragments and background in a laboratory environment and visualize the materials and positions of different fragments and background materials. The purpose of this study is to verify the feasibility of using hyperspectral imaging to identify explosive fragments. At the same time, it provides guidance for the intelligent identification and recovery of fragments in the field in the future and has practical significance for evaluating the power of fragments.

2. Materials and Methods

2.1. Hyperspectral Imaging System

The hyperspectral camera model used in this study is a Hypersec VNIR-A (Headwall, Boston, MA, United States), which consists of a 1600×2500 high-resolution CCD lens, a total reflection grating, and a slit. The band interval for the sample data is 2.5 nm, and a total of 234 bands from 400 to 1000 nm are used for the spectral imaging of the experimental samples. The hyperspectral image acquisition system includes a hyperspectral camera, a light source, a light source control system, a reference whiteboard, a mobile loading platform, a motor control system, and a computer. The light source adopts a halogen lamp, which can meet the detector’s requirements for the wavelength range of the light source. The schematic diagram of the hyperspectral image acquisition system is shown in Figure 1. The specific parameters of the components of the hyperspectral imaging system used in this study are shown in Table 1.
First, we performed image acquisition in an optical darkroom to reduce the impact of ambient light on the spectral information of the sample. We adjusted the brightness of the halogen light source to avoid image oversaturation. Second, we set the camera integration time to 40 ms and the gaze time to 39 ms to prevent image distortion. The distance between the test sample and the camera slit was 30 cm. Finally, we set the speed of the mobile loading platform to 2.168 mm/s to ensure that the collected sample images were clear and frame-accurate. Through these control measures, we could ensure that the data collected in the experiment met the experimental requirements and that the spectral and spatial information of the samples was complete and accurate. The actual scene image of the hyperspectral image acquisition system is shown in Figure 2a, and the darkroom environment image is shown in Figure 2b.

2.2. Black-and-White Board Correction

To reduce the image noises caused by the camera’s dark current and light, it is necessary to perform the black-and-white board correction on the collected hyperspectral images. We used a reference whiteboard to perform the black-and-white correction on the collected hyperspectral images. The correction formula is defined as:
R n o r m = R r a w R d a r k R w h i t e R d a r k
where R n o r m is the hyperspectral image after the black-and-white correction, R r a w is the original hyperspectral image, R d a r k is the full dark image collected without removing the lens cover, and R w h i t e is the image obtained from a standard reference whiteboard with a reflectance of 99%.

2.3. Experiment Datasets

2.3.1. Fragments and Background Samples

Generally, fragments are often made of different materials such as steel, synthetic materials, tungsten, copper, iron, and aluminum to achieve varying degrees of damage effects on different types of targets [34]. We collected five kinds of explosive fragments, as well as five common background materials. The true-color images of the collected fragments and background samples are shown in Figure 3.
The first line displays images (a) to (e), which depict tungsten fragments, copper fragments, iron fragments, aluminum fragments, and composite materials, respectively. The second line shows images (f) to (j), representing background materials, stones, soil, leaves, foams, and barks, respectively.

2.3.2. Extracting Regions of Interest

Extracting regions of interest (RoI) not only removes redundant data but also improves the speed of image feature processing and analysis and eliminates interference from other irrelevant data [35]. For hyperspectral remote-sensing images, the closed boundaries of the target area can be obtained by extracting interested objects, which can be used for subsequent object classification. Before performing dimensionality reduction, the input data needs to be converted from 3D to 2D. Region of interest extraction and data conversion of bark samples are illustrated in Figure 4.
In Figure 4, we captured a hyperspectral sample image using a hyperspectral system, with dimensions of length (H), width (W), and band number (B). Employing three rectangular boxes, sized a ×b, c × d, and e × f, we extracted the regions of interest, resulting in three distinct regions with dimensions of a × b × B, c × d × B, and e × f × B. Each 3D region of interest was converted into 2D hyperspectral data with respective lengths of a × b, c × d, e × f, and a width of B. Finally, all the data was concatenated along the length direction to form a 2D hyperspectral image data of (a × b + c × d + e × f) × B. In this experiment, we extracted regions of interest from ten collected samples, with 4000 sample points extracted from each sample. Finally, a total of 40,000 data points were used as the datasets of the spectral information. The dataset was divided into training sets, validation sets, and test sets in a 6:2:2 ratio through random sampling. The training set consisted of 24,000 sample points, while the validation and testing sets each contained 8000 sample points.

2.4. Convolutional Neural Network–Bidirectional Long Short-Term Memory Network

CNN-BiLSTM is the deep-learning model that combines CNN and LSTM [36]. It extracts features of different sizes through the convolution layers in CNN and effectively captures the contextual relationships in text using BiLSTM, thus endowing the model with strong feature-extraction and memory capabilities [37]. The model is mainly divided into the 1D-CNN part, the LSTM part, and the output part. The detailed process is as follows.
During the training stage, the CNN component of the network employs a structure comprising four convolutional layers and four maxpooling layers. The features extracted by the CNN are then passed through a dense layer and fed into the BiLSTM network [38]. The BiLSTM component consists of two layers of LSTM networks, leveraging the distinct network structure of LSTM to further extract the time series features that may have been overlooked by the 1D-CNN component. This process enhances the accuracy of the classification model. The features extracted by the BiLSTM are then passed through a dense layer and fed into the dropout layer. Finally, the classification results are obtained through dense layers. The model structure is shown in Figure 5.
Due to the deep network structure employed in the aforementioned model, there exists a risk of overfitting during the training process. To mitigate this risk, a dropout layer has been included between dense layer 2 and dense layer 3. This layer serves the purpose of adjusting the weights of neurons and decreasing reliance on any specific neuron. Furthermore, the dropout layer aids in reducing the influence of noise and enhancing the model’s robustness. The loss function utilized in this model is the cross-entropy loss function, while the gradient descent optimization algorithm employs the Adam optimizer.

2.5. U-Shaped Network

In order to obtain the spatial distribution pattern of fragments through a streamlined research process, traditional methods were anticipated to be substituted with artificial intelligence techniques to enhance the intelligence of this study. Therefore, this study adopted a semantic segmentation model based on U-Net to solve the problem of spatial region segmentation. U-Net is a neat end-to-end neural network architecture, which is known for its “U” shape [39]. The U-Net structure diagram is shown in Figure 6.
In the above figure, the numbers 1, 64, 128, 256, 512, and 1024 represent the number of image channels. The size of the input image is 572 × 572. Blue/white boxes represent feature maps. Blue arrows represent the 3 × 3 convolution for feature extraction. Grey arrows represent the skip connections for feature fusion. Red arrows indicate the maximum pooling, used to reduce dimensions. Green arrows indicate the up-convolution, used to restore dimensions. Baby-blue arrows represent the 1 × 1 convolution, used to output the results [31].
The compiler of U-Net consists of multiple convolutional and pooling layers. In this article, the convolutional kernel size used in the structure of U-Net is 3 × 3, with a step size of 1. The pooling layer adopts maximum pooling, which means that the pooled value of the area is the maximum value of the area. The size of each pooling layer kernel is 2. The padding is 0, and the step size is 2.
A decoder is a process of restoring the original size of a feature map, which consists of convolution, up-sampling, and skip structures. U-Net uses bilinear interpolation for up-sampling, which has two independent variables. Linear interpolation is applied in both directions, using the original image matrix as the source matrix and the interpolated matrix as the target matrix. The up-sampling process, based on bilinear interpolation, is shown in Figure 7.
As shown in Figure 7, the horizontal and vertical axes represent the positions of the image pixels. f x , y represents the color or grayscale value of the pixel at position x , y . We know the coordinates and corresponding pixel values of red dots Q 11 ( x 1 , y 1 ) , Q 12 ( x 1 , y 2 ) , Q 21 ( x 2 , y 1 ) , Q 22 ( x 2 , y 2 ) , f Q 11 , f Q 12 ,   f Q 21 , and f Q 22 . Our goal is to obtain the pixel values of the unknown function f at the green point P ( x , y ) , which is f P . First, perform linear interpolation in the x direction to obtain f R 1 and f R 2 . f R 1 is the value of blue point R 1 ( x , y 1 ) obtained by performing one-dimensional linear interpolation in the x-axis direction based on f Q 11 , f Q 12 . f R 2 is the value of blue point R 2 ( x , y 2 ) obtained by performing one-dimensional linear interpolation in the x-axis direction based on f Q 21 , f Q 22 .
f ( R 1 ) x 2 x x 2 x 1 f ( Q 11 ) + x x 1 x 2 x 1 f ( Q 21 )
f ( R 2 ) x 2 x x 2 x 1 f ( Q 12 ) + x x 1 x 2 x 1 f ( Q 22 )
Then, perform linear interpolation in the y-direction to obtain the f P based on f R 1 , f R 2 .
f ( P ) y 2 y y 2 y 1 f ( R 1 ) + y y 1 y 2 y 1 f ( R 2 )
Overall, the result of bilinear interpolation is f x , y :
f ( x , y ) f ( Q 11 ) ( x 2 x 1 ) ( y 2 y 1 ) ( x 2 x ) ( y 2 y ) + f ( Q 21 ) ( x 2 x 1 ) ( y 2 y 1 ) ( x x 1 ) ( y 2 y ) + f ( Q 12 ) ( x 2 x 1 ) ( y 2 y 1 ) ( x 2 x ) ( y y 1 ) + f ( Q 22 ) ( x 2 x 1 ) ( y 2 y 1 ) ( x x 1 ) ( y y 1 )
In terms of skip structure, U-Net adopts the method of concatenating features from different levels in the channel dimension to form more “thick” features, which are the blue and white channels in Figure 5.

2.6. Spatial–Spectral Combination

Building upon previous research, this article utilized a spatial–spectral combination approach, leveraging the distinctive attributes of fragments and hyperspectral images. Following the spatial segmentation, the subsequent step involved identifying each segmented region using spectral information. This process enabled an accurate differentiation between the different types of targets and the background materials. The flow chart of the spatial–spectrum combination is shown in Figure 8.
This study obtained a binary image through U-Net segmentation, where black represented the background and white represented the target to be classified. We used the concept of connected domains to write algorithms that numbered all regions to be classified and extracted the coordinates of all pixels within each region. This created a mask based on these coordinates to cover the hyperspectral image and converted the 3D hyperspectral data of each region into 2D. Input the 2D hyperspectral data into a pre-trained CNN-BiLSTM model to classify each pixel in the region. By counting the categories of pixels within each region, the category of the region was determined by the categories of the majority of pixels within that region. Then, the region was colored based on the different classification results to obtain a spatial map of the fragment classification results.

2.7. Semantic Segmentation Datasets

To obtain the spatial distribution of fragments, this study employed a semantic segmentation model based on U-Net to accurately segment targets in explosive fragment dispersion fields. However, due to the limited number of experimental samples, it was challenging to train a semantic segmentation model with sufficient generalization ability. To address this issue, the study adopted the concept of transfer learning and utilized training datasets from two sources. The first part of the training dataset comprised 1000 images from the saliency object detection datasets that met the required criteria. The second part consisted of 124 images collected from laboratory experiments and online sources. Additionally, the study incorporated an extended dataset specifically created for this research. This dataset was expanded by applying various transformations, such as rotation, brightness adjustment, cropping, and other methods, resulting in a collection of 5620 images. The composition of the semantic segmentation dataset based on transfer learning is shown in Figure 9. These images were calibrated in the image label software included in MATLAB 2019b.

2.8. Hyperspectral Image of the Dispersion of Fragments

The spatial distribution and spectral data of the fragments could be obtained from the hyperspectral images of the dispersion of fragments. The hyperspectral image of the dispersion of fragments is shown in Figure 10.

2.9. Construction of Classifying Explosive Fragments

The overall procedure of the proposed framework for classifying explosive fragments is depicted in Figure 11. First, the spectral information classification model based on CNN-BiLSTM was trained. Ten types of hyperspectral data samples were collected in a laboratory environment, and RoIs were extracted from each sample type, resulting in 4000 sample points per type. These datasets were randomly divided into training sets, validation sets, and test sets. The training sets had 24,000 sample points, and the validation sets and testing sets both had 8000 sample points. The input data were fed into the CNN-BiLSTM model, and network parameters were adjusted to achieve the desired classification accuracy. The main parameters of the CNN-BiLSTM model are shown in Table 2. Second, a semantic segmentation model based on U-Net was trained. The U-Net model was trained using an extended dataset that included saliency object detection datasets, as well as images collected from laboratory experiments. We input the spatial image of scattered fragments to obtain the segmented binary image of the scattered fragments. The main parameters of the U-Net model are shown in Table 3. Finally, we input the spectral data of scattered fragments into the trained CNN-BiLSTM model to obtain the classification results. Then, the classification results were combined with the segmented binary image, where the category of the region was determined by the majority of pixels within that region.
From Table 2, 24,000 training sets were divided into 240 subsets, and every subset contains 100 samples. These subsets were called mini-batches, and the mini-batch size of the network was 100. During the training process of each subset, a gradient descent was performed, and 240 iterations were performed. The behavior of traversing all samples once was called an epoch, and the total epochs of the model were 80. An Adam loss function was selected to update the model parameters as the model optimizer. The number of filters was 20, and the size of the filter was 5*5. The number of hidden units was 100, which allowed the network to have strong computing power while preventing overfitting. Due to the experiment datasets having a total of 10 types of samples, the number of classes in the model was set to 10.
In Table 3, we aimed to segment both the background and target, constituting a total of 2 classes. The input image dimensions were 512 × 512 pixels. Each batch consisted of 5 samples, with training extending 50 epochs. The initial learning rate was set at 0.001, and the model optimization was handled by the Adam optimizer. When the momentum parameter was excessively high, it failed to effectively capture the overall data trend. Conversely, if it was too low, the model could overfit, leading to substantial fluctuations between adjacent data points. A momentum value of 0.9 struck a balance, enabling an accurate reflection of the data trend without excessive oscillations. In practice, momentum values around 0.9 were commonly employed. Throughout the training, the model weights and corresponding loss function values were periodically saved every five iterations. This approach facilitated monitoring changes in the loss function and segmentation accuracy over the training course, with a save interval of 5 in the model structure. Following the 235th training iteration, the model achieved its lowest loss value of 0.025, and the weights at this point were fed into the U-Net architecture. The training speed heavily relied on the number of workers, influencing data loading efficiency, which had been set at 4 in that scenario. The dataset comprised 3264 training samples and 816 validation samples, crucial for model evaluation and performance assessment.
In order to reduce the influence of random factors, all experiments were repeated ten times, and the average value was taken as the final experimental result. The running environment of the experimental code in this article is as follows. The CPU was a 13th Gen Intel(R) Core (TM) i9-13900K. The primary frequency was 3.00 GHz. The memory capacity was 128 GB. The GPU was an NVIDIA GeForce GTX 3090. The memory was 64 GB. The programming language was Python3.11, and the deep-learning architecture was PyTorch1.10.0.

3. Experimental Results and Analysis

3.1. Spectral Curve Analysis

There are 50 randomly selected representative sample points from the RoIs of each type of sample. The calculated average values of the spectral information for each category and obtained the spectral reflectance curve are shown in Figure 12.
From Figure 12, it could be seen that only the reflectance of the leaves had a significant difference compared to other species. In the visible light spectrum, chlorophyll was the main factor governing plant spectral response. Chlorophyll absorbed most of the incident energy in two spectral bands, with central wavelengths of 0.45 μm (blue) and 0.67 μm (red), respectively. Between these two chlorophyll absorption bands, chlorophyll had a relatively small absorption effect around 0.54 μm (green), forming a reflection peak, so our eyes saw the plant as green. In the near-infrared band, the spectral characteristics of vegetation were mainly influenced by the internal structure of the plant leaves. The spectral characteristic of healthy green plants in the near-infrared band was high reflectivity. Between the visible and near-infrared bands, around 0.76 μm, the reflectivity sharply increased, forming the so-called “red edge”, which was the most obvious feature of the plant curve. There were also certain differences in reflection intensity or overall trend for different fragments and other types of background materials. We could establish classification models for different fragments and background materials based on these differences.

3.2. Spectral Information Classification Model Results and Analysis

In this study, the classification accuracy was used as the evaluation metric for the model. The overall accuracy (OA) and average accuracy (AA) were used to evaluate the classification performance [40]. The OA is calculated as the number of samples predicted correctly divided by the total samples of the test set, as shown in Formula (6). AA is the ratio of the sum of classification accuracy for each class to the number of classes, as shown in Formula (7). The precision, recall, and F1 score are shown in Formulas (8)–(10), respectively. The above evaluation indicators are dimensionless. The range of these parameters is from zero to one, where one represents the best performance and zero represents the worst performance.
Overall   Accuracy = T P + T N F N + T P + F P + T N
Average   Accuracy = Overall   Accuracy N
Precision = T P T P + F P
Recall = T P F N + T P
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
TP denotes the number of samples labeled as positive and classified as positive. FP denotes the number of samples labeled as negative and classified as positive. TN denotes the number of samples labeled as negative and classified as negative. FN denotes the number of samples labeled as positive and classified as negative. N is the number of classes. Because the above variables represent different types of sample sizes, their dimensions are “units”.
In order to demonstrate the good performance of CNN-BiLSTM in fragment hyperspectral image classification, the experiment used decision trees, support vector machines (SVM), CNN, BiLSTM, and CNN-BiLSTM to classify 8000 samples in the test set. The confusion matrix was a visualization method that displays the performance of an algorithm in a specific matrix [41]. Each column in the matrix represents the predicted label, and each row represents the actual label. Since the number of samples in each category is 800, the OA is equal to the AA.
The confusion matrix results of the five-class classification models are shown in Figure 13. Labels 0, 1, 2, 3, 4, 5, 6, 7, 8, and 9, respectively, represent tungsten fragments, aluminum fragments, iron fragments, copper fragments, composite fragments, bark, soil, stone, foam, and leaves.
From Figure 13, the misjudgment rate between different types of fragments and background materials was relatively low. The classification between the fragments had a higher misjudgment rate, such as tungsten fragments and iron fragments. This finding indicated that distinguishing between non-metals was easier to achieve than between metals. Finally, the recognition rate of each class of samples and the AA are shown in Table 4. Bold numbers represent the highest accuracy.
According to Table 4, the accuracies of CNN, BiLSTM, and CNN-BiLSTM were significantly higher than that of decision trees and SVM. This result indicated that deep learning had significantly improved classification accuracy. The average accuracy of the CNN BiLSTM classification model reached 93.8%, surpassing decision trees, SVM, CNN, and BiLSTM. The CNN-BiLSTM model was only slightly inferior to the BiLSTM in bark sample classification and had the highest classification accuracy for other categories of samples. Compared to using CNN and BiLSTM separately, the CNN-BiLSTM model was more suitable for the spectral information classification of fragments. Among the three classification models, the classification accuracies for leaves and foam were particularly high, more than 99%. On the other hand, the classification accuracy of tungsten fragments and iron fragments was relatively low, but still above 80%. The precision, recall, and F1 scores of the different classification models are shown in Table 5.

3.3. Semantic Segmentation Results and Analysis

The model segmented hyperspectral spatial images with fragments scattered in the background, where the white part represented the target to be classified. The results are shown in Figure 14, where (a) was the segmentation map and (b) was the calibration map.
The intersection over union (IoU and true positivity ratio (TPR) were used to evaluate the semantic segmentation results. The IoU and TPR are dimensionless. The range of these parameters is from zero to one, where one represents the best performance and zero represents the worst performance. The calculation formulas are as follows.
I o U = T P F N + T P + F P
T P R = T P T P + F N
TP (true positive) is the intersection of the segmentation results and the calibration results. FP (false positive) is the over-segmentation part. FN (false negative) is the under-segmentation part [42]. Because the above variables represent different types of sample sizes, their dimensions are “units”. The evaluation results are shown in Table 6.
Since the IoU and TPR of this experiment were both more than 0.9, it indicated that the spatial segmentation method based on U-Net had achieved good results in the images.

3.4. Identification Results and Analysis of Explosive Fragments Based on the Spatial–Spectral Combination

After training on spectral reflectance classification, three deep-learning classification models were developed. Among these models, the CNN-BiLSTM model exhibited the highest classification accuracy. Consequently, this model was employed to classify the hyperspectral data of scattered fragments on a pixel-by-pixel basis. The resulting classification outcomes are presented in Figure 15.
In Figure 15, the different colors of the pixels represent various materials. Specifically, gray represented sand, black represented bark, green represented stones, yellow represented leaves, dark blue represented foam, brown represented composite fragments, orange represented tungsten fragments, blue represented aluminum fragments, and purple represented copper fragments.
Utilizing semantic segmentation, the classification results obtained from the CNN-BiLSTM model were further analyzed by counting the pixels in each region. The confusion matrix results of the CNN-BiLSTM models are shown in Figure 16. Labels 0, 1, 2, 3, 4, 5, 6, 7, 8, and 9, respectively, represented tungsten fragments, aluminum fragments, iron fragments, copper fragments, composite fragments, bark, soil, stone, foam, and leaves. The recognition of each sample class based on the CNN-BiLSTM model is shown in Table 7, and the results of the evaluation metrics based on CNN-BiLSTM are shown in Table 8.
The OA and AA of the samples were determined to be 87.2% and 85.8%, respectively. Among the different sample types, leaves exhibited the highest accuracy at 98.5%, while iron fragments had the lowest recognition accuracy at 76.2%. It should be noted that the misclassified pixels were primarily located at the edges, regions that were excessively bright or dark, and areas affected by shadows resulting from the angle of light. The number of different sample points varied in hyperspectral images. The number of pixels in the soil samples and bark samples was high, while the number of pixels in the aluminum fragment samples, iron fragment samples, and copper fragment samples was low. The reason for the low precision and F1 scores of aluminum fragments was that soil samples were misclassified as aluminum fragment samples. The reason for the low precision and F1 scores of iron and copper fragments was that the bark samples were misclassified into iron and copper fragments. Subsequently, the classification results based on the CNN-BiLSTM model were combined with the spatial segmentation map generated by U-Net. Clustering was performed based on the segmented areas, with the type of each area determined by the dominant pixel category. The recognition result is shown in Figure 17.
The ratio of the intersection points of algorithm segmentation and manual calibration to the calibration results was used to evaluate the final recognition results. The spatial–spectral combination recognition rate and the result of the evaluation index are shown in Table 9.
From Table 9, the recognition rate of all samples was higher than 94%, and the OA and AA of the samples were 96.4% and 96.3%, respectively. It showed a significant improvement compared to the recognition rate in the spectral domain.

4. Discussion

This study used hyperspectral imaging technology for fragments classification based on the differences in explosion fragments and their background spectral reflectance information. We collected ten samples of tungsten fragments, iron fragments, copper fragments, aluminum fragments, composite fragments, soil, stones, leaves, bark, and foam. In a darkroom environment, full-band hyperspectral images in the range of 400.00 to 1000.00 nm of ten types of samples and fragment simulation scattering scenes were collected. After preprocessing, a deep-learning method was used to build a spatial–spectral combination approach for fragment classification using hyperspectral imaging, aiming to address the requirements of fragment identification and spatial dispersion analysis in ex-plosive fields. In addition, some further discussions are presented below.
In our previous research, there was a problem of low accuracy in using decision trees to identify stones and iron fragments [19]. This was because the decision tree model did not extract appropriate features, and the model was susceptible to noise and outliers. From Table 4, we can see that the classification accuracy of CNN, BiLSTM, and CNN-BiLSTM based on deep learning is significantly higher than that of decision tree and SVM. This was because deep-learning algorithms can automatically capture the correlation between spectral features and obtain the intrinsic characteristics of explosive fragments. The accuracy of CNN was slightly lower than that of BiLSTM. The reason for this situation is that the spectral information of sampling points in hyperspectral images is typical sequential data. The BiLSTM is better at handling sequence-type data than the CNN. The CNN-BiLSTM combined the advantages of CNN and BiLSTM, with an accuracy rate of up to 93.8%. It proved that the method was beneficial for identifying explosive fragments using hyperspectral images.
In the application of hyperspectral image segmentation, most studies only treated segmentation as a part of preprocessing and did not delve into it in-depth [43]. Currently, most studies use machine-learning-based segmentation methods [44,45]. The segmentation method we used in our previous research was based on morphology [19]. The two methods required human involvement to some extent. Therefore, we trained a U-Net model that could automatically extract explosion fragment features for fragment and background object segmentation using the idea of transfer learning. The trained U-Net network achieved an accuracy of 93.5% in segmenting sample images. The results indicated that U-Net could effectively segment fragments in the background from other background objects and extract the spatial features of the samples.
This study used the spatial–spectral combination method to identify explosive fragments, which could classify and express targets in the scattering field, with a recognition rate of better than 94.9% for each category. Compared to using the CNN-BiLSTM model to classify each pixel in the scattering field, the spatial–spectral combination algorithm significantly improved the fragment recognition rate. Compared with previous studies, we found that the algorithm used in this article greatly reduced human involvement [19]. In recent years, many studies have focused on how to fully utilize the spatial and spectral information contained in hyperspectral images and have also proposed many novel algorithms, such as the random patches network (RPNet) [46], etc. However, these algorithms are usually used in publicly available hyperspectral datasets and are rarely applied in practical fields. We will study relevant algorithms in the future and truly apply novel algorithms to the field of explosive fragment recognition.
As above, the fragments classification method used in this article could automatically extract features in both the spatial and spectral domains and perform spatial–spectral combinations through a cascading approach. After experimental verification, it was feasible to use the method of spatial–spectral combination with hyperspectral imaging technology for the classification of explosive fragments.

5. Conclusions

This paper introduces a spatial–spectral combination approach for fragment classification using hyperspectral imaging, aiming to address the requirements of fragment identification and spatial dispersion analysis in explosive fields. Our work can be divided into three main aspects.
(1)
This study used a hyperspectral imaging system to collect data on fragments and backgrounds. After denoising, the regions of interest were extracted and sample points were randomly selected in each region as the spectral information dataset for this project. Random sampling is used to divide the sample points into a training set, a validation set, and a testing set. By observing the spectral reflectance curves of explosion fragments and background samples, it can be found that there are spectral band differences in the spectral curves of different types of samples. A semantic segmentation training dataset was collected and established based on the characteristics of the simulated explosion fragment dispersion field in this study. Finally, we expanded and calibrated the spatial segmentation dataset;
(2)
We address the limitation of CNN in considering the relationship between current data and its neighboring data, which significantly impacts classification accuracy. To overcome this, this study adopts the CNN-BiLSTM model as the spectral information classification model. The experimental results demonstrate the ability of CNN-BiLSTM to leverage the advantages of both CNN and BiLSTM, resulting in high accuracy for hyperspectral image classification even with a limited number of training samples. To further enhance the model’s performance, dropout layers are introduced to mitigate overfitting. In addition, we incorporate the U-Net semantic segmentation model to extract spatial relationships between pixels from hyperspectral images. This approach enables the precise segmentation of spatial images containing fragments, achieving high accuracy in the process;
(3)
A classification framework based on the spatial–spectral combination is proposed. In this way, U-Net is combined with CNN-BiLSTM, which can fully merge the spatial and spectral information of hyperspectral images to achieve high classification accuracy. The experimental results demonstrate that the accuracy of fragment recognition based on the spatial–spectral combination is substantially higher compared to the accuracy achieved through spectral-based recognition alone. It indicates the feasibility of using hyperspectral detection methods for fragment recognition based on the spatial–spectral combination methods, providing an experimental and theoretical basis for the recognition of fragments.
The purpose of this study is to verify the feasibility of using hyperspectral imaging to identify explosive fragments. The relevant algorithms are currently only applicable for explosive fragment recognition in laboratory darkroom environments. We will test the accuracy of recognition in outdoor environments in the future. In addition, in our next research, we will identify different environments and unique features of fragments by combining multidimensional data. Due to the wide range of fragments scattered at the scene after the explosion, we will use airborne hyperspectral imaging to image the explosion fragments in the future. Considering the challenges posed by the complex light source and small fragment volume at the explosion site, our research will delve into spectral information correction models and explore more fitting classification algorithms. Due to the possibility of fragments being obstructed by soil or vegetation, we will apply more imaging methods, such as thermal infrared imaging and synthetic aperture radar (SAR) imaging with hyperspectral imaging, for image fusion in explosive fragment recognition in the future.

Author Contributions

Conceptualization, D.Z.; Data curation, K.L.; Formal analysis, D.Z., Y.M., C.W. and B.Z.; Funding acquisition, D.Z., W.C. and B.Z.; Methodology, F.G. and X.Y.; Project administration, Y.M. and W.C.; Resources, C.W. and K.L.; Software, P.Y.; Supervision, X.Y.; Validation, P.Y. and F.G.; Visualization, P.Y.; Writing—original draft, F.G.; Writing—review and editing, P.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Nos. 62205307), the National Natural Science Foundation of China Youth Science Fund Project (Nos. 62005254), and the Shanxi Province Basic Research Program (Nos. 202203021212113).

Data Availability Statement

The data are not publicly available due to the confidentiality of the research projects.

Acknowledgments

The author thanks the Northwest Institute of Nuclear Technology for providing fragment samples and the foundation’s funding.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNN-BiLSTMConvolutional Neural Network–Bidirectional Long Short-Term Memory Network
U-NetU-shaped Network
CNNConvolutional Neural Network
1DOne Dimensional
RNNRecurrent Neural Network
BiLSTMBidirectional Long Short-Term Memory Network
RoIRegions of Interest
OAOverall Accuracy
AAAverage Accuracy
SVMSupport Vector Machines
IoUIntersection over Union
TPRTrue Positivity Ratio
TPTrue Positive
FPFalse Positive
FNFalse Negative
RPNet Random Patches Network
FSKNetFaster Selective Kernel Mechanism Network
SARSynthetic Aperture Radar

References

  1. Li, X.; Li, F.; Liang, M.; Zhang, K. Design Methods for the Lethality of Fragmentation Warhead based on a Multi-Objective Optimization Algorithm. J. Phys. Conf. Ser. 2022, 2219, 012034. [Google Scholar] [CrossRef]
  2. Li, H.S.; Hao, Y.; Zhang, X.Q. Numerical Calculation Method of Target Damage Effectiveness Evaluation Under Uncertain Information of Warhead Fragments. Mathematics 2022, 10, 1688. [Google Scholar] [CrossRef]
  3. Li, H.; Zhang, X.; Zhang, X. Calculation Model and Method of Target Damage Efficiency Assessment Based on Warhead Fragment Dispersion. IEEE Trans. Instrum. Meas. 2021, 70, 1–8. [Google Scholar] [CrossRef]
  4. Yi, J.Y.; Xie, Y.; Yin, J.P.; Li, X.D. Study on the Rule of Fragmentation Dispersion of Warhead at Different Depths on the Ground. Strength Mater. 2022, 54, 695–704. [Google Scholar] [CrossRef]
  5. Zhang, X.; Li, W.; Wang, X. Numerical Simulation of Rigid Projectiles Penetration into Concrete Targets with Different Warhead Shapes. J. Phys. Conf. Ser. 2021, 2002, 012035. [Google Scholar] [CrossRef]
  6. Li, Y.M.; Li, W.B.; Liu, J.H.; Pu, B.; Zhang, Q.; Wang, X.M. Method of evaluating energy released by a warhead charge based on average fragment quality. Lat. Am. J. Solids Struct. 2021, 18, e372. [Google Scholar] [CrossRef]
  7. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in Spectral-Spatial Classification of Hyperspectral Images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Li, W.; Sun, W.; Tao, R.; Du, Q. Single-Source Domain Expansion Network for Cross-Scene Hyperspectral Image Classification. IEEE Trans. Image Process. 2023, 32, 1498–1512. [Google Scholar] [CrossRef]
  9. Yan, L.; Zhao, M.; Wang, X.; Zhang, Y.; Chen, J. Object Detection in Hyperspectral Images. IEEE Signal Process. Lett. 2021, 28, 508–512. [Google Scholar] [CrossRef]
  10. Zhu, M.; Huang, D.; Hu, X.J.; Tong, W.H.; Han, B.L.; Tian, J.P.; Luo, H.B. Application of hyperspectral technology in detection of agricultural products and food: A Review. Food Sci. Nutr. 2020, 8, 5206–5214. [Google Scholar] [CrossRef]
  11. Gao, Y.; Li, W.; Zhang, M.; Wang, J.; Sun, W.; Tao, R.; Du, Q. Hyperspectral and Multispectral Classification for Coastal Wetland Using Depthwise Feature Interaction Network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  12. Audebert, N.; Saux, B.L.; Lefevre, S. Deep Learning for Classification of Hyperspectral Data: A Comparative Review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef]
  13. Ma, J.; Sun, D.W.; Pu, H.B.; Cheng, J.H.; Wei, Q.Y. Advanced Techniques for Hyperspectral Imaging in the Food Industry: Principles and Recent Applications. Annu. Rev. Food Sci. Technol. 2019, 10, 197–220. [Google Scholar] [CrossRef] [PubMed]
  14. Le, Y.; Hu, M.Z.; Chen, Y.F.; Yan, Q.; Zhang, D.F.; Li, S.; Zhang, X.H.; Wang, L.Z. Investigating the Shallow-Water Bathymetric Capability of Zhuhai-1 Spaceborne Hyperspectral Images Based on ICESat-2 Data and Empirical Approaches: A Case Study in the South China Sea. Remote Sens. 2022, 14, 3406. [Google Scholar] [CrossRef]
  15. Zhao, J.L.; Zhou, B.; Wang, G.L.; Liu, J.; Ying, J.J. Camouflage Target Recognition Based on Dimension Reduction Analysis of Hyperspectral Image Regions. Photonics 2022, 9, 640. [Google Scholar] [CrossRef]
  16. Wang, P.; Zhang, G.; Wang, L.; Leung, H.; Bi, H. Subpixel Land Cover Mapping Based on Dual Processing Paths for Hyperspectral Image. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1835–1848. [Google Scholar] [CrossRef]
  17. Gao, H.; Yang, Y.; Li, C.; Zhang, X.; Zhao, J.; Yao, D. Convolutional neural network for spectral–spatial classification of hyperspectral images. Neural Comput. Appl. 2019, 31, 8997–9012. [Google Scholar] [CrossRef]
  18. Lee, H.; Kwon, H. Contextual deep CNN based hyperspectral classification. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 3322–3325. [Google Scholar]
  19. Guo, F.; Zhao, D.E.; Yang, X.F.; Chu, W.B.; Zhang, B.; Zhang, D.S.; Meng, F.J. Research on Hyperspectral Image Recognition of Iron Fragments. Spectrosc. Spectr. Anal. 2023, 43, 997–1003. [Google Scholar]
  20. Gan, Y.; Shi, J.-c.; He, W.-m.; Sun, F.-j. Parallel classification model of arrhythmia based on DenseNet-BiLSTM. Biocybern. Biomed. Eng. 2021, 41, 1548–1560. [Google Scholar] [CrossRef]
  21. Chen, Y.; Zhao, X.; Jia, X. Spectral–Spatial Classification of Hyperspectral Data Based on Deep Belief Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  22. Hu, W.; Huang, Y.Y.; Wei, L.; Zhang, F.; Li, H.C. Deep Convolutional Neural Networks for Hyperspectral Image Classification. J. Sens. 2015, 2015, 258619. [Google Scholar] [CrossRef]
  23. Meedeniya, D.A.; Mahakalanda, I.; Lenadora, D.S.; Perera, I.; Hewawalpita, S.G.S.; Abeysinghe, C.; Nayak, S.R. Chapter 13—Prediction of paddy cultivation using deep learning on land cover variation for sustainable agriculture. In Deep Learning for Sustainable Agriculture; Poonia, R.C., Singh, V., Nayak, S.R., Eds.; Academic Press: Cambridge, MA, USA, 2022; pp. 325–355. [Google Scholar]
  24. Zhao, W.; Du, S. Spectral–Spatial Feature Extraction for Hyperspectral Image Classification: A Dimension Reduction and Deep Learning Approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  25. Shu, Z.; Wang, Y.; Yu, Z. Dual attention transformer network for hyperspectral image classification. Eng. Appl. Artif. Intell. 2024, 127, 107351. [Google Scholar] [CrossRef]
  26. Andrushia, A.D.; Neebha, T.M.; Patricia, A.T.; Sagayam, K.M.; Pramanik, S. Capsule network-based disease classification for Vitis Vinifera leaves. Neural Comput. Appl. 2024, 36, 757–772. [Google Scholar] [CrossRef]
  27. Xiong, P.; Xue, Y.; Zhang, J.; Liu, M.; Du, H.; Zhang, H.; Hou, Z.; Wang, H.; Liu, X. Localization of myocardial infarction with multi-lead ECG based on DenseNet. Comput. Methods Programs Biomed. 2021, 203, 106024. [Google Scholar] [CrossRef]
  28. Xu, X.; Jeong, S.; Li, J. Interpretation of Electrocardiogram (ECG) Rhythm by Combined CNN and BiLSTM. IEEE Access 2020, 8, 125380–125388. [Google Scholar] [CrossRef]
  29. Zhao, H.; Deng, K.; Li, N.; Wang, Z.; Wei, W. Hierarchical Spatial-Spectral Feature Extraction with Long Short Term Memory (LSTM) for Mineral Identification Using Hyperspectral Imagery. Sensors 2020, 20, 6854. [Google Scholar] [CrossRef]
  30. Yang, M.T.; Liu, W.C.; Zhang, H.G. A robust multiple heartbeats classification with weight-based loss based on convolutional neural network and bidirectional long short-term memory. Front. Physiol. 2022, 13, 982537. [Google Scholar] [CrossRef]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  32. Fan, X.; Ding, W.; Li, X.; Li, T.; Hu, B.; Shi, Y. An Improved U-Net Infrared Small Target Detection Algorithm Based on Multi-Scale Feature Decomposition and Fusion and Attention Mechanism. Sensors 2024, 24, 4227. [Google Scholar] [CrossRef]
  33. Azad, R.; Aghdam, E.K.; Rauland, A.; Jia, Y.; Avval, A.H.; Bozorgpour, A.; Karimijafarbigloo, S.; Cohen, J.P.; Adeli, E.; Merhof, D. Medical Image Segmentation Review: The Success of U-Net. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 1–20. [Google Scholar] [CrossRef]
  34. Li, X.Y.; Raga, R.C.; Shi, X.M. GloVe-CNN-BiLSTM Model for Sentiment Analysis on Text Reviews. J. Sens. 2022, 2022, 7212366. [Google Scholar] [CrossRef]
  35. Zhang, X.S.; Ma, Y.L.; Wang, M.H. An attention-based Logistic-CNN-BiLSTM hybrid neural network for credit risk prediction of listed real estate enterprises. Expert Syst. 2024, 41, e13299. [Google Scholar] [CrossRef]
  36. Sun, S.; Sun, J.Z.; Wang, Z.L.; Zhou, Z.Y.; Cai, W. Prediction of Battery SOH by CNN-BiLSTM Network Fused with Attention Mechanism. Energies 2022, 15, 4428. [Google Scholar] [CrossRef]
  37. Çevik Bektaş, S.; Altaş, I.H. DWT-BILSTM-based models for day-ahead hourly global horizontal solar irradiance forecasting. Neural Comput. Appl. 2024, 36, 13243–13253. [Google Scholar] [CrossRef]
  38. Yang, J.; Liu, J.; Guo, J.; Tao, K. Track Irregularity Identification Method of High-Speed Railway Based on CNN-Bi-LSTM. Sensors 2024, 24, 2861. [Google Scholar] [CrossRef]
  39. Liu, L.; Cheng, J.; Quan, Q.; Wu, F.-X.; Wang, Y.-P.; Wang, J. A survey on U-shaped networks in medical image segmentations. Neurocomputing 2020, 409, 244–258. [Google Scholar] [CrossRef]
  40. Hay, A.M. The derivation of global estimates from a confusion matrix. Int. J. Remote Sens. 1988, 9, 1395–1398. [Google Scholar] [CrossRef]
  41. Liu, B.; Zhang, H.; Zhu, J.; Chen, Y.; Pan, Y.; Gong, X.; Yan, J.; Zhang, H. Pixel-Level Recognition of Trace Mycotoxins in Red Ginseng Based on Hyperspectral Imaging Combined with 1DCNN-Residual-BiLSTM-Attention Model. Sensors 2024, 24, 3457. [Google Scholar] [CrossRef]
  42. Hao, S.; Zhou, Y.; Guo, Y. A Brief Survey on Semantic Segmentation with Deep Learning. Neurocomputing 2020, 406, 302–321. [Google Scholar] [CrossRef]
  43. Bhargava, A.; Sachdeva, A.; Sharma, K.; Alsharif, M.H.; Uthansakul, P.; Uthansakul, M. Hyperspectral imaging and its applications: A review. Heliyon 2024, 10, e33208. [Google Scholar] [CrossRef]
  44. Xue, H.; Xu, X.; Meng, X. Estimation Model for Maize Multi-Components Based on Hyperspectral Data. Sensors 2024, 24, 6111. [Google Scholar] [CrossRef] [PubMed]
  45. Ghezal, A.; Peña, C.J.; König, A. Varroa Mite Counting Based on Hyperspectral Imaging. Sensors 2024, 24, 4437. [Google Scholar] [CrossRef] [PubMed]
  46. Xu, Y.; Du, B.; Zhang, F.; Zhang, L. Hyperspectral image classification via a random patches network. ISPRS J. Photogramm. Remote Sens. 2018, 142, 344–357. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of hyperspectral image acquisition system.
Figure 1. Schematic diagram of hyperspectral image acquisition system.
Sensors 24 07131 g001
Figure 2. Hyperspectral image acquisition system. (a) Actual scene diagram; (b) darkroom environment diagram.
Figure 2. Hyperspectral image acquisition system. (a) Actual scene diagram; (b) darkroom environment diagram.
Sensors 24 07131 g002
Figure 3. True-color images of the collected fragments and background samples. (a) Tungsten fragments; (b) copper fragments; (c) iron fragments; (d) aluminum fragments; (e) composite fragments; (f) stone samples; (g) soil samples; (h) leaf samples; (i) foam samples; (j) bark samples.
Figure 3. True-color images of the collected fragments and background samples. (a) Tungsten fragments; (b) copper fragments; (c) iron fragments; (d) aluminum fragments; (e) composite fragments; (f) stone samples; (g) soil samples; (h) leaf samples; (i) foam samples; (j) bark samples.
Sensors 24 07131 g003aSensors 24 07131 g003b
Figure 4. Region of interest extraction and data conversion of bark samples.
Figure 4. Region of interest extraction and data conversion of bark samples.
Sensors 24 07131 g004
Figure 5. Model structure of CNN-BiLSTM.
Figure 5. Model structure of CNN-BiLSTM.
Sensors 24 07131 g005
Figure 6. U-Net network structure diagram.
Figure 6. U-Net network structure diagram.
Sensors 24 07131 g006
Figure 7. Up-sampling process based on bilinear interpolation.
Figure 7. Up-sampling process based on bilinear interpolation.
Sensors 24 07131 g007
Figure 8. Flow chart of the spatial–spectrum combination.
Figure 8. Flow chart of the spatial–spectrum combination.
Sensors 24 07131 g008
Figure 9. Composition of semantic segmentation dataset based on transfer learning.
Figure 9. Composition of semantic segmentation dataset based on transfer learning.
Sensors 24 07131 g009
Figure 10. Hyperspectral image of the dispersion of fragments.
Figure 10. Hyperspectral image of the dispersion of fragments.
Sensors 24 07131 g010
Figure 11. Overall procedure of the proposed classification of hyperspectral images of the explosive fragments framework including data augmentation, spectral information classification model based on CNN-BiLSTM, semantic segmentation model based on U-Net, and spatial–spectral combination method.
Figure 11. Overall procedure of the proposed classification of hyperspectral images of the explosive fragments framework including data augmentation, spectral information classification model based on CNN-BiLSTM, semantic segmentation model based on U-Net, and spatial–spectral combination method.
Sensors 24 07131 g011
Figure 12. Full-band reference spectral curve of samples.
Figure 12. Full-band reference spectral curve of samples.
Sensors 24 07131 g012
Figure 13. Confusion matrix results of the five-class classification models. (a) Decision tree; (b) SVM; (c) CNN; (d) BiLSTM; (e) CNN-BiLSTM.
Figure 13. Confusion matrix results of the five-class classification models. (a) Decision tree; (b) SVM; (c) CNN; (d) BiLSTM; (e) CNN-BiLSTM.
Sensors 24 07131 g013
Figure 14. Semantic segmentation results and calibration results. (a) Segmentation results; (b) calibration results.
Figure 14. Semantic segmentation results and calibration results. (a) Segmentation results; (b) calibration results.
Sensors 24 07131 g014
Figure 15. Classification results of CNN-BiLSTM model.
Figure 15. Classification results of CNN-BiLSTM model.
Sensors 24 07131 g015
Figure 16. Confusion matrix results of the CNN-BiLSTM models.
Figure 16. Confusion matrix results of the CNN-BiLSTM models.
Sensors 24 07131 g016
Figure 17. Results of spatial–spectral combination.
Figure 17. Results of spatial–spectral combination.
Sensors 24 07131 g017
Table 1. Main parameters of experimental system.
Table 1. Main parameters of experimental system.
System CompositionParameterValues
Hyperspectral cameraSpectral range400–1000 nm
Spectral channels923
Spatial channels1600
Spectral resolution2.5 nm
Pixel size6.5 μm
Slit width25 μm
Maximum frame rate100 fps
Detector TypeCCD
LensFocal length23 mm
Field of view25°
Halogen lampPower150 W (adjustable)
IlluminationConcave aluminum surface reflection
Reference whiteboardReflectivity>99%
MaterialPolytetrafluoroethylene (PTFE)
Mobile loading platformLength of rail250 mm
Step precision0.5 μm
Table 2. Main parameters of spectral information classification model based on CNN-BiLSTM.
Table 2. Main parameters of spectral information classification model based on CNN-BiLSTM.
Main ParameterValues
Mini Batch Size100
Number of Hidden Units100
Filter Size5
Number of Filters20
Epochs80
Optimizer typeAdam
Number of classes10
Number of test sets8000
Number of train sets24,000
Number of validation sets24,000
Table 3. Main parameters of U-Net training model.
Table 3. Main parameters of U-Net training model.
Main ParameterValues
Num classes2
Input shape(512, 512)
Epoch50
Batch size5
Initial learning rate0.0001
Optimizer typeAdam
momentum0.9
Save period5
Number of workers4
Number of train sets3264
Number of validation sets816
Table 4. Accuracy statistics of different model classification test sets.
Table 4. Accuracy statistics of different model classification test sets.
ModelAccuracy, %
TungstenAluminumIronCopperCompositeBarkSoilStoneFoamLeavesAverage
Decision Tree76.491.079.390.983.194.988.389.197.910089.1
SVM78.192.180.192.983.795.388.389.998.710089.9
CNN81.495.884.194.986.695.589.894.898.710092.1
BiLSTM84.595.986.494.889.995.890.895.698.710093.2
CNN-BiLSTM85.695.987.894.991.595.392.196.099.110093.8
Table 5. Precision, recall, and F1 scores for different classification models.
Table 5. Precision, recall, and F1 scores for different classification models.
ModelIndexTungstenAluminumIronCopperCompositeBarkSoilStoneFoamLeaves
Decision TreePrecision0.7710.9500.7950.8740.8970.8820.8520.8990.9941.000
Recall0.7640.9100.7930.9090.8310.9490.8830.8910.9791.000
F1 score0.7670.9300.7940.8910.8630.9140.8670.8950.9861.000
SVMPrecision0.7900.9500.8120.8800.9040.8910.8640.9070.9941.000
Recall0.7810.9210.8010.9290.8360.9530.8830.8990.9861.000
F1 score0.7850.9350.8060.9040.8690.9210.8730.9030.9901.000
CNNPrecision0.8340.9620.8440.8110.9350.8940.9220.9710.9761.000
Recall0.8140.9580.8410.9490.8660.9550.8980.9480.9861.000
F1 score0.8240.9600.8420.8750.8990.9230.9100.9590.9811.000
BiLSTMPrecision0.8570.9650.8650.9150.9360.9130.9250.9700.9761.000
Recall0.8450.9590.8640.9480.8990.9580.9080.9560.9861.000
F1 score0.8510.9620.8640.9310.9170.9350.9160.9630.9811.000
CNN-BiLSTMPrecision0.8670.9620.8740.9380.9370.9230.9310.9710.9771.000
Recall0.8550.9590.8780.9490.9150.9530.9210.9600.9911.000
F1 score0.8610.9600.8670.9430.9260.9380.9260.9650.9841.000
Table 6. Semantic segmentation results.
Table 6. Semantic segmentation results.
Evaluation IndexIoUTPR
Values0.9350.974
Table 7. Classification accuracy of the CNN-BiLSTM models.
Table 7. Classification accuracy of the CNN-BiLSTM models.
SampleSizeSample Point ClassificationAccuracy, %
TungstenAluminumIronCopperCompositeBarkSoilStoneFoamLeaves
Tungsten19,09614,589012549880018740039176.4
Aluminum46,835295837,26261235900398516590079.6
Iron131,20110,398125100,03412350115611,50636230312476.2
Copper71,81498212411266,20418784300500092.2
Composite19,90983300134216,3840997251010282.3
Bark306,00311,98410,897000255,24718,87590000083.4
Soil2,902,425188,976337195,89494,68520674892,502,160867253144186.2
Stone234,432189209002384800138,98205,7920087.8
Foam83,5950192816880000079,979095.7
Leaves424,66884500003943176500418,12498.5
Table 8. Evaluation index of CNN-BiLSTM.
Table 8. Evaluation index of CNN-BiLSTM.
IndexTungstenAluminumIronCopperCompositeBarkSoilStoneFoamLeaves
Precision0.8130.6690.4800.5610.9770.9530.9780.8990.9930.990
Recall0.7640.7960.7620.9220.8230.8340.8620.8780.9570.985
F1 score0.7880.7270.5890.6980.8930.8900.9160.8880.9750.987
OA, %87.2
AA, %85.8
Table 9. Statistical table of the results of the spatial–spectral combination.
Table 9. Statistical table of the results of the spatial–spectral combination.
SamplesTungstenAluminumIronCopperCompositeBarkSoilStoneFoamLeaves
Test sample points19,09646,835131,20171,81419,909306,0032,902,425234,43283,595424,668
Identify points18,59045,274126,14969,16318,993298,4272,795,035222,46179,567415,112
Accuracy, %97.496.796.196.395.497.596.394.995.297.5
OA, %96.4
AA, %96.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, D.; Yu, P.; Guo, F.; Yang, X.; Ma, Y.; Wang, C.; Li, K.; Chu, W.; Zhang, B. Classification of Hyperspectral Images of Explosive Fragments Based on Spatial–Spectral Combination. Sensors 2024, 24, 7131. https://doi.org/10.3390/s24227131

AMA Style

Zhao D, Yu P, Guo F, Yang X, Ma Y, Wang C, Li K, Chu W, Zhang B. Classification of Hyperspectral Images of Explosive Fragments Based on Spatial–Spectral Combination. Sensors. 2024; 24(22):7131. https://doi.org/10.3390/s24227131

Chicago/Turabian Style

Zhao, Donge, Peiyun Yu, Feng Guo, Xuefeng Yang, Yayun Ma, Changli Wang, Kang Li, Wenbo Chu, and Bin Zhang. 2024. "Classification of Hyperspectral Images of Explosive Fragments Based on Spatial–Spectral Combination" Sensors 24, no. 22: 7131. https://doi.org/10.3390/s24227131

APA Style

Zhao, D., Yu, P., Guo, F., Yang, X., Ma, Y., Wang, C., Li, K., Chu, W., & Zhang, B. (2024). Classification of Hyperspectral Images of Explosive Fragments Based on Spatial–Spectral Combination. Sensors, 24(22), 7131. https://doi.org/10.3390/s24227131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop