Next Article in Journal
Electrospun PEO/PEDOT:PSS Nanofibers for Wearable Physiological Flex Sensors
Next Article in Special Issue
A Rail-Temperature-Prediction Model Based on Machine Learning: Warning of Train-Speed Restrictions Using Weather Forecasting
Previous Article in Journal
Parameter Design and Performance Evaluation of a Large-Swath and High-Resolution Space Camera
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rock Particle Motion Information Detection Based on Video Instance Segmentation

1
School of Electrical and Information Engineering, Changsha University of Science and Technology, Changsha 410114, China
2
Xiaoxiang Research Institute of Big Data, Changsha 410199, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(12), 4108; https://doi.org/10.3390/s21124108
Submission received: 19 April 2021 / Revised: 2 June 2021 / Accepted: 8 June 2021 / Published: 15 June 2021
(This article belongs to the Special Issue AI-Oriented Sensing for Civil Engineering Applications)

Abstract

:
The detection of rock particle motion information is the basis for revealing particle motion laws and quantitative analysis. Such a task is crucial in guiding engineering construction, preventing geological disasters, and verifying numerical models of particles. We propose a machine vision method based on video instance segmentation (VIS) to address the motion information detection problem in rock particles under a vibration load. First, we designed a classification loss function based on Arcface loss to improve the Mask R-CNN. This loss function introduces an angular distance based on SoftMax loss that distinguishes the objects and backgrounds with higher similarity. Second, this method combines the abovementioned Mask R-CNN and Deep Simple Online and Real-time Tracking (Deep SORT) to perform rock particle detection, segmentation, and tracking. Third, we utilized the equivalent ellipse characterization method for segmented particles, integrating with the proportional calibration algorithm to test the translation and detecting the rotation by calculating the change in the angle of the ellipse’s major axis. The experimental results show that the improved Mask R-CNN obtains an accuracy of 93.36% on a self-created dataset and also has some advantages on public datasets. Combining the improved Mask R-CNN and Deep SORT could fulfill the VIS with a low ID switching rate while successfully detecting movement information. The average detection errors of translation and rotation are 5.10% and 14.49%, respectively. This study provides an intelligent scheme for detecting movement information of rock particles.

1. Introduction

Rock particles are common in geological disasters and are a type of material that is widely adopted in construction and transportation. The development of a general theory of the motion of granular materials is one of the 125 scientific frontier issues of Science. Revealing the movement laws and quantitatively analyzing rock particles in a moving state can provide reasonable guidance for the construction process and to improve construction efficiency [1]. It can also offer meso-analysis related to the occurrence of geological disasters, which can assist in their prevention [2,3]. In addition, with the research and application of numerical simulation methods (e.g., finite element method and discrete element method) in geotechnical engineering, the lack of effective verification data is a principal reason for restricting their development [4,5,6]. The motion information detection of rock particles not only is the basis of revealing their motion laws and quantitative analysis, but also can provide reliable data for particle numerical models, which is of great value in engineering and scientific research [7].
The detection of particle motion information is done using video instance segmentation (VIS). This means detecting, segmenting, and tracking instances simultaneously [8]. It is more challenging than image instance segmentation in that it not only requires instance segmentation on individual frames, but also the tracking of instances across frames.
In this study, we focused on detecting the motion information of rock particles under a vibration load and proposed a VIS-based method. This method combines the current mainstream instance segmentation and object tracking algorithms. Experimental equipment was established to verify the proposed method, and the effectiveness of our technique was demonstrated through experiments. We summarize them as follows:
  • Because the objects and backgrounds have a high degree of similarity, we improved the Mask R-CNN to develop the feature recognition ability and the instance segmentation effect of particles by designing a classification loss function based on Arcface loss.
  • The multiparticle VIS with a low ID switching rate was achieved by combining the improved Mask R-CNN and SORT.
  • We utilized the equivalent ellipse characterization method to process the segmented particles, combining characterization results with the proportional calibration algorithm to detect its translation and detecting the rotation by calculating the change in the angle of the major axis of the ellipse.
  • We verified the effectiveness of the method through segmentation experiments, VIS, and motion information detection. The experimental results showed that the improved Mask R-CNN had a better detection and segmentation effect. The ID switching rate of the VIS was low. This method can successfully detect the movement information of rock particles. The average detection errors of translation and rotation are 5.10% and 14.49%, respectively.

2. Related Work

Few studies on particle VIS have been conducted. However, particle segmentation as the basis of VIS has been extensively researched with the development of machine vision. In addition, there is substantial work on image-based deformation and stress measurement.

2.1. Traditional Particle Segmentation Method

The main traditional methods have relied on regional information and then combined with related traditional algorithms to accomplish particle segmentation. With the operation of edge connection, Yen et al. [9] used the Canny edge detection algorithm to detect particle images so that the edge gaps were changed to multiple closed regions. Zhang et al. [10] combined the bi-windows and maximum between-class variance (OSTU) to obtain binary images. Then, the distance transformation principle was employed to obtain the best seed area with a higher gray value, and a Watershed algorithm based on markers was applied to segment the particles. Amankwah et al. [11] proposed a method for segmenting particles, which utilized the mean shift algorithm to identify pixel clusters of particular modes from the probability density function of the image data. The pixel clusters were then used to generate markers for the watershed transform and shadow areas in images. However, the particle images often have characteristics of density. The abovementioned methods cannot accurately extract particle features, and the segmentation over-relies on manual selection of features.

2.2. Particle Segmentation Method Based on Deep Learning

Deep learning [12] has several advantages in image processing and, therefore, has been used in particle segmentation tasks. Yuan et al. [13] focused on solving the inaccuracy problem caused by mutual adhesion and shadows in the ore images using advanced ore image segmentation method based on the holistically-nested edge detection method. Compared with the traditional algorithm, their proposed method was more robust. Duan et al. [14] designed a lightweight U-net deep learning network to automatically detect particles from images and to obtain the probability maps of particle contours. This method could be applied to particle product quality monitoring. Liu et al. [15] proposed an image segmentation method based on the U-net and Res-Unet networks. The proposed method preprocessed the original images captured from an open-pit mine to reduce noise and extract the object region with grayscale, median filter, and adaptive histogram using equalization techniques. The abovementioned deep learning methods are mainly used for the semantic segmentation of particles that can automatically realize particle feature extraction with higher accuracy. These methods have become the main research direction for particle segmentation. However, because semantic segmentation can only acquire the complete information of all particles, it is difficult to analyze the meso-mechanism on a single-particle scale.

2.3. The Image-Based Measurement of Deformation and Stresses

Over the years, many visual sensors have been used in the monitoring and prediction of natural disasters (e.g., landslides, slope failure, and rockfalls) because of their low cost compared to expensive terrestrial laser scanner (TLS) or Interferometric Synthetic Aperture Radar (InSAR) [16,17]. Concurrently, image-based measurement of deformation and stresses were also being studied extensively. The most widely known method is particle image velocimetry (PIV) [18]. The PIV can combine a range of advances in image analysis algorithms (e.g., image intensity interpolation [19] and deformation parameter optimization [20]) that are best suited to geotechnical applications, which would be faster and more precise than the standard PIV approach [21]. However, these image-based methods could not analyze challenging scenes, such as images with insufficient lighting and fast-moving objects [22]. However, many related complementary methods have been studied and applied. Gance et al. [23] proposed a target detection and tracking (TDT) method that was based on simple binary image processing for the analyses of long time series. This method can be used as a complement to image correlation and other displacement observation methods for rapid assessments. In terms of moving-object tracking methods, feature tracking, color feature tracking, and outline tracking have proven to be effective tracking methods [24]. However, 3D reconstruction technology has some advantages compared to feature tracking because it also approximates the 3D shape of an object using a 3D mesh. Guccione et al. [25] used four high-speed cameras and two tilted mirrors for 3D reconstruction to track the objects. Then, the 3D rotational velocities were estimated accurately using a new postprocessing algorithm. The abovementioned devices and methods have made important contributions to meso-analysis and have helped in preventing geological disasters.

3. Motion Information Detection of Rock Particles Based on VIS

3.1. Method Framework

Because of the complexity of detecting the motion information of rock particles, it is necessary to segment the video instance first. For resiliency in environments in which the objects and backgrounds are highly similar, we designed a new classification loss function for Mask R-CNN based on Arcface loss to improve the feature recognition ability and ameliorate the instance segmentation effect of particles. The combination of the improved Mask R-CNN and Deep SORT can further accomplish the VIS of rock particles. The use of the equivalent ellipse characterization method can extract the center position of each particle frame by frame, and the pixel translation can be obtained by subtracting the center position of the previous frame. Then, the proportional calibration algorithm can calibrate the pixel translation and actual translation. On the other hand, the rotation can be detected by calculating the change in the angle of the ellipse’s major axis. The overall algorithm framework is illustrated in Figure 1.

3.2. Particle Segmentation

3.2.1. Mask R-CNN

Mask R-CNN [26] adds a mask prediction branch based on Faster R-CNN [27] and achieves pixel-level segmentation of images by combining the advantages of object detection and semantic segmentation networks. It has been used in fields such as agriculture, engineering, and daily life [28,29,30]. The region of interest alignment (RoI Align) is exploited to replace the region of interest pooling (RoI pooling) in Mask R-CNN, which solves the region mismatch problem. The overall network structure of Mask R-CNN is shown in Figure 2. It is composed mainly of four parts: backbone, region proposal network (RPN), RoI Align, and classifier.
The backbone is a series of convolutional layers that can extract the feature maps. In this study, we chose ResNet-101 as the backbone. ResNet-101 had five convolutional layers. The samples were convolved, regularized, and activated in each layer to obtain feature maps of different sizes. Then, the use of feature pyramid networks (FPNs) [31] could fuse multilayered semantic features and acquire feature maps after fusion of various sizes.
The RPN is a network that can extract region proposals through classification and regression. The inputs of the RPN are the outputs of the last layer of the FPN. First, a certain number of anchors are generated for each pixel of the feature maps. The likelihood that these anchors are the foreground or background and the offset between these anchors and their corresponding ground truth is calculated. Then, the loss of classification and regression could be assessed according to the loss function of the RPN.
RoI Align utilizes bilinear interpolation to replace the quantization rounding operation of RoI pooling. It can convert the region proposals into fixed-size feature maps without losing spatial information so that each pixel can maintain accurate coordinates.
The classifier had three parallel branches. Two of these are categories and coordinates that can obtain a precisely positioned bounding box. The other branch utilizes a fully convolutional network (FCN) [32] to predict the mask and map the feature size to the image size.

3.2.2. Classification Loss Function Based on Arcface Loss

The rock particles were densely distributed and are in close contact with each other in the experimental environment. Considering the task of extracting the translation and rotation of particles in this study, we attributed the coarse particles that were heavily obscured as the background for reducing the interference of such coarse particles on the accuracy of motion information. The characteristics of such particles in the background were similar to those of the objects. In addition, many small particles in the background were similar to the objects. The diameters of the coarse and small particles were 20–30 mm and 2.5–7.5 mm, respectively, and the diameter ratio of coarse particles to small particles was about 2.7–12. These two types of particles (coarse particles that are heavily occluded and small particles) in the background were highly similar to the object particles.
SoftMax [33] loss is usually used to calculate the possibility and ensure the separability of objects as a common classification loss function in an object detection model. As many particles are heavily obscured and small in the experimental environment, it is difficult to distinguish features when using the traditional SoftMax loss. The A-SoftMax loss [34], which is based on the traditional SoftMax loss, was used to improve the feature discrimination ability of the model. It adds the angular distance based on SoftMax loss, which can distinguish the objects and backgrounds with higher similarity. The A-SoftMax loss can be expressed as follows:
L c l s = 1 N i = 1 N log ( e x i φ ( θ y i , i ) e x i φ ( θ y i , i ) + j y i e x i φ ( θ j , i ) )
φ ( θ y i , i ) = ( 1 ) k cos ( m θ y i , i ) 2 k
where N is the number of training samples, xi is the quantitative data of object features, θ y i is the angle between the object feature vector and Xi weight W y i , θ y i [ k π m , ( k + 1 ) π m ] , k [ 0 , m 1 ] , and m is the angular magnification.
The above loss function can narrow the within-class distance of the network model and enlarge the between-class distance, which can improve the feature classification ability of the model. However, the angle amplification mechanism also makes it difficult for the model to converge. Therefore, we used a classification loss function based on Arcface loss [35] by integrating the features of SoftMax loss, which can be expressed as follows:
L c l s = 1 N i = 1 N log ( e s ( cos ( θ y i + n ) ) e s ( cos ( θ y i + n ) ) + j y i e s cos ( θ j ) ) ,
where s is the fixed scale factor and n is the angle margin.
The theory of Arcface loss is shown in Figure 3. Xi is feature and W is the ground truth weight. Based on Xi and W normalization, we could get cos θ j . The inverse cosine function can calculate the angle between Xi and W y i . The cosine distance can be obtained by adding n the object angle. Then, it outputs to the SoftMax layer after multiplying with s. Compared with A-SoftMax loss, Arcface loss utilizes angular distance and abandons the angle magnification mechanism to make the network converge more easily.
The other two parts of loss function are the regression loss function L r e g and mask loss function L m a s k . L r e g obtains the bounding-box regression loss by the smooth function and L m a s k is the binary cross entropy loss function. Therefore, the overall loss function can be expressed as follows:
L = L c l s + L r e g + L m a s k
where L r e g and L m a s k are shown in the equations below:
L r e g = s m o o t h ( t i + t i ) = { 0.5 x , i f | t i t i | < 1 | t i t i | 0.5 , o t h e r w i s e ,
L m a s k = i p i log ( p i ) + ( 1 p i ) ( 1 p i ) ,
where t i is the vector with four-coordinate data of the bounding box and t i is the vector with the data of the ground truth. p i is the binary classification probability and p i is the i-th output.

3.3. Particle Tracking

As an improvement of SORT [36], Deep SORT [37] adds appearance information to matching problems, which can reduce the switching of object IDs. It is divided mainly into three aspects: state estimation, trajectory processing, and matching problems. The main tracking framework is shown in Figure 4.
The state of the trajectory at a certain moment is represented by an eight-dimensional space of ( u , v , r , h , u · , v · , r · , h · ) , where ( u , v ) is the center position of the bounding box, r is the ratio of length to width, h is the height, and the other four variables represent the speed information in the image coordinates. A standard Kalman filter [38] based on a constant-velocity model and a linear observation model can predict the motion state. The predicted result was ( u , v , r , h ) .
Trajectory processing consists of two parts: new detection result processing and tracker processing. If the detection result cannot be associated with the existing trajectory, it is regarded as a new detection result, and a new trajectory hypothesis is applied. If the new trajectory can be successfully associated in the first three frames, it is determined as a new object tracking. Otherwise, the trajectory was deleted. On the other hand, we set up a tracker that increases progressively during Kalman filter prediction and sets it to 0. If the tracker exceeds the predefined maximum threshold A max , we consider that object tracking is over.
The Deep SORT considers the association of motion information and appearance information simultaneously and combines them to calculate the degree of matching between the detection results and trajectories.
Deep SORT takes advantage of the Mahalanobis distance to describe the degree of motion information association. The Mahalanobis distance reflects the uncertainty of the state measurement by calculating the standard deviation between the detected position and the average tracking position. This can be expressed as follows:
d ( 1 ) ( i , j ) = ( d j y i ) T S i 1 ( d j y i ) ,
where d j represents the position of the detected bounding box j-th, y i represents the predicted position of the object by the tracker, and S i is the covariance matrix between the detection position and average tracking position. In addition, Deep SORT excludes impossible associations by Mahalanobis distance thresholding within the 95% confidence interval calculated from the inverse distribution χ 2 . This can be expressed as follows:
b i , j ( 1 ) = 1 [ d ( 1 ) ( i , j ) t ( 1 ) ]
where t ( 1 ) = 9.4877 . If the Mahalanobis distance is less than the specified threshold t ( 1 ) , the motion state will be successfully associated.
In terms of the appearance information association of objects, Deep SORT obtains the 128-dimensional feature vector of r j the bounding box through d j the CNN network. The restriction condition was r i = 1 . For each trajectory k , a library of R k appearance descriptors is applied to store the feature vector of the last 100 frames. Finally, the minimum cosine distance between the trajectory i-th and trajectory j-th can be calculated [39]. This process can be expressed as follows:
d ( 2 ) ( i , j ) = min { 1 r j T r k ( i ) | r k ( i ) R }
Similarly, Deep SORT sets a threshold t ( 2 ) for this information association, which can be expressed as follows:
b i , j ( 2 ) = 1 [ d ( 2 ) ( i , j ) t ( 2 ) ]
The linear weights of the two association methods are utilized to reflect the final degree of association, which can be expressed as follows:
c i , j = λ d ( 1 ) ( i , j ) + ( 1 λ ) d ( 2 ) ( i , j )
where λ is the associated weight of the motion information.

3.4. Motion Information Detection

Translation and rotation are the most representative motion information parameters in engineering and scientific research. It is necessary to characterize the particles to detect the translation and rotation of rock particles under a vibration load. Common characterization methods include the smallest circumscribed circle, the smallest circumscribed rectangle, and the equivalent ellipse [40]. We chose the equivalent ellipse as the characterization method for the detection of the translation and rotation of particles on account of the appearance characteristics of the rock particles.
We detected and segmented the masks of particles using the improved Mask R-CNN and tracked these masks with Deep SORT. Then, these masks were processed by the equivalent ellipse characterization method, which contributed to converting them into a numerical value with practical meaning, as shown in Figure 5.
The centers of the ellipses could replace the centers of mass of the particles. We extracted the center position of the ellipse frame by frame and obtained the translation by subtracting the center position of the previous frame. The units of the results measured by the above method are pixels. To acquire the actual translation of the rock particles, calculating the physical size of a single pixel is necessary, which is the calibration of image pixels [41]. We calibrated the images using a proportional calibration algorithm, which can be expressed as follows:
k = L N
where k is the actual size of pixels, L is the actual size, and N is the number of pixels. The rotation can be acquired by calculating the change in the major axis angles of the ellipses frame by frame.

4. Experiment and Analysis

4.1. Experimental Equipment and Parameter Settings

We developed experimental equipment to collect videos, including the experiment box, vibration motor, load plate, reinforcement ring, CCD camera, transparent panel, coarse particles, and small particles. In this experiment, the vibration motor acted on the load plate and provided a vibration load for the particles; the reinforcement ring was used to fix the experiment box; the transparent panel could block particles; the CCD camera captured videos of the experimental environment. The SHL-500W CCD camera had a resolution of 1600 × 1200 pixels and a frame rate of 30 fps. The lens model used was LT-C0516-5MP with a focal length of 5 mm. The horizontal distance between the experiment box and the CCD camera was 143 cm, and the bottom width of the experiment box was 840 mm. The particles selected for this experiment were pebbles with irregular shapes and smooth surfaces. An image of the experimental environment captured by the visual sensor is shown in Figure 6.
The training and testing experiments of the model were conducted on Ubuntu 18.04. The processor was an Intel Core i7-8700K CPU @3.7 GHz (Intel, Mountain View, CA, USA), and the GPU was an NVIDIA GeForce RTX2080Ti (NVIDIA, Santa Clara, CA, USA). The framework of the deep learning framework was TensorFlow. The official weight was utilized as the pretraining weight, and the stochastic gradient descent was used during network training to accelerate the convergence speed of the Mask R-CNN. We set 80 epochs and performed 100 iterations per epoch; therefore, a total of 8 × 103 iterations were performed. The activation function was ReLU, and the batch size was 2. The other key training parameters are listed in Table 1.
In the process of object tracking by Deep SORT, the particle movement speed of particles was relatively slow, and the motion uncertainty was low. Therefore, the association of motion information was regarded as the main matching index. λ and A max took the values 0.8 and 30, respectively.

4.2. Dataset

A binary mixture of coarse and small particles often assumes multiple spatial distribution states. The degree of mixing is an index used to evaluate the distribution uniformity of the number of binary particles in the mixing process. Zero indicates that the coarse and small particles are completely separated, and 100 indicates that the coarse and small particles are evenly distributed. The collected images, with mixing degrees of 0, 20, 40, 60, 80, and 100, were divided into three categories according to the proportion of particles (i.e., coarse particles account for 25%, 50%, and 75%). There were 18 categories in total, each of which contained 10 images in different compaction states. We then obtained a total of 180 images of particles, and each image of particles was cropped to make the effective data in the image more obvious. Because of insufficient data, we used image enhancement technology (e.g., translation, rotation, and affine transformation) so that each image could increase the other four additional images. Finally, 900 images were utilized for subsequent training and analysis. Specifically, we used the LabelMe image annotation tool to generate a file in .json format and then divided the dataset into a training set, validation set, and test set at a ratio of 8:1:1.

4.3. Algorithm Evaluation Index

To evaluate the effects of Mask R-CNN comprehensively and objectively, we utilized the accuracy, precision, recall, and F1 score as algorithm evaluation indicators.
Accuracy is the percentage of correct prediction results in the total samples, and it is calculated using Equation (13):
A c c u r a c y = T P + T N T P + F P + T N + F N
where T P is the number of pixels that actually belong to particles and are predicted to be particles by the model, F P is the number of pixels that actually belong to the background but are predicted to be particles by the model, T N is the number of pixels that actually belong to the background and are predicted to be background by the model, and F N is the number of pixels that actually belong to particles but are predicted to be background by the model. Therefore, T P + T N is the number of pixels correctly predicted by the model, and F P + F N is the number of pixels incorrectly predicted by the model, where T P + F P + T N + F N is the total number of pixels.
Precision is the ratio of the number of samples that are correctly identified as a certain class to the actual number of this class. Recall is the ratio of the number of samples that are correctly identified as a certain class to the predicted total number of this class. These are shown in the following equations:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
where TP is the number of samples that are actually positive and divided into positive classes by the classifier, FP is the number of samples that are actually negative and divided into positive classes by the classifier, and FN is the number of samples that are actually positive and divided into negative classes by the classifier.
The F1 score is the harmonic mean of accuracy and recall, and the calculation formula is as follows:
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l .
If the objects switched frequently in the process of tracking the movement of rock particles using Deep SORT, the tracking effect would be worse. Therefore, we operated the ID switching rate ( I D s w i t c h ) of the objects as the evaluation index of the tracking effect of the rock particles. I D s w i t c h can be expressed as follows:
I D s w i t c h = I s w i t c h I a l l .
where I s w i t c h is the number of IDs switching during the tracking, and I a l l is the total number of detected objects in the video stream.
We utilized the detection errors δ T and δ R to indicate the accuracy of translation and rotation in a video. These calculation methods are shown in Equations (18) and (19):
δ T = 1 P F p = 1 P f = 1 F | T p f t p f | t p f ,
δ R = 1 P F p = 1 P f = 1 F | R p f r p f | r p f ,
where P is the number of particles selected to calculate the detection errors, which is set to 10 in this experiment. F is the number of images selected to calculate the detection errors in the video. Moreover, we selected an image every 5 s (150 frames) in this experiment. T p f and R p f respectively represent the translation and rotation detected by the visual method of the p-th particle in the f-th image, with t p f and r p f respectively being the true translation and rotation of the p-th particle in the f-th image.

4.4. Instance Segmentation Experiment

To verify the effectiveness of the improved Mask R-CNN for the detection and segmentation of rock particles, we compared it with the standard Mask R-CNN and other state-of-the-art methods. Firstly, we calculated detection and segmentation performance evaluation indicators on the self-created dataset and recorded them in Table 2. Secondly, we reported instance segmentation results on COCO using the standard metrics in Table 3. All models were trained on the train2017 and tested on the val2017. Final results were on test-dev. Finally, we also compared the improved results of the classification loss with some enhancement results of Mask R-CNN on COCO minival (Table 4).
The original images of particles are shown in Figure 7a,d. Figure 7b,e show the detection and segmentation results of the rock particles by the Mask R-CNN. Some rock particles were not detected. For particles largely obscured and surrounded by small particles, the missing detection is more serious. This means that it is difficult to classify coarse particles with a high degree of similarity in the backgrounds, which results in a poor detection effect. The segmented mask can cover the object, and the edge contour of the coarse particles does not have obvious undersegmentation or oversegmentation. Figure 7c,f show the results of the improved Mask R-CNN. After adding the classification loss function based on Arcface loss, the Mask R-CNN has a certain improvement in the detection effect of rock particles, which means that it can detect and segment some rock particles that are partially covered or surrounded by small particles. The segmentation effect is similar to the previous standard Mask R-CNN algorithm. More particles cannot be detected as the degree of mixing and the proportion of coarse particles increase.
Table 2 shows that all indicators have been increased through the improved Mask R-CNN to detect and segment the rock particles. Recall has the largest increase of 5.99%, which indicates that the algorithm has improved in distinguishing the particles and backgrounds in pictures, and more objects can be detected. The comprehensive evaluation indexes accuracy and F1 score increased by 4.20% and 4.28%, respectively. This shows that the overall detection and segmentation effect of the improved algorithm is better than that of the standard algorithm. In addition, the improved Mask R-CNN also shows better results than other state-of-the-art methods. The improved Mask R-CNN is more suitable for subsequent VIS, ellipse characterization, and detection of motion information of particles.
As shown in Table 3, the improved Mask R-CNN offers competitive instance segmentation performance. The instance segmentation results of the improved Mask R-CNN on COCO are slightly better than the standard Mask R-CNN. Table 3 also reports a certain mask AP improvement compared to FCIS, RetinaMask, and YOLACT. Our improvement to Mask R-CNN also has some advantages on public datasets.
Table 4 shows that all improvements increase mask AP by 2.9 points (from 36.7 to 39.6) and box AP by 4.4 points (from 39.6 to 44.0). The addition of the classification loss function increases both mask AP and box AP. This shows that the new loss function can be used as a component to improve the detection and segmentation effect of Mask R-CNN.

4.5. VIS Experiment

We also processed four videos to verify the efficacy of the improved Mask R-CNN and Deep SORT. The fps of the videos was 30, and each video contained a complete compaction process. As shown in Figure 8, we selected five frames with obvious changes from each video as qualitative experimental results to reflect the VIS effect during vibrational compaction. The tracking evaluation indicators of the particles were measured and are listed in Table 5.
Figure 8 shows that the combination of the improved Mask R-CNN and Deep SORT can realize the VIS of rock particles. The improved Mask R-CNN can detect and segment objects frame by frame, and then we input the bounding box into Deep SORT to achieve object tracking. This experiment could process the video stream and integrate the detection, segmentation, and tracking of rock particles to achieve VIS, which provided the basis of machine vision for motion information detection of particles.
Table 5 lists the information statistics of the video stream. IDswitch is low, which shows that the switching probability of the particle ID is low, and its tracking effect is significant. The IDswitch of the improved Mask R-CNN and Deep SORT are lower than that of standard Mask R-CNN, which indicates that better detection and segmentation benefit the tracking quality. We also found the IDswitch was improved even more for the videos with higher degree of mixing and proportion of coarse particles. Videos with more coarse particles are easier to switch IDs. This means that complex and crowded scenes may cause difficulty in information correlation in tracking, which negatively influences the tracking effect of the Deep SORT.

4.6. Particle Motion Information Detection Experiment

Based on the improved Mask R-CNN and Deep SORT, we measured the translation and rotation of coarse particles. The data collection frequency was high (30 times per second) because the videos were analyzed frame by frame. We used motor vibrations to produce translation and rotation so that translation and rotation curves had several jitters with small amplitudes and high frequencies. Therefore, the Savitzky–Golay filter, which can filter high-frequency jitter without changing the shape of the curve, was used to process the curve. The filtered translation and rotation curves of some particles are shown in Figure 9.
We developed a method to extract the translation and rotation of particles by manually marking the long axis to compare the effects of the machine vision method. We selected a frame every 5 s from the videos and used LabelMe to manually label the long axes of selected particles. Then, the long axes coordinates were extracted from the .json file generated by the annotation. The translation and rotation of particles were calculated using the long axes coordinates. As shown in Figure 9, the translation and rotation extracted by this method were taken as real values and were used to construct a scatter diagram. To further verify the accuracy of the vision method proposed in this study, we used Equations (18) and (19) to calculated the detection errors δT and δR (Table 6).
As shown in Figure 9, the translation and rotation of rock particles increase with time under the vibration load. The long axes of the equivalent ellipses are difficult to determine for particles with shapes close to the standard circle. Therefore, the detection errors for rotation are greater than those for translation (Table 6). We found that the improvement of the Mask R-CNN algorithm does not significantly reduce the detection error (only 0.12% and 0.69%). However, the improvement of Mask R-CNN reduced the ID switching rate, which is important for the tracking effects of particles. The analysis is presented below.
Figure 10 shows examples of the standard Mask R-CNN and Deep SORT failures. Figure 10a shows the translation curve of the standard and improved algorithms and Figure 10b shows the rotation curve. As the particle tracked by the standard Mask R-CNN and Deep SORT switched ID is in the range of 30–35 s, the translation and rotation of the particle changed dramatically during this time. In contrast, the tracking curves of the improved Mask R-CNN and Deep SORT remained stable. The improved Mask R-CNN could reduce the ID switching rate, although it did not significantly improve the segmentation effect. Therefore, the improvement in Mask R-CNN was beneficial to the tracking effect and accuracy of detection because any ID switch would have a significant impact on the tracking result.

4.7. Discussion

The motion information of rock particles under a vibration load is related to the vibration amplitude and frequency as well as the shape, position, and material characteristics of the particles themselves. This experiment can detect the translation and rotation of many particles based on VIS.
The thickness of the transparent panel was 10 mm. The particles in the middle are closer to the lens because the transparent panel near the center is under more pressure, which might cause the translation of middle particles to be greater than the actual translation. The deformation of the transparent panel was 3–6 mm by measurement, and the overall movement range of the particles in this experiment was small, so we think this error could be ignored. In addition, the deformation of the transparent panel may change the light path and lens distortion, but the errors of translation and rotation will not be particularly large because they are calculated by relative position.
The detection of rock particle motion information proposed in this study has the following three limitations. First, the machine vision method cannot detect the internal and occluded particles. Second, this method approximately converts a three-dimensional space to a two-dimensional space, which adversely affects the test results. Finally, this method causes particle ID switching during the tracking process and this can significantly affect the tracking process, even though the probability of switching is low.

5. Conclusions and Outlook

We proposed a method for detecting the motion information of rock particles under a vibration load based on VIS. First, we improved the Mask R-CNN and designed a classification loss function based on Arcface loss to improve feature discrimination. The method can adapt to the phenomenon of high similarity between objects and backgrounds in the experimental environment. The combination of the improved Mask R-CNN and Deep SORT can achieve multiparticle VIS. Finally, we used the equivalent ellipse characterization method for segmented particles. Our method can combine with the proportional calibration algorithm to obtain the translation and rotation of particles by calculating the change in angle of the long axis of the characterizing ellipse. Experimental results show that the improved Mask R-CNN is better than the standard algorithm. The improved Mask R-CNN can perform VIS by combining it with a Deep SORT. The ID switching rate is low, and the motion information of the rock particles can be detected successfully. This is a foundation for researching the movement law and quantitative analysis of particles under a vibration load. In the follow-up, we will continue to research the motion information detection of small particles and analyze the detection results.

Author Contributions

Conceptualization, M.C. and W.Y.; methodology, W.Y.; software, M.C. and Y.L.; validation, M.C. and M.L.; formal analysis, Y.L.; investigation, M.C. and W.Y.; resources, M.C.; data curation, M.C.; writing—original draft preparation, M.C.; writing—review and editing, M.C. and M.L.; visualization, M.C. and Y.L.; supervision, M.L. and W.Y.; project administration, W.Y.; funding acquisition, W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 61601061.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jerónimo, P.; Resende, R.; Fortunato, E. An assessment of contact and laser-based scanning of rock particles for railway ballast. Transp. Geotech. 2020, 22, 100302. [Google Scholar] [CrossRef]
  2. Zhou, Y.; Shi, Z.; Zhang, Q.; Jang, B.; Wu, C. Damming process and characteristics of landslide-debris avalanches. Soil Dyn. Earthq. Eng. 2019, 121, 252–261. [Google Scholar] [CrossRef]
  3. Gao, G.; Meguid, M.A.; Chouinard, L.E.; Xu, C. Insights into the Transport and Fragmentation Characteristics of Earthquake-Induced Rock Avalanche: Numerical Study. Int. J. Géoméch. 2020, 20, 04020157. [Google Scholar] [CrossRef]
  4. Kh, A.B.; Mirghasemi, A.A.; Mohammadi, S. Numerical simulation of particle breakage of angular particles using combined DEM and FEM. Powder Technol. 2011, 205, 15–29. [Google Scholar]
  5. Liu, G.Y.; Xu, W.J.; Sun, Q.C.; Govender, N. Study on the particle breakage of ballast based on a GPU accelerated discrete element method. Geosci. Front. 2020, 11, 461–471. [Google Scholar] [CrossRef]
  6. Bagherzadeh, H.; Mansourpour, Z.; Dabir, B. Numerical analysis of asphaltene particles evolution and flocs morphology using DEM-CFD approach. J. Pet. Sci. Eng. 2021, 199, 108309. [Google Scholar] [CrossRef]
  7. Li, B.; Liang, Y.; Zhang, L.; Zou, Q. Breakage law and fractal characteristics of broken coal and rock masses with different mixing ratios during compaction. Energy Sci. Eng. 2019, 7, 1000–1015. [Google Scholar] [CrossRef] [Green Version]
  8. Yang, L.; Fan, Y.; Xu, N. Video instance segmentation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 5188–5197. [Google Scholar]
  9. Yen, Y.K.; Lin, C.L.; Miller, J.D. Particle overlap and segregation problems in on-line coarse particle size measurement. Powder Technol. 1998, 98, 1–12. [Google Scholar] [CrossRef]
  10. Zhang, G.Y.; Liu, G.Z.; Zhu, H. Segmentation algorithm of complex ore images based on templates transformation and reconstruction. Int. J. Miner. Metall. Mater. 2011, 18, 385. [Google Scholar] [CrossRef]
  11. Amankwah, A.; Aldrich, C. Automatic ore image segmentation using mean shift and watershed transform. In Proceedings of the 21st International Conference Radioelektronika (RADIOELEKTRONIKA), Brno, Czech Republic, 19–20 April 2011; pp. 1–4. [Google Scholar]
  12. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  13. Yuan, L.; Duan, Y. A Method of Ore Image Segmentation Based on Deep Learning. In Proceedings of the International Conference on Intelligent Computing (ICIC), Wuhan, China, 15–18 August 2018; pp. 508–519. [Google Scholar]
  14. Duan, J.; Liu, X.; Wu, X.; Mao, C. Detection and segmentation of iron ore green pellets in images using lightweight U-net deep learning network. Neural Comput. Appl. 2019, 32, 5775–5790. [Google Scholar] [CrossRef]
  15. Liu, X.; Zhang, Y.; Jing, H.; Wang, L.; Zhao, S. Ore image segmentation method using U-Net and Res_Unet convolutional networks. RSC Adv. 2020, 10, 9396–9406. [Google Scholar] [CrossRef] [Green Version]
  16. Jaboyedoff, M.; Oppikofer, T.; Abellán, A.; Derron, M.H.; Loye, A.; Metzger, R.; Pedrazzini, A. Use of LIDAR in landslide investigations: A review. Nat. Hazards 2012, 61, 5–28. [Google Scholar] [CrossRef] [Green Version]
  17. Scaioni, M. Modern Technologies for Landslide Monitoring and Prediction; Springer: Berlin/Heidelberg, Germany, 2015; pp. 69–88. [Google Scholar]
  18. Adrian, R.J. Particle-imaging techniques for experimental fluid-mechanics. Annu. Rev. Fluid Mech. 1991, 23, 261–304. [Google Scholar] [CrossRef]
  19. Sutton, M.A.; Mcneill, S.R.; Helm, J.D.; Chao, Y.J. Advances in two-dimensional and three-dimensional computer vision. Top. Appl. Phys. 2000, 77, 323–372. [Google Scholar]
  20. Bing, P.; Wu, D.; Xia, Y. Incremental calculation for large deformation measurement using reliability-guided digital image correlation. Opt. Lasers Eng. 2012, 50, 586–592. [Google Scholar]
  21. Stanier, S.A.; Blaber, J.; Take, W.A.; White, D. Improved image-based deformation measurement for geotechnical applications. Can. Geotech. J. 2016, 53, 727–739. [Google Scholar] [CrossRef]
  22. Travelletti, J.; Malet, J.P. Characterization of the 3d geometry of flow-like landslides: A methodology based on the integration of heterogeneous multi-source data. Eng. Geol. 2012, 128, 30–48. [Google Scholar] [CrossRef]
  23. Gance, J.; Malet, J.P.; Dewez, T.; Travelletti, J. Target detection and tracking of moving objects for characterizing landslide displacements from time-lapse terrestrial optical images. Eng. Geol. 2014, 172, 26–40. [Google Scholar] [CrossRef]
  24. Shum, H.; Komura, T. Tracking the translational and rotational movement of the ball using high-speed camera movies. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Genoa, Italy, 11–14 September 2005; pp. 3–1084. [Google Scholar]
  25. Guccione, D.E.; Thoeni, K.; Giacomini, A.; Buzzi, O.; Fityus, S. Efficient Multi-View 3d Tracking of Arbitrary Rock Fragments upon Impact. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, 43, 589–596. [Google Scholar] [CrossRef]
  26. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  27. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Li, D.; Chen, Y.; Zhang, K.; Li, Z. Mounting behaviour recognition for pigs based on deep learning. Sensors 2019, 19, 4924. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Fang, X.; Jie, W.; Feng, T. An industrial micro-defect diagnosis system via intelligent segmentation region. Sensors 2019, 19, 2636. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Li, X.; Fan, Z.; Liu, Y.; Li, Y.; Dai, Q. 3d pose detection of closely interactive humans using multi-view cameras. Sensors 2019, 19, 2831. [Google Scholar] [CrossRef] [Green Version]
  31. Wang, Y.; Li, H.; Jia, P.; Zhang, G.; Wang, T.; Hao, X. Multi-Scale DenseNets-Based Aircraft Detection from Remote Sensing Images. Sensors 2019, 19, 5270. [Google Scholar] [CrossRef] [Green Version]
  32. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Santiago, Chile, 13–16 December 2015; pp. 3431–3440. [Google Scholar]
  33. Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-margin softmax loss for convolutional neural networks. In Proceedings of the 33rd International Conference on Machine Learning (ICML), New York, NY, USA, 19–24 June 2016; Volume 2, p. 7. [Google Scholar]
  34. Liu, W.; Wen, Y.; Yu, Z.; Li, M.; Raj, B.; Song, L. Sphereface: Deep hypersphere embedding for face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 22–25 July 2017; pp. 212–220. [Google Scholar]
  35. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 4690–4699. [Google Scholar]
  36. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  37. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijin, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
  38. Chen, Z.; Han, Z.; Hao, J.; Zhu, Q.; Soh, Y.C. Fusion of wifi, smartphone sensors and landmarks using the kalman filter for indoor localization. Sensors 2015, 15, 715–732. [Google Scholar] [CrossRef]
  39. Qiu, Z.; Zhao, N.; Zhou, L.; Wang, M.; Yang, L.; Fang, H.; Liu, Y. Vision-based moving obstacle detection and tracking in paddy field using improved yolov3 and deep SORT. Sensors 2020, 20, 4082. [Google Scholar] [CrossRef]
  40. Hamzeloo, E.; Massinaei, M.; Mehrshad, N. Estimation of particle size distribution on an industrial conveyor belt using image analysis and neural networks. Powder Technol. 2014, 261, 185–190. [Google Scholar] [CrossRef]
  41. Yang, Z.; Yuan, Y.; Zhang, M.; Zhao, X.; Tian, B. Safety distance identification for crane drivers based on mask r-cnn. Sensors 2019, 19, 2789. [Google Scholar] [CrossRef] [Green Version]
  42. Li, Y.; Qi, H.; Dai, J.; Ji, X.; Wei, Y. Fully convolutional instance-aware semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2359–2367. [Google Scholar]
  43. Fu, C.Y.; Shvets, M.; Berg, A.C. RetinaMask: Learning to predict masks improves state-of-the-art single-shot detection for free. arXiv 2019, arXiv:1901.03353. [Google Scholar]
  44. Bolya, D.; Zhou, C.; Xiao, F.; Lee, Y.J. Yolact: Real-time instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 9157–9166. [Google Scholar]
Figure 1. Method framework.
Figure 1. Method framework.
Sensors 21 04108 g001
Figure 2. Mask R-CNN network structure.
Figure 2. Mask R-CNN network structure.
Sensors 21 04108 g002
Figure 3. Arcface loss theory. * indicates multiplication.
Figure 3. Arcface loss theory. * indicates multiplication.
Sensors 21 04108 g003
Figure 4. Deep SORT tracking framework.
Figure 4. Deep SORT tracking framework.
Sensors 21 04108 g004
Figure 5. Equivalent ellipse characterization method.
Figure 5. Equivalent ellipse characterization method.
Sensors 21 04108 g005
Figure 6. Experimental equipment.
Figure 6. Experimental equipment.
Sensors 21 04108 g006
Figure 7. Experiment effect images of instance segmentation: (a) shows an image with a mixing degree of 40, and the proportion of coarse particles is 25; (b) shows the segmentation effect of (a) with the standard Mask R-CNN; (c) shows the segmentation effect of (a) with the improved Mask R-CNN; (d) shows an image with a mixing degree of 60, and the proportion of coarse particles is 50; (e) shows the segmentation effect of (d) with the standard Mask R-CNN; (f) shows the segmentation effect of (d) with the improved Mask R-CNN.
Figure 7. Experiment effect images of instance segmentation: (a) shows an image with a mixing degree of 40, and the proportion of coarse particles is 25; (b) shows the segmentation effect of (a) with the standard Mask R-CNN; (c) shows the segmentation effect of (a) with the improved Mask R-CNN; (d) shows an image with a mixing degree of 60, and the proportion of coarse particles is 50; (e) shows the segmentation effect of (d) with the standard Mask R-CNN; (f) shows the segmentation effect of (d) with the improved Mask R-CNN.
Sensors 21 04108 g007
Figure 8. Experiment effect of VIS: (a) shows the tracking effect of video1 with a mixing degree of 80, and the proportion of coarse particles is 25; (b) shows the tracking effect of video2 with a mixing degree of 40, and the proportion of coarse particles is 25; (c) shows the tracking effect of video3 with a mixing degree of 60, and the proportion of coarse particles is 50; (d) shows the tracking effect of video4 with a mixing degree of 60, and the proportion of coarse particles is 75.
Figure 8. Experiment effect of VIS: (a) shows the tracking effect of video1 with a mixing degree of 80, and the proportion of coarse particles is 25; (b) shows the tracking effect of video2 with a mixing degree of 40, and the proportion of coarse particles is 25; (c) shows the tracking effect of video3 with a mixing degree of 60, and the proportion of coarse particles is 50; (d) shows the tracking effect of video4 with a mixing degree of 60, and the proportion of coarse particles is 75.
Sensors 21 04108 g008
Figure 9. Motion information of some rock particles.
Figure 9. Motion information of some rock particles.
Sensors 21 04108 g009
Figure 10. The example of the failure of standard Mask R-CNN and Deep SORT. (a) is the translation comparison chart. (b) is the rotation comparison chart.
Figure 10. The example of the failure of standard Mask R-CNN and Deep SORT. (a) is the translation comparison chart. (b) is the rotation comparison chart.
Sensors 21 04108 g010
Table 1. Key training parameters.
Table 1. Key training parameters.
Parameter TypeParameterValue
Global ParameterNumber of categories2
Initial weight1
Learning rate10−3
Attenuation coefficient of learning rate10−4
Attenuation step length100
Momentum Factor0.9
RPN Mesh ParameterAnchor size16, 32, 64, 128, 256
Anchor shape ratio0.5, 1, 2
Anchor step length1
Maximum number of Anchors per image256
Non-Maximum Suppression (NMS) threshold for anchor0.7
The number of anchors after NMS2000
Header Network ParameterConfidence threshold0.7
Maximum number of instances100
NMS threshold0.3
Table 2. Experimental results of instance segmentation on the self-created dataset.
Table 2. Experimental results of instance segmentation on the self-created dataset.
MethodAccuracy%Precision%Recall%F1%
FCIS [42]71.9776.2659.6866.96
RetinaMask [43]86.3485.7169.5376.78
YOLACT [44]73.2874.4560.1166.52
Standard Mask R-CNN89.1689.7272.1779.99
Improved Mask R-CNN93.3691.4178.1684.27
Table 3. The instance segmentation results on COCO.
Table 3. The instance segmentation results on COCO.
Method A P m A P 50 m A P 75 m
FCIS [42]29.551.530.2
RetinaMask [43]34.755.436.9
YOLACT [44]29.848.531.2
Standard Mask R-CNN35.758.537.8
Improved Mask R-CNN36.058.938.3
Table 4. Comparison of the improved results of the classification loss with some enhancement results of Mask R-CNN on COCO minival. All the results use ResNet-101 as the backbone. Each row adds an extra component to the above row. The Mask R-CNN results are from Table 8 in the Appendix of Mask R-CNN [26] (Copyright Year: 2017. Copyright Owner’s Name: Kaiming He).
Table 4. Comparison of the improved results of the classification loss with some enhancement results of Mask R-CNN on COCO minival. All the results use ResNet-101 as the backbone. Each row adds an extra component to the above row. The Mask R-CNN results are from Table 8 in the Appendix of Mask R-CNN [26] (Copyright Year: 2017. Copyright Owner’s Name: Kaiming He).
Method A P m A P 50 m A P 75 m A P b b A P 50 b b A P 75 b b
Standard Mask R-CNN36.759.538.939.661.543.2
+ update baseline37.059.739.040.563.043.7
+ c2c training37.660.439.941.764.145.2
+ ImageNet–5k38.661.740.942.765.146.6
+ train-time augmentation39.262.541.643.565.947.2
+ loss function (ours)39.663.142.344.066.547.7
Table 5. The tracking evaluation indicators of the particles. “SM + D” is the experimental results of the standard Mask R-CNN and Deep SORT. “IM + D” represents the experimental results of the improved Mask R-CNN and Deep SORT.
Table 5. The tracking evaluation indicators of the particles. “SM + D” is the experimental results of the standard Mask R-CNN and Deep SORT. “IM + D” represents the experimental results of the improved Mask R-CNN and Deep SORT.
VideoMethodIswitchIallIDswitch%
Video1SM + D73256210.28
Video1IM + D45256210.18
Video2SM + D1059595491.78
Video2IM + D638595491.07
Video3SM + D35541912571.86
Video3IM + D28111912571.47
Video4SM + D39771425342.79
Video4IM + D27221425341.91
Table 6. The detection errors. “SM + D” and “IM + D” are the same as in Table 2. No ID switching occurs for the 10 selected particles in each video.
Table 6. The detection errors. “SM + D” and “IM + D” are the same as in Table 2. No ID switching occurs for the 10 selected particles in each video.
VideoMethodδt%δr%
Video1SM + D3.5311.84
Video1IM + D3.4712.31
Video2SM + D5.5115.43
Video2IM + D5.1413.81
Video3SM + D6.2715.28
Video3IM + D6.4116.06
Video4SM + D5.5818.16
Video4IM + D5.3615.79
AverageSM + D5.2215.18
AverageIM + D5.1014.49
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, M.; Li, M.; Li, Y.; Yi, W. Rock Particle Motion Information Detection Based on Video Instance Segmentation. Sensors 2021, 21, 4108. https://doi.org/10.3390/s21124108

AMA Style

Chen M, Li M, Li Y, Yi W. Rock Particle Motion Information Detection Based on Video Instance Segmentation. Sensors. 2021; 21(12):4108. https://doi.org/10.3390/s21124108

Chicago/Turabian Style

Chen, Man, Maojun Li, Yiwei Li, and Wukun Yi. 2021. "Rock Particle Motion Information Detection Based on Video Instance Segmentation" Sensors 21, no. 12: 4108. https://doi.org/10.3390/s21124108

APA Style

Chen, M., Li, M., Li, Y., & Yi, W. (2021). Rock Particle Motion Information Detection Based on Video Instance Segmentation. Sensors, 21(12), 4108. https://doi.org/10.3390/s21124108

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop