Next Article in Journal
An Impedance Readout IC with Ratio-Based Measurement Techniques for Electrical Impedance Spectroscopy
Previous Article in Journal
A Robust Deep-Learning Model for Landslide Susceptibility Mapping: A Case Study of Kurdistan Province, Iran
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Method of Multitype Hybrid Rock Lithology Classification Based on Convolutional Neural Networks

School of Resources and Safety Engineering, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(4), 1574; https://doi.org/10.3390/s22041574
Submission received: 11 January 2022 / Revised: 5 February 2022 / Accepted: 15 February 2022 / Published: 17 February 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Rock lithology recognition plays a fundamental role in geological survey research, mineral resource exploration, mining engineering, etc. However, the objectivity of researchers, rock variable natures, and tedious experimental processes make it difficult to ensure the accurate and effective identification of rock lithology. Additionally, multitype hybrid rock lithology identification is challenging, and few studies on this issue are available. In this paper, a novel multitype hybrid rock lithology detection method was proposed based on convolutional neural network (CNN), and neural network model compression technology was adopted to guarantee the model inference efficiency. Four fundamental single class rock datasets: sandstone, shale, monzogranite, and tuff were collected. At the same time, multitype hybrid rock lithologies datasets were obtained based on data augmentation method. The proposed model was then trained on multitype hybrid rock lithologies datasets. Besides, for comparison purposes, the other three algorithms, were trained and evaluated. Experimental results revealed that our method exhibited the best performance in terms of precision, recall, and efficiency compared with the other three algorithms. Furthermore, the inference time of the proposed model is twice as fast as the other three methods. It only needs 11 milliseconds for single image detection, making it possible to be applied to the industry by transforming the algorithm to an embedded hardware device or Android platform.

1. Introduction

Rock lithology classification has always been an indispensable part of engineering fields. However, in-situ rock shows different physical and mechanical properties due to varying mineral compositions, geological mineralization conditions, and internal structures. It is significant for engineers to understand in-situ rock lithology accurately and efficiently prior to engineering design and construction, excavation, and support schedules. In the past, it mainly depended on physical or chemical analysis methods. Technicians classified in-situ rock types by observing rock mineral composition and crystalline structure through magnifying glasses. They also brought rock samples back to the laboratory, made thin sections, and then analyzed the internal structure under the microscope to finish rock type classification. In addition, chemical analysis is also employed for rock lithology classification. Among the aforementioned traditional methods, their final classification accuracy directly depends on the experiences and professionalism of technicians. Hence, enormous subjectivity exists. Moreover, the data preparation process is extremely tedious, time-consuming, and inefficient.
Recently, computer vision technologies and artificial intelligence (AI) have rapidly developed and are widely applied in our daily lives. There are two mainstream AI algorithms: machine learning and neural networks. Machine learning algorithms such as support vector machine (SVM), random forest (RF), decision tree (DT), and logistic regression (LR) aim to solve continuous variable prediction or classification problems. However, the scale and complexity of data that machine learning methods could address are relatively small and uncomplicated. In contrast, neural networks mainly imitate human beings’ biological neural instincts. The neural cells selectively retain active or inactive status to the input information, transmit this status signal to the subsequent adjacent neurons, and finally finish the response. Usually, the more data there are, the more complex of data uncertainties, nonlinearities, and interrelationships, while compared to machine learning, neural networks have powerful abilities of self-learning and automatic feature extraction on big data problems. Therefore, it has quickly been promoted over the last decades.
Feedforward neural network (FNN), convolutional neural network (CNN), recurrent neural network (RNN), and generative adversarial network (GAN) constitute the basic format of neural networks. However, CNN is one of the most successful and widely used methods. The applications include the earliest MNIST handwritten digital numbers identification [1], Cifar10 image classification [2], object detection [3,4,5,6,7] on open-source dataset: the PASCAL Visual Object Classes Homepage (Pascal VOC; ox.ac.uk)) and Common Objects in Context (COCO; cocodataset.org. accessed 13 December 2021)), face recognition [8], natural language processing [9,10], remote sensing [11,12], autonomous driving perception technologies [13], as well as industrial equipment fault detection, and medical CT analysis [14,15,16], etc.
CNN shows excellent performance on the image process aspect. Many researchers are combining CNN with empirical methods or numerical simulation methods to address rock mechanics engineering problems, and the final results proved to be more scientific and optimized. Karimpouli et al. [17,18,19,20] combined CNN to estimate rock physical properties. Chen et al. [21,22,23] studied landslide automatic recognition with satellite imagery based on CNN. Dong [24] diagnosed structural tunnel damage through CNN, and Yang [25] optimized TBM parameters by adopting CNN to analyze rock fragment size. Kovačević et al. [26,27] proposed a CNN to predict tunnel deformation and slope susceptibility. In addition, rock lithology automatic classification has also attracted the attention of researchers in recent years, and the existing experimental results illustrated that the overall classification performance based on CNN is more robust than traditional methods. However, rock lithology automatic classification research has undergone three development stages.
In the first stage, researchers used thin section images or features extracted from thin-section images or microscopic images as the input of the convolutional neural network. Cheng [28] applied CNN to recognize three types of sandstones of different granularities with 98.5% precision based on thin-section images. Singh [29] used thin section texture features to identify different basalt rock samples, and the neural network input is 27-dimensional numerical features extracted from RGB or grayscale thin section images, in this way, the accuracy could also reach 92.22%. Anjos [30] achieved three different types of carbonate rock identification by micro-CT images, and the best performance was over 81.33% precision. Li [31] used a transferred TrAdaBoost method to solve four interregional sandstone microscopic image classifications. Marmo [32] trained a multilayer perception neural network to identify four types of Dunham carbonate. They used numerical methods to extract 23-dimensional features from thin-section images, and the method showed 93% precision. Su [33] trained three neural network models based on thin-section images and assigned different weights to each model, and then the combined result of three models was viewed as the final output label.
In recent years, image processing technology has become increasingly reliable, and CNN has been directly adopted to classify single-type rock images. Wang [34] realized a lightweight neural network algorithm for identifying rock images based on MobileNets [35], and the method can accurately classify 25 single-type rocks with 93.45% precision. Ran [36] proposed RTCNNs to identify six typical rock types based on CNN. They cut the original high-resolution image into several patches, the input size was defined as 128 × 128 × 3, and the final classification accuracy was 97.96%. ShuffleNet [37], a commonly used lightweight convolutional neural network, was transferred to recognize rock lithology [38]. Wang [39] introduced CNN to realize the identification of four types of slope rock, and the test dataset accuracy was 90%. In Mars exploration, CNN was also adopted. Li [40] used VGG16 as a backbone network to classify four groups of Martian rocks, and the accuracy achieved approximately 100% on the test dataset. Pham [41] used a deep residual neural network (ResNet) and combined some data augmentation technologies to identify ten typical rock types with an overall accuracy of 84% on the test dataset. Fan [42] performed a comparison experiment of two standard convolutional neural networks, SqueezeNet and MobileNet, and the classification results were 94.55% and 93.27%, respectively, on 28 kinds of single-class rock.
Thirdly, different from the previous two methods, Liu [43] realized the precise and intelligent identification of rock types by using the object detection method. Object detection needs to not only detect the location of all objects in an image but also classify all targets. Liu [43] used Faster R-CNN [7], a deep learning neural network method, and achieved single-type rock recognition with 96% precision, while for hybrid multitype rock detection, the accuracy was only over 80%. Xu [44] also adopted the Faster R-CNN architecture and ResNet structure to classify 30 types of rock lithologies, and the accuracy was over 93.916%, but it was also a single-type image.
In this paper, a novel convolutional neural network named RDNet was developed for automatic detection of multiple types of mixed rock lithologies. The proposed method was optimized based on YOLO-V3 [4], and spatial pyramid pooling (SPP) [45] structure, which was added to detect multiscale objects as much as possible. Furthermore, neural network model compression technology was used to improve the model detection efficiency. In addition, a new data augmentation method was transferred to extend dataset diversity. For comparison purposes, the presented model and three other algorithms, including Faster R-CNN, YOLO-V3, and SSD [5], were trained based on the same four types of hybrid rock data: sandstone, shale, monzogranite, and tuff. Finally, the experimental results showed that our method (RDNet) exhibits excellent performance, and the model inference time is extremely fast up to real time, requiring only 11 milliseconds of single image detection. Consequently, it is feasible to transplant the algorithm to the embedded hardware device or Android platform to realize productization.

2. Materials and Methods

Algorithms and dataset are the prerequisites for the application and promotion of a neural network model. The illustration of algorithm and data used in this work are elaborated in the following sections.

2.1. Background

As mentioned above, convolutional neural networks (CNNs) have been widely used in many fields, including rock lithology classification. Further, object detection algorithms based on CNN possess stronger functions and broader applications.

2.1.1. Convolutional Neural Network

Convolutional neural network (CNN) was first proposed in the 1980s for handwritten digit number recognition and achieved excellent performance compared to conventional methods (linear classifier, radial basis function, K-nearest neighbor and SVM). After decades of development, CNN has achieved great success in many fields. Even though the application fields are different, it could be summarized that CNN could not only automatically learn features from (usually large-scale) data but also generalize the results to anonymous data of the same type with the same performance. CNN is usually composed of three main modules: convolutional layers, activation layers, and pooling layers.
Convolutional layers are responsible for obtaining feature maps, every convolutional layer includes several convolutional kernels, and each convolutional kernel has defined parameters (such as kernel width, height, and depth). As shown in Figure 1, the input shape is 6 × 6 × 3. There are four convolutional kernels of the convolutional layer, and each kernel generates a feature map. So, there are four feature maps.
Theoretically, the process of convolution is a matric calculation, as shown in Formula (1), and the detailed calculation steps are shown in Figure 2.
F 00 = { 1 × 0 + 1 × 1 + 0 × 0 + 1 × 0 + 0 × 0 + 0 × 1 + 0 × 1 + 1 × 2 + 1 × 2 1 × 0 + 1 × 2 + 0 × 0 + 0 × 1 + 0 × 0 + 0 × 0 + 1 × 0 + 0 × 2 0 × 1 + 0 × 0 + 1 × 1 + 0 × 0 + 1 × 2 + 0 × 1 + 1 × 1 + 1 × 0 + 1 × 2 F 00 = 5 + 2 + 0 = 3
Equation (1) mentioned above is a linear equation, while to improve the robustness of the CNN, it is necessary to add another layer, the activation layer. The activation layer is usually expressed as a kind of nonlinearity function, the basic form is a = σ z , and σ is an activation function. There are several kinds of activation functions as follows:
σ z = max 0 ,   z
σ z = 1 1 + e z
σ z = e z e z e z + e z
σ z = z , z 0 z a , z < 0
where Equations (2)–(5) are the relu, sigmoid, tanh, and leaky relu activation functions, respectively.
The pooling layer usually follows the activation layer, which means subsampling, which aims to decrease feature map resolution and further reduce parameters and computations. The pooling layer includes two formats: MaxPooling and AveragePooling. For the pooling layer, only two parameters, kernel size and sliding window stride, need to be defined. MaxPooling takes a maximum value over the kernel region, and in contrast, the mean value is for AveragePooling. Figure 3 shows the result of two pooling ways.

2.1.2. Object Detection Networks

There are two kinds of mainstream object detection algorithms based on convolutional neural networks: two-step and one-stage networks. The prominent representative of the two-stage algorithm is Faster R-CNN. For Faster R-CNN, the first step of the algorithm is to propose a certain number of candidate boxes, and the second step is to further optimize the regression of the candidate box location and the object classification. The two-stage algorithm ensures detection accuracy but sacrifices efficiency since the model parameters and the computations are relatively large. After Faster R-CNN was proposed, a new idea emerged: using convolutional neural networks to directly detect object boxes and predict the label, which is named one-stage network. It actually simplified the entire process compared to two-stage algorithm and then be quickly promoted, SSD and YOLO-V3 are both belong to this kind of methods. One-stage neural network inference speed is several times faster than that of the two-stage networks. Usually, the more parameters and calculations, the more computing resources are needed and the longer of the inference time. Detailed information on the model parameters and computations of Faster-RCNN, SSD, and YOLO-V3 are shown in Table 1.

2.2. Rock Lithology Identification

Most convolutional neural network parameters are enormous, ranging from millions to billions, so they all face high computing resource cost phenomena. As shown in Table 1, the Faster R-CNN, SSD, and YOLO-V3 computations are 149.25 GMac, 87.86 GMac, and 49.62 GMac, respectively.
The ground-truth (GT) box size and the ratio between width and height of the original dataset were comprehensively analyzed, as shown in Figure 4. In Figure 4a, the square root of the area is used to represent the size of the target box, and the y-axis represents the corresponding object box numbers. In Figure 4b, the x-axis is w/h ratio of object box, y-axis indicates the box numbers of different w/h ratios. It is evident that the distribution of all GT boxes is relatively single whether on boxes size or w/h ratio, thus it turns out to be that the multitype hybrid rock detection could be viewed as a simple task in some extent. Given the computing resource cost, this paper developed a novel method for multitype hybrid rock lithology detection based on YOLO-V3.

Developed Method

To begin, the SPP module rather than the way in the original YOLO-V3 structure was adopted as the new pattern of multiscale feature fusion. At the same time, the third pipeline of the YOLO-V3 network structure is removed, named Simplified-Net, and the structure is shown in Figure 5. The parameters and calculations of Simplified-Net and YOLO-V3 are listed in Table 2. It is obvious that the gap between them is slight, and the training results are depicted in the next section.
However, trials demonstrate that a significant proportion of parameters are redundant for CNN, their weight values are close to zero, and the importance of these kinds of neurons is thought to be negligible [46]. Thus, the overall performance of the model could not be affected if these kinds of neurons were deleted.
Therefore, in recent years, many scholars have tried to propose more lightweight convolutional neural networks by designing an innovative light architecture network or combining model compression technologies. Usually, the first method has more uncertainties and challenges and longer research cycles. In contrast, it is more effective to accelerate neural network inference speed through model compression technologies [47], such as model pruning, knowledge distillation, and parameter quantization. Model pruning is composed of unstructured pruning and structured pruning. Unstructured pruning refers to pruning for individual weights and structured pruning prunes for the channel or layer level. One of the disadvantages of the unstructured pruning method is that the weight matrix obtained is sparse, and acceleration effects cannot be achieved without dedicated hardware/library. In contrast, the structured pruning method performs pruning at the channel or layer level, and as a consequence, the network structure becomes more simplified after structured pruning, and the parameters and computations are also decreased to a large extent.
The simplified net proposed above still has many parameters and calculations, as shown in Table 2. Therefore, another branch was removed as well and keeping only one pipeline in the end. In addition, structured pruning technology combined with batch normalization [48] were adopted to further lighten it. Finally, a more lightweight network was obtained, named RDNet.
The batch normalization (BN) layer is usually between the convolutional and activation layers in the neural network structure, and the dominant goal is to normalize the input of the activation layer and then eliminate the influence of abnormal data during the model training process. The normalization calculation process is as Equation (6):
μ B = 1 m i = 1 m x i σ B 2 = 1 m i = 1 m x i μ B 2 x ^ i = x i μ B σ B 2 + ε y i = γ x ^ i + β
where μ B and σ B 2 are the mean values and variances, γ is the scale factor, and β is the translation factor.
Scale factor γ is selected as the reference value for structured pruning, and the depth of the γ parameters is consistent with the feature map channels of the convolutional layer output. The entire pruning process is as follows: (1) collect γ parameters of all BN layers in a neural network and sort γ values in ascending order way; (2) set a pruning rate (for example, 0.7 or others), which indicates the proportion of convolutional layer channels that will be pruned; (3) get the pruned threshold value. The total pruned numbers are equal to the multiplication of the total amount of all γ values and the pruning rate, and then the pruned threshold value is defined as the value of total pruned numbers index over sorted γ values; (4) compare γ values of all BN layers with the threshold value and record the channel index where γ value is smaller than the threshold value. For example, as shown in Figure 6, (I) denotes the feature map channels of the i-th convolutional layer, and (II) indicates the γ parameters of the i-th batch normalization layer. It can be seen that the second and third γ values are smaller than the threshold value (threshold is 0.5 in this paper), thus, the second and third γ values are saved as the pruning index, and then the recorded index is used to prune convolutional kernels. As shown in (III), channels C2 and C3 of the i-th convolutional layer feature map could be deleted, in turn, the i-1-th convolutional layer kernels could be reduced from N to N-2. Similarly, other convolutional layers are updated in the same way, and then the entire network structure is lightweighted.
The detailed numbers of convolutional kernels of YOLO-V3, Simplified-Net, and RDNet are shown in Figure 7, and the corresponding model parameters and calculations are exhibited in Table 3. It can be seen that the Simplified-Net only has one branch less than YOLO-V3, so the parameters and calculations are almost the same. RDNet has not only two branches less than YOLO-V3 but also convolutional kernels pruned to a considerable extent, as shown in Table 3. The parameters and computations are reduced almost 20 times compared to YOLO-V3 and Simplified-Net.

2.3. Database Description

In this study, the fundamental four single class rock datasets were acquired from the National Infrastructure of Mineral, Rock and Fossil for Science and Technology of China, including sandstone, shale, monzogranite and tuff, as shown in Figure 8. And multitype hybrid rock lithologies dataset was obtained through data augmentation technology, which will depict in the later section, based on these fundamental datasets. Finally, the total datasets were split into three parts, training data, validation data and test data, at a ratio of 6:3:1.

2.3.1. Data Process

To obtain a robust object detection model based on CNN, training dataset should be prepared to train the neural network model. LabelImg, an open-source tool, is widely selected to label ground-truth (GT) information of all training datasets. The labeled GT information includes object location and class, and the object location consists of two points of the bounding box, top left (x1, y1) and bottom right (x2, y2), as shown in Figure 9.
In our method, arbitrary image size is acceptable in the process of training or testing since a series of preprocessing steps were conducted, therefore the final input shape will be unified resize to the defined input size (in this paper, the input size is defined as 512 × 512 × 3). The preprocessing steps are as follows: (1) Resize, firstly calculating the maximum of the defined input size divided by the max value of original image width and height to get the scale_ratio. For example, if the original input shape is 720 × 640 × 3, the scale_ratio is equal to max(512, 512)/max(720, 640), and the scale ratio is 0.71. Then original width and height multiplied by the scale_ratio gives the resized shape. Also taking the above as an example, the resized image shape is (512, 455, 3). (2) Padding, using the defined input width and height subtracted from the resized image shape, respectively, and then the deviation of width and height between them was obtained (the width and height deviation is (0, 57), respectively), padding the deviation with zero pixel values and the final image shape turns out to be (512, 512, 3).

2.3.2. Multitype Hybrid Rock Lithologies Datasets

Diverse apparent characteristics of in-situ rock, such as different illumination, brightness, blurriness and so on, are considered. To ensure the generalization of the trained model as much as possible, it is necessary to take advantage of the existing data augmentation technologies to enhance dataset complexity.
Accompanying the development of deep learning techniques, many bag of tricks have been put forward consecutively to obtain better feedback [49]. A simple data augmentation principle was introduced by [50], it randomly mixed two samples under a ratio, but it is more suitable for classification tasks than object detection. Similar to occlusion, another method (cutout) was proposed by [51]. A square area, with a defined size and filled in with zero-pixel values, was pasted on the original image at any location to make the neural network learn global feature information instead of local features. The method also has good performance on the CIFAR10 and CIFAR-100 classification tasks. Rather than imitating occlusion, Yun [52] proposed a cut-mix augmentation strategy, it cut a part of the area of other training images and pasted on another image. It can not only be used for classification but also has consistent performance gains in detection. Bochkovskiy [3] put a new method by using image stitching technology, four random images were stitched together, which significantly expanded the diversity and complexity of data, and meanwhile solved the unbalance for data distribution.
In this paper, based on the provided four fundamental single class rock datasets, online data augmentation was conducted during the training process based on the final data augmentation method mentioned above to obtain multitype hybrid rock lithologies datasets. Figure 10 shows the four kinds of data augmentation results.

3. Experimental Results and Discussion

YOLO-V3, Simplified-Net, RDNet, RDNet + Aug, Faster R-CNN and SSD, together six models mentioned in this paper were all trained and validated under the same dataset. Where, YOLO-V3, Simplified-Net, RDNet and RDNet + Aug used the same parameters setting, Faster R-CNN and SSD retained the algorithms default parameters.

3.1. Training

The models of YOLO-V3, Simplified-Net, RDNet and RDNet + Aug were trained under the PyTorch framework using the Stochastic gradient descent (SGD) optimizer. The initial learning rate (lr) is 1 × 104, and the learning rate is updated by a decreasing factor, which is expressed as follows:
n e w l r = λ × i n i t i a l _ l r
λ = i 500 , i 500 1.0 , 500 < i 10,000 0.3 , 10,000 < i 20,000 0.1 , i > 20,000
where λ indicates the decreasing factor and i is the iteration number.
In addition, RTX3090 GPU is applied to train the model, the input data size is 512 × 512 × 3, the batch size is 32, and the number of training iterations is 60,000.

3.2. Results Analysis

The results of YOLO-V3, Simplified-Net, RDNet, RDNet + Aug as well as other two algorithms like Faster R-CNN and SSD were analyzed and compared in the next section.

3.2.1. Evaluation Metrics

Precision (P) and recall (R) are the major indicators for validating the performance of an object detection model. The intersection over union (IOU) indicates the area overlap ratio between two rectangular boxes, and the calculation equation is I O U = S A S B / S A S B . Only when the IOU between GTs is greater than the threshold can the prediction result be marked as correctly detected, otherwise, error is detected, and the IOU threshold value in this paper is set to 0.5. Each rock lithology has its own precision and recall evaluation indicators P and R, and precision equals the number of truly correctly detected objects of a certain rock type divided by the sum numbers of detected objects that are marked as this type rock lithology. Recall equals the number of truly correctly detected objects of a certain rock type divided by the GT numbers in this type. The calculation formula is as follows:
P = T P T P + F P
R = T P T P + F N
where true positive (TP) is the number of prediction results marked as correctly detected in this type, false positive (FP) is the number of prediction results actually belonging to another type but marked as this type, which is viewed as incorrect detection, and false negative (FN) is the number of prediction results marked as other types instead of this type. Therefore, in this paper, precision (P) and recall (R) are used as tools to measure the overall performance of the model.

3.2.2. YOLO-V3 and Simplified-Net

Firstly, with iterative training, the validation results of YOLO-V3 and Simplified-Net on four types of rock lithologies are shown in Figure 11. The x-axis indicates training iterations, the total number of iterations is 60,000, the y-axis indicates P and R, the solid curve is the YOLO-V3 results, and the dotted line is the Simplified-Net.
Table 4 exhibits the best model performance of YOLO-V3 and Simplified-Net on all rock types. The parameters, calculations, and model inference speed between YOLO-V3 and Simplified-Net are shown in Table 5.
The experimental results revealed that the gap between the Simplified-Net and YOLO-V3 on the four types of rock lithologies is negligible, and the average precision error and recall error are 2.3% and 1.25% respectively, proving that this kind of simplified approach is feasible.

3.2.3. Simplified-Net and RDNet

Secondly, with iterative training, the evaluation effects of Simplified-Net and RDNet on four types of rock lithologies are shown in Figure 12, where the solid curve and dotted line are the results of Simplified-Net and RDNet, respectively. The best model performances of Simplified-Net and RDNet on all rock types are compared in Table 6. The model parameters, computations, and model inference speed are shown in Table 7.
According to the final experimental results in Figure 12 and Table 7, it is clear that there exists a certain disparity between Simplified-Net and RDNet, and the average precision error and recall error are 4.775% and 5.45% respectively, while, considering that RDNet was further simplified and pruned based on Simplified-Net mentioned in Section 2.3.1, the structure of RDNet has been greatly simplified. Table 8 shows that the parameters and computations of RDNet are reduced almost 20 times, and the inference time is shortened by half, requiring only 11 milliseconds for single image detection. Therefore, it is reasonable for a network to have a certain discount on overall performance when it is largely pruned.
In other words, the structured pruning technology on our task is acceptable, it not only guarantees that the precision is not greatly affected but also reduces the parameters and calculations.

3.2.4. RDNet and RDNet + Aug

Thirdly, data augmentation skills are merged into the preprocessing module, and trials have been conducted based on RDNet. The results are summarized in Figure 13 and Table 8. The solid curve is RDNet + Aug, and the dotted line is RDNet without data augmentation. It is obvious that whether on precision or recall, combined with data augmentation skills, the model performance obtained significant improvement, the average precision for detection of four types of rock is over 10%, and the recall is improved almost 2-fold. Since, data augmentation skills are integrated in the data preprocessing phase, no parameters or calculations are added to the network, as shown in Table 9.

3.2.5. RDNet + Aug and Other Models

The two-stage object detection algorithm Faster R-CNN and one-stage network SSD are also trained on the same dataset, and with the same training iterations, both of their results are evaluated and compared with ours. As shown in Figure 14 and Figure 15, dotted lines represent our method, solid lines are SSD and Faster R-CNN, respectively, where the x-axis indicates training iterations and the y-axis indicates three evaluation indicators. It is obvious that our method performances more stable on aspect of all evaluation indicators, while SSD and Faster R-CNN has higher vibration and variance on four types of rock data. The best effects for our method, SSD and Faster R-CNN on four types of rock data are listed in Table 10, and the compared detection results are shown in Figure 16.
It can be summarized from Table 10 that among the three algorithms, SSD has the lowest recall on the test dataset, and many objects were missed, while Faster R-CNN is more sensitive to apparent rock characteristics, therefore, the precision is not good compared to SSD, especially on monzogranite, shale and tuff classes.
Meanwhile, it is clear that our method achieved the best stability on four kinds of rock datasets, and the precision is 10%~30% higher than Faster R-CNN and SSD. In addition, the recall also performs better. Furthermore, the inference speed is twice as fast as s Faster R-CNN and SSD, and only 11 ms is needed for single image detection.

3.3. Discussion

The total experimental results of YOLO-V3, Simplified-Net, RDNet, RDNet + Aug, Faster R-CNN and SSD were summarized. The best model performance of each algorithm on four type of rock lithologies is shown in Table 11. It can be concluded that YOLO-V3, Simplified-Net and RDNet possess almost the same evaluation results. The average precision on the four types of rock data is 75.325%, 77.325%, and 73.45%, respectively, and the average recall is 34.425%, 34.025%, and 52.63%, which demonstrates that the initial simplification method and the compression technology used in this paper are practical. On the other hand, low recall reveals that multitype hybrid rock lithology detection is challenging.
In addition, the average precision of RDNet (73.45%) is higher than that of Faster R-CNN (65.6%), and lower than that of SSD (80.05%), and the average recall of RDNet (52.63%) is lower than that of Faster R-CNN (75.92%) and SSD (56.5%), which illustrates that RDNet has no advantages compared to Faster R-CNN and SSD.
Combined with the data augmentation technology, RDNet + Aug achieved 82.1% average accuracy (higher than RDNet 73.45%), and 78% average recall (higher than RDNet 52.63%). Figure 16 shows the comparison results of the labeled information of GT (the yellow box is the labeled targets box, the top left is the corresponding label) and the detection results (blue box) of RDNet + Aug, Faster R-CNN and SSD.
The comparison of validation results between RDNet + Aug, Faster R-CNN, SSD, and YOLO-V3 is shown in Figure 17, which indicates that a suitable data augmentation method is of great importance for training the convolutional neural network model.
It is also worth noting that in addition to the more stable detection performance, RDNet + Aug still has obvious advantages on parameters and calculations. As is shown in Table 12, the parameters and calculations of RDNet + Aug were reduced almost 20 times compared to YOLO-V3 and far less than Faster R-CNN and SSD. The inference speed is only 11 milliseconds for a single image detection, which is shortened by half compared to the others.
In the future study, multitype hybrid rock lithologies detection under complicated environment is still the key issue. High quality lithology database is needed to further expand the types of lithology and the number of each type of lithology. With the optimized network, the stability and generalization performance of the rock lithology detection model can show a big improvement.

4. Conclusions

In this paper, a novel intelligent multitype rock classification method was proposed based on a convolutional neural network, and the main conclusions are as follows:
(1)
Based on the YOLOv3 network and model compression technology, an extremely lightened neural network named RDNet was proposed. Trials indicate that the performance of the pruned model is acceptable and not influenced to a large extent, but the efficiency is greatly improved. The parameters and computations are reduced almost 20-fold, and the forward speed reaches 11 milliseconds for single image detection.
(2)
A data augmentation method was transferred to obtain a multitype rock lithology dataset, which solved the imbalance between classes and expanded the quantity of train datasets. The average precision of the trained model on four type rock hybrid datasets reaches 82.1%, the highest is 93.6% on monzogranite. Compared to the original YOLO-V3, the average accuracy is nearly 8% higher, And the average recall is almost double improved, which demonstrates that the model possesses excellent stability and remarkable generalization performance.
(3)
The other two mainstream object detection algorithms Faster R-CNN and SSD are also compared. The average accuracy of our method is 16.5% higher than that of Faster R-CNN and 2% higher than SSD, besides, the average recall of our method is 2% higher than that of Faster R-CNN and 21.5% higher than SSD. Furthermore, the inference speed is twice faster than that of Faster R-CNN and SSD.

Author Contributions

D.L. contributed to the creative idea, supervised the work and corrected the manuscript. J.Z. wrote the manuscript, implemented the training and validation code, and analyzed the results. Z.L. obtained the original dataset and processed it. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by the National Natural Science Foundation of China (Grant No.: 52074349), the Hunan Natural Science Foundation for Distinguished Young Scholars (Grant No.: 2019JJ20028).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Thanks to the public datasets used in this research. We also thank the reviewers for their comments and suggestions to improve the quality of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  2. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  3. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. Available online: https://arxiv.org/abs/2004.10934 (accessed on 21 October 2021).
  4. Farhadi, A.; Redmon, J. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. Available online: https://arxiv.org/abs/1804.02767 (accessed on 21 October 2021).
  5. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single shot MultiBox detector. In Computer Vision—ECCV 2016, Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  6. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [Green Version]
  8. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 8–12 June 2015; pp. 815–823. [Google Scholar]
  9. Hirschberg, J.; Manning, C.D. Advances in natural language processing. Science 2015, 349, 261–266. [Google Scholar] [CrossRef]
  10. Chowdhury, G.G. Natural language processing. Annu. Rev. Inf. Sci. Technol. 2003, 37, 51–89. [Google Scholar] [CrossRef] [Green Version]
  11. Khelifi, L.; Mignotte, M. Deep learning for change detection in remote sensing images: Comprehensive review and meta-analysis. IEEE Access 2020, 8, 126385–126400. [Google Scholar] [CrossRef]
  12. Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.-S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef] [Green Version]
  13. Huang, Y.; Chen, Y. Autonomous driving with deep learning: A survey of state-of-art technologies. arXiv 2020, arXiv:2006.06091. Available online: https://arxiv.org/abs/2006.06091 (accessed on 21 October 2021).
  14. Albarqouni, S.; Baur, C.; Achilles, F.; Belagiannis, V.; Demirci, S.; Navab, N. Aggnet: Deep learning from crowds for mitosis detection in breast cancer histology images. IEEE Trans. Med. Imaging 2016, 35, 1313–1321. [Google Scholar] [CrossRef]
  15. Yang, L.; Zhang, Y.; Chen, J.; Zhang, S.; Chen, D.Z. Suggestive annotation: A deep active learning framework for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; pp. 399–407. [Google Scholar]
  16. Zhu, W. Deep Learning for Automated Medical Image Analysis; University of California: Irvine, CA, USA, 2019. [Google Scholar]
  17. Karimpouli, S.; Tahmasebi, P. Image-based velocity estimation of rock using convolutional neural networks. Neural Netw. 2019, 111, 89–97. [Google Scholar] [CrossRef]
  18. Mha, B.; Zzab, C.; Jie, R.A.; Jh, A.; Gl, A.; Yc, A.; Ning, L. Deep convolutional neural network for fast determination of the rock strength parameters using drilling data. Int. J. Rock Mech. Min. Sci. 2019, 123, 104084. [Google Scholar]
  19. Han, S.; Li, H.; Li, M.; Luo, X. Measuring rock surface strength based on spectrograms with deep convolutional networks. Comput. Geosci. 2019, 133, 104312. [Google Scholar] [CrossRef]
  20. Han, S.; Li, H.; Li, M.; Rose, T. A deep learning based method for the non-destructive measuring of rock strength through hammering sound. Appl. Sci. 2019, 9, 3484. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, Z.; Zhang, Y.; Ouyang, C.; Zhang, F.; Ma, J. Automated landslides detection for mountain cities using multi-temporal remote sensing imagery. Sensors 2018, 18, 821. [Google Scholar] [CrossRef] [Green Version]
  22. Ding, A.; Zhang, Q.; Zhou, X.; Dai, B. Automatic recognition of landslide based on CNN and texture change detection. In Proceedings of the Chinese Association of Automation (YAC), Youth Academic Annual Conference, Wuhan, China, 11–13 November 2016; pp. 444–448. [Google Scholar]
  23. Yu, H.; Ma, Y.; Wang, L.; Zhai, Y.; Wang, X. A landslide intelligent detection method based on CNN and RSG_R. In Proceedings of the 2017 IEEE International Conference on Mechatronics and Automation (ICMA), Takamatsu, Japan, 6–9 August 2017; pp. 40–44. [Google Scholar]
  24. Dong, Y.; Wang, J.; Wang, Z.; Zhang, X.; Gao, Y.; Sui, Q.; Jiang, P. A deep-learning-based multiple defect detection method for tunnel lining damages. IEEE Access 2019, 7, 182643–182657. [Google Scholar] [CrossRef]
  25. Yang, Z.; He, B.; Liu, Y.; Wang, D.; Zhu, G. Classification of rock fragments produced by tunnel boring machine using convolutional neural networks. Autom. Constr. 2021, 125, 103612. [Google Scholar] [CrossRef]
  26. Kovačević, M.S.; Bačić, M.; Gavin, K.; Stipanović, I. Assessment of long-term deformation of a tunnel in soft rock by utilizing particle swarm optimized neural network. Tunn. Undergr. Space Technol. 2021, 110, 103838. [Google Scholar] [CrossRef]
  27. Fang, Z.; Wang, Y.; Peng, L.; Hong, H. Integration of convolutional neural network and conventional machine learning classifiers for landslide susceptibility mapping. Comput. Geosci. 2020, 139, 104470. [Google Scholar] [CrossRef]
  28. Cheng, G.; Guo, W. Rock images classification by using deep convolution neural network. J. Phys. Conf. Ser. 2017, 887, 012089. [Google Scholar] [CrossRef] [Green Version]
  29. Singh, N.; Singh, T.; Tiwary, A.; Sarkar, K.M. Textural identification of basaltic rock mass using image processing and neural network. Comput. Geosci. 2010, 14, 301–310. [Google Scholar] [CrossRef]
  30. Anjos, C.; Avila, M.; Vasconcelos, A.; Neta, A.; Landau, L. Deep learning for lithological classification of carbonate rock micro-CT images. Comput. Geosci. 2021, 25, 971–983. [Google Scholar] [CrossRef]
  31. Li, N.; Hao, H.; Gu, Q.; Wang, D.; Hu, X. A transfer learning method for automatic identification of sandstone microscopic images. Comput. Geosci. 2017, 103, 111–121. [Google Scholar] [CrossRef]
  32. Marmo, R.; Amodio, S.; Tagliaferri, R.; Ferreri, V.; Longo, G. Textural identification of carbonate rocks by image processing and neural network: Methodology proposal and examples. Comput. Geosci. 2005, 31, 649–659. [Google Scholar] [CrossRef]
  33. Su, C.; Xu, S.-j.; Zhu, K.-y.; Zhang, X.-c. Rock classification in petrographic thin section images based on concatenated convolutional neural networks. Earth Sci. Inform. 2020, 13, 1477–1484. [Google Scholar] [CrossRef]
  34. Wang, C.; Li, Y.; Fan, G.; Chen, F.; Wang, W. Quick Recognition of Rock Images for Mobile Applications. J. Eng. Sci. Technol. Rev. 2018, 11, 111–117. [Google Scholar] [CrossRef]
  35. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. Available online: https://arxiv.org/abs/1704.04861 (accessed on 22 October 2021).
  36. Ran, X.; Xue, L.; Zhang, Y.; Liu, Z.; Sang, X.; He, J. Rock classification from field image patches analyzed using a deep convolutional neural network. Mathematics 2019, 7, 755. [Google Scholar] [CrossRef] [Green Version]
  37. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  38. Fan, G.; Chen, F.; Chen, D.; Li, Y.; Dong, Y. A Deep Learning Model for Quick and Accurate Rock Recognition with Smartphones. Mob. Inf. Syst. 2020, 2020, 7462524. [Google Scholar] [CrossRef]
  39. Wang, P.; Wang, S.; Zhu, C.; Zhang, Z. Classification and extent determination of rock slope using deep learning. Geomech. Geophys. Geo-Energy Geo-Resour. 2020, 6, 33. [Google Scholar] [CrossRef]
  40. Li, J.; Zhang, L.; Wu, Z.; Ling, Z.; Cao, X.; Guo, K.; Yan, F. Autonomous Martian rock image classification based on transfer deep learning methods. Earth Sci. Inform. 2020, 13, 951–963. [Google Scholar] [CrossRef]
  41. Pham, C.; Shin, H.-S. A Feasibility Study on Application of a Deep Convolutional Neural Network for Automatic Rock Type Classification. Tunn. Undergr. Space 2020, 30, 462–472. [Google Scholar]
  42. Fan, G.; Chen, F.; Chen, D.; Dong, Y. Recognizing Multiple Types of Rocks Quickly and Accurately Based on Lightweight CNNs Model. IEEE Access 2020, 8, 55269–55278. [Google Scholar] [CrossRef]
  43. Liu, X.; Wang, H.; Jing, H.; Shao, A.; Wang, L. Research on intelligent identification of rock types based on faster R-CNN method. IEEE Access 2020, 8, 21804–21812. [Google Scholar] [CrossRef]
  44. Xu, Z.; Ma, W.; Lin, P.; Shi, H.; Pan, D.; Liu, T. Deep learning of rock images for intelligent lithology identification. Comput. Geosci. 2021, 154, 104799. [Google Scholar] [CrossRef]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Transact. Pattern Anal. Machine Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [Green Version]
  46. Zhuang, L.; Li, J.; Shen, Z.; Gao, H.; Zhang, C. Learning Efficient Convolutional Networks through Network Slimming. In Proceedings of the IEEE Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2736–2744. [Google Scholar]
  47. Cheng, Y.; Wang, D.; Zhou, P.; Zhang, T. A survey of model compression and acceleration for deep neural networks. arXiv 2017, arXiv:1710.09282. Available online: https://arxiv.org/abs/1710.09282 (accessed on 21 October 2021).
  48. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 7–9 July 2015; Volume 37, pp. 448–456. [Google Scholar]
  49. He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of Tricks for Image Classification with Convolutional Neural Networks. CoRR 2018. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 558–567. [Google Scholar]
  50. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. Available online: https://arxiv.org/abs/1710.09412 (accessed on 21 October 2021).
  51. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with cutout. arXiv 2017, arXiv:1708.04552. Available online: https://arxiv.org/abs/1708.04552 (accessed on 21 October 2021).
  52. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
Figure 1. Schematic diagram of the convolutional layer.
Figure 1. Schematic diagram of the convolutional layer.
Sensors 22 01574 g001
Figure 2. The process of convolution calculation.
Figure 2. The process of convolution calculation.
Sensors 22 01574 g002
Figure 3. Two forms of pooling ways. MaxPooling takes the maximum value among the four dotted box and AveragePooling takes the average value.
Figure 3. Two forms of pooling ways. MaxPooling takes the maximum value among the four dotted box and AveragePooling takes the average value.
Sensors 22 01574 g003
Figure 4. Data distribution. (a) the distribution of object box size; (b) object w/h ratio distribution.
Figure 4. Data distribution. (a) the distribution of object box size; (b) object w/h ratio distribution.
Sensors 22 01574 g004
Figure 5. Simplified-Net Flow chart. (a) is the Res_Block module and (b) is Simplified-Net network structure.
Figure 5. Simplified-Net Flow chart. (a) is the Res_Block module and (b) is Simplified-Net network structure.
Sensors 22 01574 g005
Figure 6. I, II, and III indicate the i-th feature map channels, γ parameters of the i-th batch normalization layer, and the i-th feature map channels which could be deleted, respectively.
Figure 6. I, II, and III indicate the i-th feature map channels, γ parameters of the i-th batch normalization layer, and the i-th feature map channels which could be deleted, respectively.
Sensors 22 01574 g006
Figure 7. Comparison of convolutional kernels parameters of YOLO-V3, Simplified-Net and RDNet.
Figure 7. Comparison of convolutional kernels parameters of YOLO-V3, Simplified-Net and RDNet.
Sensors 22 01574 g007
Figure 8. All types rock data distribution.
Figure 8. All types rock data distribution.
Sensors 22 01574 g008
Figure 9. Labeled ground-truth(GT) information, including object position(x1,y1),(x2,y2) and rock class: shale.
Figure 9. Labeled ground-truth(GT) information, including object position(x1,y1),(x2,y2) and rock class: shale.
Sensors 22 01574 g009
Figure 10. Four kinds of data augmentation results. (a) mix-up, (b) cut-out, (c) cut-mix, and (d) stitching methods.
Figure 10. Four kinds of data augmentation results. (a) mix-up, (b) cut-out, (c) cut-mix, and (d) stitching methods.
Sensors 22 01574 g010
Figure 11. YOLO-V3 and Simplified-Net evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Figure 11. YOLO-V3 and Simplified-Net evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Sensors 22 01574 g011
Figure 12. Simplified-Net and RDNet evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Figure 12. Simplified-Net and RDNet evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Sensors 22 01574 g012aSensors 22 01574 g012b
Figure 13. RDNet and RDNet + Aug evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Figure 13. RDNet and RDNet + Aug evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Sensors 22 01574 g013
Figure 14. RDNet + Aug and SSD evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Figure 14. RDNet + Aug and SSD evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Sensors 22 01574 g014
Figure 15. RDNet + Aug and Faster R-CNN evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Figure 15. RDNet + Aug and Faster R-CNN evaluation results. (a) monzogranite, (b) sandstone, (c) shale, and (d) tuff.
Sensors 22 01574 g015
Figure 16. Comparison of detection results of part test datasets. (a,e) show the labeled ground-truth of targets. (b,f) is the detection result obtained by using improved method in this paper (RDNet + Aug). (c,g) is the detection result of Faster R-CNN method. (d,h) is the detection result of SSD method.
Figure 16. Comparison of detection results of part test datasets. (a,e) show the labeled ground-truth of targets. (b,f) is the detection result obtained by using improved method in this paper (RDNet + Aug). (c,g) is the detection result of Faster R-CNN method. (d,h) is the detection result of SSD method.
Sensors 22 01574 g016
Figure 17. Comparison results of different models. (a) average precision on four multitype rock lithology of five models; (b) average recall on four multitype rock lithology of five models.
Figure 17. Comparison results of different models. (a) average precision on four multitype rock lithology of five models; (b) average recall on four multitype rock lithology of five models.
Sensors 22 01574 g017
Table 1. Parameters and calculations of different models.
Table 1. Parameters and calculations of different models.
ModelsFaster R-CNNSSD YOLO-V3
Parameters136.46 M23.88 M61.52 M
Computations149.25 GMac87.86 GMac49.62 GMac
Note: M represents the number of CNN parameters, GMac indicates the Computational complexity.
Table 2. Comparison between original and simplified network.
Table 2. Comparison between original and simplified network.
ModelsYOLO-V3Simplified-Net
Parameters61.52 M60.48 M
Computations49.62 GMac45.46 GMac
Table 3. Comparison between YOLO-V3, Simplified-Net and RDNet.
Table 3. Comparison between YOLO-V3, Simplified-Net and RDNet.
ModelsYOLO-V3Simplified-Net RDNet
Parameters61.52 M60.48 M3.84 M
Computations49.62 GMac45.46 GMac3.38 GMac
Table 4. The best model performance of YOLO-V3 and Simplified-Net on four types rock lithology.
Table 4. The best model performance of YOLO-V3 and Simplified-Net on four types rock lithology.
Rock TypesEvaluationMethods
YOLO-V3Simplified-Net
MonzograniteP75.9%77.2%
R49.2%47.1%
SandstoneP71.8%71.2%
R24.3%24.8%
ShaleP83.1%83.9%
R35.2%34.0%
TuffP70.5%77.0%
R29.0%30.2%
Table 5. Model parameters, calculations, and inference time of YOLO-V3 and Simplified-Net.
Table 5. Model parameters, calculations, and inference time of YOLO-V3 and Simplified-Net.
Speed PerformanceMethods
YOLO-V3Simplified-Net
Model parameters61.52 M60.48 M
Model calculations49.62 GMac45.46 GMac
Inference time (ms)2119
Table 6. The best model performance of Simplified-Net and RDNet on four types rock lithology.
Table 6. The best model performance of Simplified-Net and RDNet on four types rock lithology.
Rock TypesEvaluationMethods
Simplified-NetRDNet
MonzograniteP77.2%69.3%
R47.1%53.7%
SandstoneP71.2%65.7%
R24.8%30.4%
ShaleP83.9%85.7%
R34.0%41.5%
TuffP77.0%73.1%
R30.2%32.3%
Table 7. Model parameters, calculations, and inference time of Simplified-Net and RDNet.
Table 7. Model parameters, calculations, and inference time of Simplified-Net and RDNet.
Speed PerformanceMethods
Simplified-NetRDNet
Model parameters60.48 M3.84 M
Model calculations45.46 GMac3.38 GMac
Inference time (ms)1911
Table 8. The best model performance of RDNet and RDNet + Aug on four types rock lithology.
Table 8. The best model performance of RDNet and RDNet + Aug on four types rock lithology.
Rock TypesEvaluationMethods
RDNetRDNet + Aug
MonzograniteP69.3%93.6%
R53.7%77.9%
SandstoneP65.7%75.2%
R30.4%79.1%
ShaleP85.7%81.0%
R41.5%83.0%
TuffP73.1%78.5%
R32.3%72.0%
Table 9. Model parameters, calculations, and inference time of RDNet and RDNet + Aug.
Table 9. Model parameters, calculations, and inference time of RDNet and RDNet + Aug.
Speed PerformanceMethods
RDNetRDNet + Aug
Model parameters3.84 M3.84 M
Model calculations3.38 GMac3.38 GMac
Inference time (ms)1111
Table 10. The best model performance of Faster R-CNN, SSD, and RDNet + Aug on four types rock lithology.
Table 10. The best model performance of Faster R-CNN, SSD, and RDNet + Aug on four types rock lithology.
Rock TypesEvaluationMethods
Faster R-CNNSSDRDNet + Aug
MonzograniteP61.5%88.5%93.6%
R85.7%56.9%77.9%
SandstoneP63.7%69.2%75.2%
R67.2%69.8%79.1%
ShaleP71.6%82.1%81.0%
R81.9%56.1%83.0%
TuffP65.6%80.4%78.5%
R68.9%43.2%72.0%
Inference
time (ms)
242211
Table 11. Evaluation results of different methods.
Table 11. Evaluation results of different methods.
Rock TypesEvaluationMethods
YOLO-V3Simplified-NetRDNetRDNet + AugFaster R-CNNSSD
MonzograniteP75.9%77.2%69.3%93.6%61.5%88.5%
R49.2%47.1%53.7%77.9%85.7%56.9%
SandstoneP71.8%71.2%65.7%75.2%63.7%69.2%
R24.3%24.8%30.4%79.1%67.2%69.8%
ShaleP83.1%83.9%85.7%81.0%71.6%82.1%
R35.2%34.0%41.5%83.0%81.9%56.1%
TuffP70.5%77.0%73.1%78.5%65.6%80.4%
R29.0%30.2%32.3%72.0%68.9%43.2%
Table 12. Parameters, calculations and inference time of different methods.
Table 12. Parameters, calculations and inference time of different methods.
Speed PerformanceMethods
YOLO-V3Simplified-NetRDNetRDNet + AugFaster R-CNNSSD
Parameters61.52 M60.48 M3.84 M3.84 M136.46 M23.88 M
Calculations49.62 GMac45.46 GMac3.38 GMac3.38 GMac149.25 GMac87.86 GMac
Inference time (ms)211911112422
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, D.; Zhao, J.; Liu, Z. A Novel Method of Multitype Hybrid Rock Lithology Classification Based on Convolutional Neural Networks. Sensors 2022, 22, 1574. https://doi.org/10.3390/s22041574

AMA Style

Li D, Zhao J, Liu Z. A Novel Method of Multitype Hybrid Rock Lithology Classification Based on Convolutional Neural Networks. Sensors. 2022; 22(4):1574. https://doi.org/10.3390/s22041574

Chicago/Turabian Style

Li, Diyuan, Junjie Zhao, and Zida Liu. 2022. "A Novel Method of Multitype Hybrid Rock Lithology Classification Based on Convolutional Neural Networks" Sensors 22, no. 4: 1574. https://doi.org/10.3390/s22041574

APA Style

Li, D., Zhao, J., & Liu, Z. (2022). A Novel Method of Multitype Hybrid Rock Lithology Classification Based on Convolutional Neural Networks. Sensors, 22(4), 1574. https://doi.org/10.3390/s22041574

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop