Next Article in Journal
Using Wearable Sensors to Measure Goal Achievement in Older Veterans with Dementia
Next Article in Special Issue
Novel Meta-Learning Techniques for the Multiclass Image Classification Problem
Previous Article in Journal
An Open Surface Drifter for River Flow Field Characterization
Previous Article in Special Issue
Dimension Reduction of Digital Image Descriptors in Neural Identification of Damaged Malting Barley Grains
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Defect Severity Identification for a Catenary System Based on Deep Semantic Learning

1
School of Electrical Engineering, Southwest Jiaotong University, Chengdu 610031, China
2
College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China
3
Institute of Oceanographic Instrumentation, Qilu University of Technology (Shandong Academy of Sciences), Qingdao 266075, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(24), 9922; https://doi.org/10.3390/s22249922
Submission received: 26 November 2022 / Revised: 10 December 2022 / Accepted: 14 December 2022 / Published: 16 December 2022
(This article belongs to the Special Issue Advances in Deep Learning for Intelligent Sensing Systems)

Abstract

:
A variety of Chinese textual operational text data has been recorded during the operation and maintenance of the high-speed railway catenary system. Such defect text records can facilitate defect detection and defect severity analysis if mined efficiently and accurately. Therefore, in this context, this paper focuses on a specific problem in defect text mining, which is to efficiently extract defect-relevant information from catenary defect text records and automatically identify catenary defect severity. The specific task is transformed into a machine learning problem for defect text classification. First, we summarize the characteristics of catenary defect texts and construct a text dataset. Second, we use BERT to learn defect texts and generate word embedding vectors with contextual features, fed into the classification model. Third, we developed a deep text categorization network (DTCN) to distinguish the catenary defect level, considering the contextualized semantic features. Finally, the effectiveness of our proposed method (BERT-DTCN) is validated using a catenary defect textual dataset collected from 2016 to 2018 in the China Railway Administration in Chengdu, Lanzhou, and Hengshui. Moreover, BERT-DTCN outperforms several competitive methods in terms of accuracy, precision, recall, and F 1 -score value.

1. Introduction

A pantograph–catenary system for a high-speed railway bridges the traction power supply system and electric locomotive, whose operating conditions are of great significance for the safety and reliability performances of railway transport [1]. In particular, a catenary system is a fixed installation that consists of multiple components (as shown in Figure 1), such as the pillar, contact wire, catenary wire, feeder line, etc. Owning to the complicated failure mechanisms and dynamic outdoor environments, the operational risks of the catenary are prone to inevitably increase [2,3]. To timely mitigate the underlying risk triggers, various monitoring techniques (e.g., an automatic detection and monitoring system, also called the 6C system) and maintenance activities [4] are implemented to discover and report all potential defects of a catenary system. At the same time, a large number of catenary defect texts have been recorded and collected after manual judgment. Such textual defect text records are closely associated with the characteristics of catenary accidents and failure risks. If mined efficiently and accurately, they can provide powerful and credible information bases for discovering valuable defect knowledge and efficient defect severity identification, facilitating the subsequent defect disposal and adjustment of relevant maintenance activities [5]. Therefore, it is crucial to mine defect text records in-depth, extract critical textual semantic information, and finally identify defect severity for a catenary system.
In pace with the rapid advancement of natural language processing (NLP) technology, Chinese text mining has attracted extensive attention [6,7]. In railway systems, the associated text mining technologies have become emerging concerns and are gradually applied in several classification tasks, such as extracting traffic-relevant information, fault type classification, analyses of rail accidents, etc. Chen et al. used the long short-term memory-convolutional neural network (LSTM-CNN) with bag-of-word features to judge whether social media data are related to traffic [8]. Wang et al. applied a multichannel network with a LSTM layer and a convolution layer (MC-LSTM-Conv) and keyword fuzzy matching to detect traffic events from microblogs (i.e., traffic jams versus non-traffic jams) [9]. In [10], the support vector machine (SVM) is used to classify fault class with respect to railway signaling maintenance. Brown et al. utilized ensemble methods to predict rail accident severity [11]. To summarize, with regard to operation and maintenance defect text record data of a catenary system, text mining techniques can be applied to provide more insights into helping establish the associations between the defect event description and defect severity degree, which can ultimately improve the efficiency and accuracy of defect severity identification.
In this paper, we present an investigation to efficiently extract contextual semantic information in-depth and automatically identify catenary defect severity levels based on operation and maintenance text records. To this end, the approach integrates a combination of bidirectional encoder representation from transformers (BERT) [12] and a deep text categorization network (DTCN) with deepened Conv layers. It can abstract long-range semantic features and more global data, which are more valuable to enable defect-level classifications. Firstly, we collected catenary defect text records from 2016 to 2018 in the China Railway Administration of Chengdu, Lanzhou, and Hengshui, and analyzed the source, content, and characteristics of catenary defect texts. Afterward, Chinese word vector representations were learned based on the deep pre-trained language model BERT via the constructed catenary defect textual data set. On the basis of the generated word embedding representations, the DTCN is proposed to identify catenary defect severity (i.e., defect level category). Compared with state-of-the-art methods for text classification, the proposed BERT-DTCN can capture long-range semantic features and utilize deep semantics to distinguish the defect level. The effectiveness and superiority of our proposal are demonstrated by experimental results in terms of accuracy, precision, recall, and the F 1 -score value.
The primary contributions of this paper are as follows:
  • A deep semantic neural network named BERT-DTCN is proposed to effectively extract long-range semantic features and automatically identify defect severity from catenary text records.
  • Different from existing domain text representation approaches that extract vector representations with bag-of-words features, we applied BERT to learn word embedding vectors and extract semantic features of domain vocabularies in defective texts. An ablation study on the constructed catenary defect text dataset validates that the generated word embedding vectors contribute to beneficial impacts on the devised text categorization model.
  • Based on the obtained defect word embeddings, we used the DTCN to distinguish defect severity degree. Experimental results demonstrate that the proposed algorithm (BERT-DTCN) achieves a superior performance in the binary classification problem (level 1 defect or level 2 defect) over competitive text classification methods, which can reduce the workload of manual discrimination and improve the accuracy and efficiency of classification.
The remainder of the paper is organized as follows: Section 2 introduces the related work on text presentation and classification. Section 3 overviews the proposed catenary defect text classification approach in terms of the defect text database, word embedding presentation learning, and classification model. Section 4 presents our experimental results and analysis of the proposed model. The conclusions and suggestions for future work are drawn in Section 5.

2. Related Work

This study integrates methods for the defect information analysis with operational records of the catenary, word embedding representations, and catenary defect text classification. Thus, this section mainly focuses on text representation and text classification techniques.

2.1. Text Classification Methods

The conventional process of text classification consists of text dataset acquisition, text preprocessing, text representation, and text classification models. Text preprocessing involves taking several measures to process textual data, including deleting stop words, word segmentation, and so on. Text representation mainly refers to the representation of converting words or sequences in a text into a form that can be recognized and handled by a machine for subsequent text classification tasks. Text classification involves utilizing the appropriate classifiers, including rule-based methods and machine learning or deep learning models, to excavate the critical features and predict the text category. In general, the development process of text classification methods can be categorized into three stages (as shown in Figure 2).
Firstly, the text classification process is investigated by human-made linguistic rules [13,14], constructing a set of rules connecting word patterns and class labels. Such human comprehensible rules can be convinced as time goes by. However, the whole constructing process consumes time and manpower and requires abundant domain knowledge [15]. However, it is not feasible to create a huge rule base for a specific domain task.
Secondly, several scholars have carried out relevant research on text mining based on traditional machine learning techniques, including naive Bayes [16], decision tree [17], SVM [18], etc., which depends on learning classification features from a pre-labeled dataset. Although machine learning-based classification methods outperform rule-based classification classifiers, they need to carefully extract classification features from text data, which play significant roles in text classifiers [19]. As a result, several feature selection approaches, such as latent Dirichlet allocation (LDA), term frequency-inverse document frequency (TF-IDF), χ 2 statistic, mutual information, etc., are widely applied to mine topics to reduce dimensions. For instance, Wang et al. introduced a latent Dirichlet allocation (LDA) with prior knowledge to extract features [10]. Brown et al. utilized LDA to discover accident characteristics and the contributors to rail accidents [11]. However, it is obvious that these techniques do not consider the mutual position of words in text sequences and the relevant results are misleading, especially in accident texts with high similarities.
Thirdly, in pace with the emergence of deep neural networks in NLP, such deep learning techniques have been gradually utilized in text classification tasks due to their excellent performances. Various studies on text classification models, such as CNN [20], recurrent neural network (RNN) [21], recurrent convolutional neural network (RCNN) [22], attention mechanism-based bidirectional LSTM (Att-Bi-LSTM) [23], have achieved superior results toward text categorization tasks in the areas of computers, medical treatments, and electric power. Wang et al. established a CNN-LSTM-based text emotion recognition model to shed light on the developing direction of the network’s public sentiment [24]. Dai et al. used multi-sieving CNN to efficiently cope with the unbalanced microaneurysm detection problem [25]. Guerrero et al. [26] proposed a customer filtering based on the classification–regression and self-organizing map to analyze the information of inspectors’ commentaries.

2.2. Text Representation Models

The purpose of text presentation is to embed the words in a text into low-dimensional vectors because existing machine learning methods cannot handle text data directly. The text representation techniques (as shown in Figure 2) can be classified into discrete representation and distributed presentation. One-hot vector is one typical solution of discrete representation. It encodes an index to each word within a text corpus, and then each word can be represented by a binary vector, of which the dimension is equal to the size of this vocabulary [27]. Similar to the one-hot vector, the bag of words representation means that the vector representation of the document can be directly obtained by summing the vectors of words. Meanwhile, n-gram encodes indexes to n adjacent words, considering the order of words. However, such a discrete counting vector representation suffers from the dimension curse and discards the semantic relationships between words.
In the past few years, word embedding representations have been focused on; they attempt to learn low-dimensions and continuous vector representations. Word embedding represents each word with a real-value vector and uses the similarity in the vector space to represent the semantic similarity of text words, inspired by the training idea of deep learning. SkipGram and CBOW are two common word embedding methods with lower computational complexity, using a shallow neural network to perform context-based prediction [28]. Moreover, the word2vec model can compute accurate, high-dimensional word vectors from huge data sets. For instance, Li et al. [8] introduced the continuous bag-of-word (CBOW) model into generating the word embeddings and used the LSTM-CNN model to extract traffic information from 3 billion microblogs. However, such an approach neglected the distances of words (i.e., global statistical information). Hence, GloVe [29] was proposed to use the word co-occurrence and local context to learn word vectors, combining the merits of the matrix factorization and prediction-based methods. However, contextualized information is discarded in the word embedding presentation obtained by Glove, leading to great challenges of polysemy and complex syntactic features. To this end, pre-trained language models, such as embeddings from language model (ELMo) [30], generative pre-training (GPT) [31], BERT [12], etc., were investigated and explored to obtain deep contextualized word representations, integrating word embedding learning into the neural language models. In particular, with the release of BERT, it outperforms the above-mentioned models in the majority of NLP tasks, such as word embedding learning, which can efficiently capture dependencies over longer distances and excavate the actual bidirectional contextual information, improving the language understanding ability of networks via large-scale unsupervised pre-training.
Therefore, in this paper, we adopt the pre-trained language model BERT to capture long-term dependencies between text words and learning contextualized word embedding representations based on the constructed catenary defect text dataset. Moreover, motivated by the idea of ResNet [32] and the deepening of word-level convolutional neural networks (CNNs) [33], we propose a deep CNN algorithm that can efficiently capture long-range associations in text, which can extract richer semantics for domain-specific defect information identification and achieve superior performance by deepening the network without increasing computational costs by much.

3. Methodology

As illustrated in Figure 3, the methodological framework of BERT-DTCN is mainly composed of three stages:
  • Catenary defect text database: After obtaining the relevant catenary defect texts accumulated in the data center of the China Railway Administration during its long-time operation and maintenance, it is intended to conduct textual data prepossessing and construct the text dataset.
  • Word embedding presentations: The BERT model projects the Chinese texts related to catenary defects into context-aware representations that can be handled and understood by machines.
  • Classification of texts to distinguish the defect level: The DTCN module is trained to categorize the catenary defect texts by utilizing equal-width convolution and multiple convolution-residual layers with the pooling layer with stride 2 for downsampling.

3.1. Problem Definition

The problem tackled in this paper is defined as follows. Considering a collection of N defect text records that are labeled using Y = 2 classes (i.e., binary problem) in a supervised manner, the catenary defect text database is denoted as S = S 1 , S 2 , , S i , S N and the labels are represented by Y = Y 1 , Y 2 . The supervised labeling process can be denoted as f : S Y , where f is to convert the input text sequence S to binary vectors Y. The collection of all labeled training defect texts is represented as Y = f ( S ) [34].
The procedure of our method can be denoted as f : S Y ^ , which means that each defect text sequence S i S creates a label Y i ^ = f ( S i ) , Y i ^ Y .

3.2. Catenary Defect Text Database

3.2.1. Data Source and Text Content

Based on the 6C system and periodical patrolling, we collected the catenary defect text records from the China Railway Administration of Chengdu, Lanzhou, and Hengshui. There were nearly 45,000 defect text records available for three years (1 January 2016–31 December 2018) in total.
Each record can be recognized as a piece of a catenary defect, which contains the detecting time, location information, defect level, defect description, and so on, as shown in Figure 4.

3.2.2. Characteristic Analysis

Compared with conventional Chinese text, catenary defect text is characterized by the following characteristics:
  • Diversity. The operation and maintenance texts for the catenary system contain the time, number, unit, defect component, and defect description.
  • Correlation. The operation and maintenance texts are closely linked to the railway transportation, which contains a large number of rail transit terminologies.
  • Uncertainty. A great deal of defect descriptions in the catenary texts might be incomplete, noisy, fuzzy, or random.
  • Polysemy. Several polysemous words in defect texts might have multiple meanings, which need to be distinguished under different semantic meanings.

3.2.3. Data Processing

Given that catenary defect texts are mainly manually recorded by professional workers, the time-consuming and labor-intensive work may make recorders prone to lose their enthusiasm in long-term responses to repetitive works, leading to low-quality catenary defect text records. Hence, we conducted textual data preprocessing by the extra manual intervention. Several data cleaning methods, including incomplete data resolution, error values, duplicate records, detection, and elimination, are applied to improve the quality for the following catenary defect text classification task.

3.3. Word Embedding

To obtain the word embedding representations related to the catenary defect, the pre-trained language model BERT is adopted to learn context-aware information in this part. The structure of the BERT model is depicted in Figure 5, which consists of the input layer, BERT encoder, and output layer. It projects the Chinese input text for the catenary defect into context-aware representation [12]. Meanwhile, the generated word embedding vectors of each sequence are no more than 512 tokens. As for an input catenary defect text sequence with n token words s = s 1 , s 2 , , s n , the contextualized word embedding representations in the output layer can be denoted as x = x 1 , x 2 , , x n , with x R n v .

3.3.1. Input Layer

Given a token catenary defect text sequence s containing n words, s is represented as s = s 1 , s 2 , , s i , , s n , where s i ( 1 i n ) means the i th word in a Chinese text [35]. As for the input of the BERT model, special (CLS) and (SEP) tokens are added at the beginning and end of sentences respectively. In particular, the (PAD) tokens are marked at the end of the sentences to ensure their lengths are the same as the maximum sequence lengths [12]. If a single sequence consists of two sentence pairs, the sentences will be separated by the special token ((SEP)) and labeled with sentence A or sentence B, whose feature values are 0 and 1, respectively. As for each token sequence, its input representation is obtained by summing the corresponding token, segment, and position embedding. An example of the visualization of this construction is shown in Figure 6.

3.3.2. BERT Encoder

The BERT used in this paper is composed of 12 transformer blocks, 768 hidden sizes, and 12 self-attention heads [36]. The basic structure of the transformer encoder is illustrated in Figure 7. The word-embedded representation of a single sequence is the token as the input of the encoder, and the positional encoding is added. The self-attention layer enables the encoder to capture the contextual information from the word when coding, which can calculate the weighted value of each word and all words. Then, the feature vector of each word is obtained, which contains the information of the whole sentence. Afterward, multiple feature vectors obtained by the multi-headed mechanism are spliced together, the dimensions of which are descended by a full connection layer. Finally, the contextualized word embedding vectors were output through the feedforward network, two residual connection layers, and a normalization layer.
The self-attention layer performs the linear transformation on the input vector to obtain a linear value, and then calculates the attention weight, as depicted in Figure 8. The calculation of self-attention consists of three steps:
(1)
Creating three vectors (i.e., a query vector, a key vector, and a value vector) from each of the encoder’s input vectors and obtaining a weighted score by calculating the dot products of the query with all Keys. It can be calculated as:
f ( Q , K i ) = Q K i T
(2)
Dividing the scores by scaling factor d k and then normalizing the scores through a softmax operation. It can be represented as:
a i = softmax ( f ( Q , K i ) d k ) = exp ( f ( Q , K i ) d k ) j exp ( f ( Q , K i ) d k )
(3)
Multiplying each value vector by the softmax scores and summing up the weighted value vectors. It can be defined as:
Attention ( Q , K , V ) = i a i V i
where Q, K, V refer to the query, key, and value matrix, and d k represents the scaling factor.
Thus, such word embedding presentations not only contain the meanings of the words themselves, but also capture the relationships among words. Compared with the traditional word embedding methods, BERT-based embedding can obtain the representation of text with rich semantic information.

3.4. Deep Text Categorization Network

The DTCN involves taking the obtained word embedding representations as the input, categorizing the catenary defect texts, and distinguishing the severity levels using defect text records. Motivated by the idea of ResNet and deep pyramid CNN [33], we propose a deep learning-based defect text categorization network called DTCN that adopts the structures of deep CNNs, which can achieve the superior performance by deepening the network without increasing computational costs by much. As discussed previously, its structure is depicted in Figure 3, which consists of four components, namely, the embedding layer, equal-width convolution layers, stacking of convolution blocks (equal-width convolution layers and a shortcut) interleaved with max-pooling layers with stride 2 for downsampling, and a fully connected layer.

3.4.1. Embedding Layer

We used a convolution layer to transform the learned word embedding into the feature maps, the dimensions of which are the number of filters × ( seq_length -2). It is essentially a feature extractor that encodes semantic features in a given dimension (related to the number of filters and lengths of text sequences), in which words with similar semantics also have closer Euclidean or cosine distances. The convolution operation is to move a filter over the text sequence matrix (input map) and compute the dot products. In DTCN, to learn more sophisticated features, two filters are used to convolve the input word embedding vectors, and all feature vectors are concatenated into a three-dimensional convolution feature map.
Let x i R n v represent the v-dimensional word vector with respect to the i th word in a sentence with n words. The input map of DTCN can be denoted as
x 1 : n = x 1 x 2 x n
where ⊕ refers to concatenation operator. x i : i + j represents the concatenation of words x i : i + j = x i , x i + 1 , , x i + j . The filter w R m v function involves computing a new feature in the window of m words. Thus, a feature c i is produced from a window of words x i : i + m 1 by
c i = f ( w · x i : i + m 1 + b )
where b R is a bias and f is an activation function.
The feature map is generated from each possible window of words in the sentence x 1 : m , x 2 : m + 1 , , x n m + 1 : n . It is represented as
c = c 1 , c 2 , , c n m + 1
where c R n m + 1 .

3.4.2. Downsampling with the Number of Feature Maps Fixed

The increasing number of feature maps cannot improve the accuracy, but only increase the computation time substantially;l thus, the DTCN adopts equal-width convolution to enrich the semantic representation by keeping the same number of channels. After equal-width convolution layers, the convolution block (equal-width convolution layers and max-pooling with size 3 and stride 2) (as shown in Figure 3) is performed with the number of channels (also called filters) that are fixed. As a result, the length of the text sequence vectors is halved, and then the computation time of each convolutional layer is reduced by half. Moreover, the number of convolution blocks is automatically determined by the length of the defect text sequence. Therefore, the total computation time is bounded by the computation time of a convolution block. In addition, downsampling with stride 2 can efficiently double the coverage of the convolution kernel, which is computationally efficient in representing long-range associations and more global information.

3.4.3. Shortcut Connections with Pre-Activation

Due to the saturated accuracy and rapid degradation with the network depth increasing, there are great difficulties in training the deeper neural networks, such as higher training errors with more layers, and vanishing/exploding gradients [32]. To better train the deep networks, the DTCN model uses additive shortcut connections with identity mappings in [37]: z + f ( z ) , where f represents skipped convolution layers with pre-activation. In particular, pre-activation means that activation is done before weighing. Thus, the convolution layers of the DTCN can be computed by W σ ( x ) + b , where x refers to a small region (overlapping with each other) of word embedding vectors at each location, σ ( · ) is a component-wise nonlinear activation, and weights W and biases b are needed to be trained. In DTCN, activation σ ( · ) is set to σ ( x ) = max ( x , 0 ) . In addition, linear weighting W σ ( x ) + b with pre-activation eases the training of deep networks [38,39].
In our training process, the training set includes two types of labels, i.e., “0” (severity level 1) and “1” (severity level 2). The BERT-DTCN is an overall framework that is trained automatically. The details of the training algorithm are shown in Algorithm 1. The BERT-DTCN is trained by minimizing the cross-entropy loss function based on the defect text dataset. The loss function of the classifier for the BERT-DTCN is denoted as follows:
L o s s c = 1 n y i log y i ^ + ( 1 y i ) log ( 1 y i ^ )
where y i means the actual label of the input defect text sequence i; y i ^ is the probability vector corresponding to the output of the BERT-DTCN of the input defect text sequence i; and n is the number of training samples,
Algorithm 1 Pseudocode for training the BERT-DTCN.
Require: 
X s : training set for the BERT-DTCN, including the constructed defect text dataset and labels; N c : number of classifier-training iterations per mini-batch.
1:
for the number of training iterations do
2:
    Sample mini-batch of m examples from the training set X s ;
3:
    for  i = 1 N c  do
4:
        Update the BERT-DTCN by minimizing the loss: L o s s c
5:
    end for
6:
end for

4. Experiment Results and Analysis

To facilitate the performance evaluation of the BERT-DTCN, we investigated some training protocols and comparisons with competing text classification approaches on the catenary defect text dataset in this section.
We obtained nearly 45,000 catenary defect text records from 2016 to 2018 in the China Railway Administration of Chengdu, Lanzhou, and Hengshui. After data cleaning and filtering, we labeled 11,106 catenary defect text records that were applied in the catenary defect severity identification, and the proportions of the training, verification, and test defect text dataset were set to 0.7, 0.15, and 0.15, respectively. Table 1 lists the details of the catenary defect text dataset that we used for the experiments.
We applied the BERT model to generate word embedding presentations for the subsequent defect information extraction and defect level discrimination. In our experiments, we found that a majority of defect text records were concise and brief, and then the maximum length of word sequences was set to 32 (i.e., the padding size). The zeros after each sequence were padded until the length reached 32. Thus, the number of convolution blocks could be determined and was set to 4. Afterward, the general Chinese Bert language model “bert-base-chinese” was used to learn word embeddings, and then the generated word vector was a 768-dimensional vector, which was equal to the number of hidden units. The obtained word embedding vectors of each sequence can be represented by a 32 × 768 matrix, which could be fed into our classification model.
In addition, all programs were implemented under the PyTorch framework. The main configuration of the computer was a 1080Ti graphics card, Intel Xeon E5 v3, with 32G of memory [3].

4.1. Training Protocol

In this section, some training protocols of the BERT-DTCN are investigated. We focus on two key parameters: the number of convolution layers in equal-width convolution layers (called N c l ) and the number of output channels (filters) (called N f ) in the convolution. In particular, because the equal-width convolution layers were applied in BERT-DTCN, the number of input channels was equal to that of the output channels.
N c l and N f : The N c l determines the depth of neural networks, which allows each lexeme to contain more and longer contextual information. The N f is closely associated with the number of feature maps, which restricts the dimensions of the semantic space and determines the size of the output probability map. We trained the BERT-DTCN with the hyperparameter settings listed in Table 2 under various N c l and N f , and the results are given in Table 3.
We found that N c l and N f contributed to some differences in the classification performance and training time. It demonstrates that a network with a deeper structure and large sizes of feature maps is not necessary to achieve better performance. The computation burden increases with the increasing number of N c l and N f ; however, the whole computation time is indistinguishable. This is because after the convolution block, max-pooling with the number of feature maps fixed is performed, and then the computation time for each convolution layer is halved. Thus, the total computation time is almost the same. According to the results, we set the N c l = 4 and the N f = 140 . Therefore, the detailed structure parameters of the BERT-DTCN are set as listed in Table 4.

4.2. Ablation Study

There are two critical modules that work cooperatively in the BERT-DTCN model, namely, BERT-based word embedding presentation and deepening of CNN-based text classification (DTCN). In order to validate the effectiveness of the BERT, we compare the performance of BERT-DTCN and DTCN in this section. The relevant parameter settings of the DTCN and BERT-DTCN are the same as listed in Table 2 and Table 4. We trained two networks with the same training protocols based on the constructed dataset.
The most intuitive evaluation index of the classification problem is used, i.e., accuracy rate, the percentage of correctly classified samples in the total number of samples. Except for the accuracy rate, three evaluation indexes related to text categorization were adopted, namely, precision (P), recall (R), and the F 1 -score ( F 1 ) [12,35]. In this part, we take the accuracy rate as the primary evaluation of the text classification model, and the F 1 , P, and R as the auxiliary indicators.
As illustrated in Table 5, regardless of which category the defect text is in, the BERT-DTCN achieves superior results in all of the elevation metrics (Acc, P, R, and F 1 ). Moreover, the overall accuracy of the catenary text classification reaches up to 97.40%. Compared with the DTCN, the macro P, R, F 1 , and accuracy of BERT-DTCN are all improved by 0.42%. At each level of the catenary defect text dataset, the BERT-DTCN model achieves improvements ranging from 0.36% and 0.53%. In addition, the training loss curves in Figure 9 demonstrate that the BERT-DTCN converges faster and achieves better performances with fewer steps over DTCN. The ROC curve displays the trade-off between the true positive rate or sensitivity (proportion of positive tuples that are recognized) and the false-positive rate (proportion of negative tuples that are incorrectly recognized as positive) for DTCN and BERT-DTCN. The ROC curve in Figure 10 shows that the BERT-DTCN has a larger area under the ROC curve than that of DTCN, with a better severity level classification performance in the catenary defect text. Moreover, the PR curve in Figure 11 shows that the BERT-DTCN has both high precision, and high recall, characterizing the superior effectiveness of classification performance.
To summarize, it is obvious that the BERT-based word embeddings have significant positive impacts on the classification performance of the DTCN. This is because the word embedding vectors obtained by BERT can accurately represent the semantic features of catenary defect vocabulary and capture contextual information, improving the performance of the catenary defect severity identification to a certain extent. Hence, we conclude that the BERT can learn contextual semantic information, which reduces the interference brought by the non-standard parts in the defect texts and contributes to beneficial effects on the text categorization performance.

4.3. Classification Performance Comparison

To better evaluate the classification performance of BERT-DTCN, we compare our method with state-of-the-art deep learning models for text classification based on the constructed catenary defect text database.
The following baselines are adopted to validate the effectiveness and efficiency of the BERT-DTCN model:
CNN: A method for sentence-level classification tasks based on CNN [20].
RNN: RNN for multitask learning [21].
RCNN: A method of extracting contextual information for text classification based on RCNN [22].
FastText: A fast training model with a large corpus [40].
Att-Bi-LSTM: A word-level text categorization model based on Att-Bi-LSTM [23].
Transformer: A model architecture that enables global dependencies captured based on an attention mechanism [36].
We trained these networks with the same dataset and applied accuracy, P, R, and F 1 to evaluate the performances of these competing approaches.
As shown in Table 6, these comparative results illustrate that the BERT-DTCN achieves better performances in all elevation metrics. On the constructed catenary defect text database, the macro P, R, F 1 , and accuracy of BERT-DTCN reach up to 97.40%. This is because the BERT-DTCN can capture richer features through specific downsampling and learned word embedding, enhancing the representation of defect texts. Our model called BERT-DTCN outperforms competing models that are widely used in the text classification task. Moreover, the CNN, RCNN, and Att-Bi-LSTM also achieve superior performance in the catenary defect text classification in terms of macros P, R, F 1 , and accuracy based on the constructed defect text dataset.
Compared with the competing methods, the training loss curves in Figure 12 demonstrate that the BERT-DTCN converges faster and achieves better performances with fewer steps. As depicted in Figure 13 and Figure 14, the ROC and PR curves show that the BERT-DTCN can enhance performance over several state-of-the-art models. In general, the BERT-DTCN model achieves superior performance in the catenary defect level classification, which validates the effectiveness of the DTCN and BERT-DTCN. It is obvious that the DTCN and BERT-DTCN with the strategy of deepening the network can capture the global information in the text, achieving a fairly superior performance.
In addition, the BERT-DTCN used word embedding presentations pertaining to BERT to initialize word embeddings in the deep text categorization network and then feed it as training proceeded (distinguishing the defect severity level). The DTCN module in BERT-DTCN can be regarded as a deep extension of shallow CNN, sharing region embedding enhancement with diverse unsupervised embeddings. Based on the experimental results of the DTCN and CNN in Table 5 and Table 6, the DTCN with deepened networks can perform improvements in capturing more global information over CNN, which has proven the conclusions in [33], i.e., the added depth is indeed useful.

5. Conclusions

In this paper, we reported on the emerging text mining based on catenary defect records collected in the operation and maintenance of the catenary. We investigated the deep semantic learning method to automatically identify the severity level of the catenary defect. Firstly, we analyzed and summarized the characteristics of the catenary text, including diversity, correlation, uncertainty, and polysemy, and established the text dataset for the deep semantic learning-based defect text classification model. Different from counting vector representations obtained by discrete representative approaches, we applied the pre-trained language model BERT to learn contextual word embedding vectors. At the training phase, BERT-DTCN was trained by the cross-entropy loss to extract relevant defect information. Then, the classifier can learn how to distinguish between the severity level 1 defect and severity level 2 defect, i.e., the complex defect information extraction problem was transformed into a simple classification problem. We thoroughly analyzed the impacts of the training protocol and word embedding presentations obtained by BERT and compared the BERT-DTCN with other competing methods. The ablation experiments showed that the word embedding vectors obtained by BERT contributed to positive effects on the superior performance of the developed DTCN, which demonstrated the effectiveness of our classification model BERT-DTCN on the constructed catenary defect text dataset. Comparative experiments showed that BERT-DTCN outperformed the competing deep learning methods, which can effectively represent long-range associations in the catenary defect texts and extract global semantic information with deepened networks. Moreover, we found that the strategy of deepening the network can improve the classification performance to a certain extent. The number of filters and the depth of the network were two significant parameters for the deep networks. The proposed model can be applied in the operation and maintenance of a catenary system to extract defect information and categorize defect severity.
As for catenary defect record-based text mining, various issues and challenges need to be further explored, which can contribute to fruitful and beneficial results for the safe operation and maintenance of a catenary system. The extended catenary maintenance corpora and terminology dictionaries might enhance the performance to a certain extent. In addition, imbalanced data are other unsolved problems in this domain-specific task, which have adverse effects on the performances of existing categorization algorithms. In the future, these NLP tasks, such as imbalanced learning for catenary defect texts, BERT-based named entity recognition in Chinese catenary defect texts, and knowledge graph construction for the health management of a catenary system [6], deserve to be investigated in depth.

Author Contributions

Conceptualization, J.W. and S.G.; methodology, J.W. and S.G.; investigation, J.W. and L.Y.; data curation, L.Y.; writing—original draft preparation, J.W.; writing—review and editing, S.G., D.Z. and L.K.; supervision, L.Y., D.Z. and L.K.; funding acquisition, L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of Sichuan Province 2022NSFSC0572, in part by the National Natural Science Foundation of China under grant 52177115, and in part by the National Key R&D Program of China 2021YFB2601500.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, J.; Gao, S.; Yu, L.; Zhang, D.; Ding, C.; Chen, K.; Kou, L. Predicting Wind-Caused Floater Intrusion Risk for Overhead Contact Lines Based on Bayesian Neural Network with Spatiotemporal Correlation Analysis. Reliab. Eng. Syst. Saf. 2022, 225, 108603. [Google Scholar] [CrossRef]
  2. Wang, J.; Gao, S.; Yu, L.; Zhang, D.; Xie, C.; Chen, K.; Kou, L. Data-Driven Lightning-Related Failure Risk Prediction of Overhead Contact Lines Based on Bayesian Network with Spatiotemporal Fragility Model. Reliab. Eng. Syst. Saf. 2023, 231, 109016. [Google Scholar] [CrossRef]
  3. Gao, S.; Wang, J.; Yu, L.; Zhang, D.; Zhan, R.; Kou, L.; Chen, K. A Multilayer Bayesian Network Approach-Based Predictive Probabilistic Risk Assessment for Overhead Contact Lines Under External Weather Conditions. IEEE Trans. Transp. Electrif. 2022, 1–18. [Google Scholar] [CrossRef]
  4. Gao, S. Automatic Detection and Monitoring System of Pantograph-Catenary in China’s High-Speed Railways. IEEE Trans. Instrum. Meas. 2021, 70, 3502012. [Google Scholar] [CrossRef]
  5. Wang, H.; Liu, Z.; Xu, Y.; Wei, X.; Wang, L. Short Text Mining Framework with Specific Design for Operation and Maintenance of Power Equipment. CSEE J. Power Energy Syst. 2021, 7, 1267–1277. [Google Scholar] [CrossRef]
  6. Wang, J.; Wang, X.; Ma, C.; Kou, L. A Survey on the Development Status and Application Prospects of Knowledge Graph in Smart Grids. IET Gener. Transm. Distrib. 2021, 15, 383–407. [Google Scholar] [CrossRef]
  7. Kou, L.; Liu, C.; Cai, G.W.; Zhou, J.N.; Yuan, Q. Data-Driven Design of Fault Diagnosis for Three-Phase PWM Rectifier Using Random Forests Technique with Transient Synthetic Features. IET Power Electron. 2020, 13, 3571–3579. [Google Scholar] [CrossRef]
  8. Chen, Y.; Lv, Y.; Member, S.; Wang, X.; Li, L.; Member, S.; Wang, F. Detecting Traffic Information from Social Media Texts with Deep Learning Approaches. IEEE Trans. Intell. Transp. Syst. 2018, 20, 3049–3058. [Google Scholar] [CrossRef] [Green Version]
  9. Wang, Y.; He, Z.; Hu, J. Traffic Information Mining From Social Media Based on the MC-LSTM-CONV Model. IEEE Trans. Intell. Transp. Syst. 2020, 23, 1132–1144. [Google Scholar] [CrossRef]
  10. Wang, F.; Xu, T.; Tang, T.; Zhou, M.; Wang, H. Bilevel Feature Extraction-Based Text Mining for Fault Diagnosis of Railway Systems. IEEE Trans. Intell. Transp. Syst. 2017, 18, 49–58. [Google Scholar] [CrossRef]
  11. Brown, D.E. Text Mining the Contributors to Rail Accidents. IEEE Trans. Intell. Transp. Syst. 2016, 17, 346–355. [Google Scholar] [CrossRef]
  12. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  13. Borko, H.; Bernick, M. Automatic Document Classification. J. ACM 1963, 10, 163–222. [Google Scholar] [CrossRef]
  14. Sasaki, M.; Kita, K. Rule-Based Text Categorization Using Hierarchical Categories. IEEE Int. Conf. Syst. Man Cybern. 1998, 3, 2827–2830. [Google Scholar] [CrossRef]
  15. Aggarwal, C.C.; Zhai, C.X. A Survey of Text Classification Algorithms; Springer: Berlin/Heidelberg, Germany, 2012; pp. 163–222. [Google Scholar]
  16. Myaeng, S.H.; Han, K.S.; Rim, H.C. Some Effective Techniques for Naive Bayes Text Classification. IEEE Trans. Knowl. Data Eng. 2006, 18, 1457–1466. [Google Scholar] [CrossRef]
  17. Jati, W.K.; Kemas Muslim, L. Optimization of Decision Tree Algorithm in Text Classification of Job Applicants Using Particle Swarm Optimization. In Proceedings of the 2020 3rd International Conference on Information and Communications Technology (ICOIACT), Yogyakarta, Indonesia, 24–25 November 2020; pp. 201–205. [Google Scholar] [CrossRef]
  18. Joachims, T. Text Categorization with Suport Vector Machines: Learning with Many Relevant Features. In Proceedings of the 10th European Conference Machine Learning 1998, Chemnitz, Germany, 21–23 April 1998; pp. 137–142. [Google Scholar]
  19. Wang, J.; Li, Y.; Shan, J.; Bao, J.; Zong, C.; Zhao, L. Large-Scale Text Classification Using Scope-Based Convolutional Neural Network: A Deep Learning Approach. IEEE Access 2019, 7, 171548–171558. [Google Scholar] [CrossRef]
  20. Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the EMNLP 2014—2014 Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar]
  21. Liu, P.; Qiu, X.; Huang, X. Recurrent Neural Network for Text Classification with Multi-Task Learning. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New Nork, NY, USA, 9–15 July 2016; pp. 2873–2879. [Google Scholar]
  22. Lai, S.; Xu, L.; Liu, K.; Zhao, J. Recurrent convolutional neural networks for text classification. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; pp. 2267–2273. [Google Scholar]
  23. Zhou, P.; Shi, W.; Tian, J.; Qi, Z.; Xu, B. Attention-Based Bidirectional Long Short-Term Memory Networks for Relation Classification. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; pp. 207–212. [Google Scholar]
  24. Wang, X.; Kou, L.; Sugumaran, V.; Luo, X.; Zhang, H. Emotion Correlation Mining Through Deep Learning Models on Natural Language Text. IEEE Trans. Cybern. 2020, 51, 4400–4413. [Google Scholar] [CrossRef]
  25. Dai, L.; Fang, R.; Li, H.; Hou, X.; Sheng, B.; Wu, Q.; Jia, W. Clinical Report Guided Retinal Microaneurysm Detection with Multi-Sieving Deep Learning. IEEE Trans. Med. Imaging 2018, 37, 1149–1161. [Google Scholar] [CrossRef]
  26. Guerrero, J.I.; Monedero, I.; Biscarri, F.; Biscarri, J.; Millán, R.; León, C. Non-Technical Losses Reduction by Improving the Inspections Accuracy in a Power Utility. IEEE Trans. Power Syst. 2018, 33, 1209–1218. [Google Scholar] [CrossRef]
  27. Jiao, Q.; Zhang, S. A Brief Survey of Word Embedding and Its Recent Development. In Proceedings of the IAEAC 2021—IEEE 5th Advanced Information Technology, Electronic and Automation Control Conference, Chongqing, China, 12–14 March 2021; pp. 1697–1701. [Google Scholar] [CrossRef]
  28. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. In Proceedings of the 1st International Conference on Learning Representations ICLR, Scottsdale, AZ, USA, 2–4 May 2013; pp. 1–12. [Google Scholar]
  29. Pennington, J.; Socher, R.; Manning, C. GloVe: Global Vectors for Word Representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing EMNLP, Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  30. Peters, M.E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; Zettlemoyer, L. Deep Contextualized Word Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics, New Orleans, LA, USA, 1–6 June 2018; pp. 2227–2237. [Google Scholar]
  31. Radford, A.; Narasimhan, K. Improving Language Understanding by Generative Pre-Training. 2018. Available online: https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf (accessed on 13 December 2022).
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  33. Johnson, R.; Zhang, T. Deep Pyramid Convolutional Neural Networks for Text Categorization. In Proceedings of the ACL 2017—55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; Volume 1, pp. 562–570. [Google Scholar] [CrossRef] [Green Version]
  34. Li, Y.; Algarni, A.; Albathan, M.; Shen, Y.; Bijaksana, M.A. Relevance Feature Discovery for Text Mining. IEEE Trans. Knowl. Data Eng. 2015, 27, 1656–1669. [Google Scholar] [CrossRef]
  35. Yu, S.; Su, J.; Luo, D. Improving BERT-Based Text Classification with Auxiliary Sentence and Domain Knowledge. IEEE Access 2019, 7, 176600–176612. [Google Scholar] [CrossRef]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5999–6009. [Google Scholar]
  37. Rogers, D.; Preece, A.; Innes, M.; Spasić, I. Real-Time Text Classification of User-Generated Content on Social Media: Systematic Review. IEEE Trans. Comput. Soc. Syst. 2022, 9, 1154–1166. [Google Scholar] [CrossRef]
  38. Isa, D.; Lee, L.H.; Kallimani, V.P.; Rajkumar, R. Text Document Preprocessing with the Bayes Formula for Classification Using the Support Vector Machine. IEEE Trans. Knowl. Data Eng. 2008, 20, 1264–1272. [Google Scholar] [CrossRef]
  39. Kolchinsky, A.; Abi-Haidar, A.; Kaur, J.; Hamed, A.A.; Rocha, L.M. Classification of Protein-Protein Interaction Full-Text Documents Using Text and Citation Network Features. IEEE/ACM Trans. Comput. Biol. Bioinform. 2010, 7, 400–411. [Google Scholar] [CrossRef] [PubMed]
  40. Joulin, A.; Grave, E.; Bojanowski, P.; Mikolov, T. Bag of Tricks for Efficient Text Classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, 3–7 April 2017; pp. 427–431. [Google Scholar]
Figure 1. Scene of a catenary system.
Figure 1. Scene of a catenary system.
Sensors 22 09922 g001
Figure 2. Development process of text classification.
Figure 2. Development process of text classification.
Sensors 22 09922 g002
Figure 3. Architecture of defect severity identification for a catenary system.
Figure 3. Architecture of defect severity identification for a catenary system.
Sensors 22 09922 g003
Figure 4. The content of the catenary defect record.
Figure 4. The content of the catenary defect record.
Sensors 22 09922 g004
Figure 5. Structure of the BERT model.
Figure 5. Structure of the BERT model.
Sensors 22 09922 g005
Figure 6. An example of the BERT input representation.
Figure 6. An example of the BERT input representation.
Sensors 22 09922 g006
Figure 7. Basic structure of the transformer encoder.
Figure 7. Basic structure of the transformer encoder.
Sensors 22 09922 g007
Figure 8. The flowchart of the self-attention layer.
Figure 8. The flowchart of the self-attention layer.
Sensors 22 09922 g008
Figure 9. The training loss of DTCN and BERT-DTCN.
Figure 9. The training loss of DTCN and BERT-DTCN.
Sensors 22 09922 g009
Figure 10. The ROC curves for DTCN and BERT-DTCN.
Figure 10. The ROC curves for DTCN and BERT-DTCN.
Sensors 22 09922 g010
Figure 11. The PR curves for DTCN and BERT-DTCN.
Figure 11. The PR curves for DTCN and BERT-DTCN.
Sensors 22 09922 g011
Figure 12. The training loss of BERT-DTCN and competing models.
Figure 12. The training loss of BERT-DTCN and competing models.
Sensors 22 09922 g012
Figure 13. The ROC curves for BERT-DTCN and competing models.
Figure 13. The ROC curves for BERT-DTCN and competing models.
Sensors 22 09922 g013
Figure 14. The PR curves for BERT-DTCN and competing models.
Figure 14. The PR curves for BERT-DTCN and competing models.
Sensors 22 09922 g014
Table 1. The details of the catenary defect text classification dataset.
Table 1. The details of the catenary defect text classification dataset.
Dataset
Name
Classes of
Defect Level
Training
Set
Verification
Set
Test
Set
Catenary
Defect Text
2761116521653
Table 2. The Hyperparameter settings of BERT-DTCN.
Table 2. The Hyperparameter settings of BERT-DTCN.
HyperparameterSettingHyperparameterSetting
Learning rate0.00005Padding size32
OptimizerAdamEmbedding768
Batch size128Epoch20
Table 3. Accuracy and training time under different N c l and N f .
Table 3. Accuracy and training time under different N c l and N f .
N cl N f Acc/%Training Time
Two equal-width
convolution layers
12096.7317 min 25 s
13096.4917 min 15 s
14096.4917 min 12 s
15096.6717 min 19 s
16097.1017 min 19 s
Three equal-width
convolution layers
12096.6117 min 13 s
13097.0417 min 46 s
14097.3417 min 26 s
15097.2217 min 54 s
16097.2218 min 31 s
Four equal-width
convolution layers
12096.6718 min 01 s
13097.2217 min 29 s
14097.4017 min 29 s
15097.1617 min 16 s
16096.4317 min 18 s
bold is with best classification performance.
Table 4. The detailed structure parameters of the BERT-DTCN.
Table 4. The detailed structure parameters of the BERT-DTCN.
LayerInput SizeKernel SizeStrideOutput SizeNumber
InputBERT 32 × 768 × 1 1
Embedding layerConv2 32 × 768 × 1 3 × 768 1 30 × 1 × 140 1
Equal-width
convolution layers
Padding1_1 30 × 1 × 140 32 × 1 × 140 1
Conv2_1 32 × 1 × 140 3 × 1 1 30 × 1 × 140
Padding1_2 30 × 1 × 140 32 × 1 × 140
Conv2_2 32 × 1 × 140 3 × 1 1 30 × 1 × 140
Padding1_3 30 × 1 × 140 32 × 1 × 140
Conv2_3 32 × 1 × 140 3 × 1 1 30 × 1 × 140
Padding1_4 30 × 1 × 140 32 × 1 × 140
Conv2_4 32 × 1 × 140 3 × 1 1 30 × 1 × 140
Convolution blockPadding2 30 × 1 × 140 31 × 1 × 140 4
Max-pooling 31 × 1 × 140 3 × 1 2 15 × 1 × 140
Padding1_1 15 × 1 × 140 17 × 1 × 140
Conv2_1 17 × 1 × 140 3 × 1 1 15 × 1 × 140
Padding1_2 15 × 1 × 140 17 × 1 × 140
Conv2_2 17 × 1 × 140 3 × 1 1 15 × 1 × 140
Padding1_3 15 × 1 × 140 17 × 1 × 140
Conv2_3 17 × 1 × 140 3 × 1 1 15 × 1 × 140
Padding1_4 15 × 1 × 140 17 × 1 × 140
Conv2_4 17 × 1 × 140 3 × 1 1 15 × 1 × 140
OutputFully connected
layer
14021
Table 5. Comparative results between BERT-DTCN with DTCN.
Table 5. Comparative results between BERT-DTCN with DTCN.
Model  Severity Level 1 (827)Severity Level 2 (826)Macro AverageAcc/%  
P/%R/% F 1 /%P/%R/% F 1 /%P/%R/% F 1 /%
DTCN96.7597.2296.9897.2096.7396.9796.9896.9896.9896.98
BERT-DTCN97.2397.5897.4097.5797.2297.3997.4097.4097.4097.40
bold is with best classification performance.
Table 6. Comparison of BERT-DTCN with competing approaches on the constructed catenary defect text database.
Table 6. Comparison of BERT-DTCN with competing approaches on the constructed catenary defect text database.
Model  Severity Level 1 (827)Severity Level 2 (826)Macro AverageAcc/%  
P/%R/% F 1 /%P/%R/% F 1 /%P/%R/% F 1 /%
CNN96.0196.0196.0196.0096.0096.0096.0196.0196.0196.01
RNN94.9492.2993.9593.1295.0494.0794.0394.0194.0194.01
RCNN96.7095.5396.1195.5796.7396.1596.1396.1396.1396.13
Atti-Bi-LSTM95.8194.0794.9494.1795.8895.0294.9994.9894.9894.98
FastText93.4192.5092.9592.5793.4693.0192.9992.9892.9892.98
Transformer94.0491.6692.8491.8594.1993.0192.9592.9292.9292.92
BERT-DTCN97.2397.5897.4097.5797.2297.3997.4097.4097.4097.40
bold is with best classification performance.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, J.; Gao, S.; Yu, L.; Zhang, D.; Kou, L. Defect Severity Identification for a Catenary System Based on Deep Semantic Learning. Sensors 2022, 22, 9922. https://doi.org/10.3390/s22249922

AMA Style

Wang J, Gao S, Yu L, Zhang D, Kou L. Defect Severity Identification for a Catenary System Based on Deep Semantic Learning. Sensors. 2022; 22(24):9922. https://doi.org/10.3390/s22249922

Chicago/Turabian Style

Wang, Jian, Shibin Gao, Long Yu, Dongkai Zhang, and Lei Kou. 2022. "Defect Severity Identification for a Catenary System Based on Deep Semantic Learning" Sensors 22, no. 24: 9922. https://doi.org/10.3390/s22249922

APA Style

Wang, J., Gao, S., Yu, L., Zhang, D., & Kou, L. (2022). Defect Severity Identification for a Catenary System Based on Deep Semantic Learning. Sensors, 22(24), 9922. https://doi.org/10.3390/s22249922

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop