Next Article in Journal
Usage of Underground Space for 3D Cadastre Purposes and Related Problems in Turkey
Previous Article in Journal
Selective D3 Receptor Antagonist SB-277011-A Potentiates the Effect of Cocaine on Extracellular Dopamine in the Nucleus Accumbens: a Dual Core-Shell Voltammetry Study in Anesthetized Rats
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Vehicle Classification Using Embedded Strain Gauge Sensors

1
Department of Automation of Testing and Control, Harbin Institute of Technology, Harbin, China
2
MEMS Center, Harbin Institute of Technology, Harbin, China
*
Author to whom correspondence should be addressed.
Sensors 2008, 8(11), 6952-6971; https://doi.org/10.3390/s8116952
Submission received: 21 July 2008 / Revised: 19 September 2008 / Accepted: 30 October 2008 / Published: 5 November 2008
(This article belongs to the Section Chemical Sensors)

Abstract

:
This paper presents a new vehicle classification and develops a traffic monitoring detector to provide reliable vehicle classification to aid traffic management systems. The basic principle of this approach is based on measuring the dynamic strain caused by vehicles across pavement to obtain the corresponding vehicle parameters – wheelbase and number of axles – to then accurately classify the vehicle. A system prototype with five embedded strain sensors was developed to validate the accuracy and effectiveness of the classification method. According to the special arrangement of the sensors and the different time a vehicle arrived at the sensors one can estimate the vehicle's speed accurately, corresponding to the estimated vehicle wheelbase and number of axles. Because of measurement errors and vehicle characteristics, there is a lot of overlap between vehicle wheelbase patterns. Therefore, directly setting up a fixed threshold for vehicle classification often leads to low-accuracy results. Using the machine learning pattern recognition method to deal with this problem is believed as one of the most effective tools. In this study, support vector machines (SVMs) were used to integrate the classification features extracted from the strain sensors to automatically classify vehicles into five types, ranging from small vehicles to combination trucks, along the lines of the Federal Highway Administration vehicle classification guide. Test bench and field experiments will be introduced in this paper. Two support vector machines classification algorithms (one-against-all, one-against-one) are used to classify single sensor data and multiple sensor combination data. Comparison of the two classification method results shows that the classification accuracy is very close using single data or multiple data. Our results indicate that using multiclass SVM-based fusion multiple sensor data significantly improves the results of a single sensor data, which is trained on the whole multisensor data set.

Graphical Abstract

1. Introduction

Traffic data collected on highways have many applications, depending on the various agencies' needs. Traffic parameters which are the subject of direct measurement include the vehicle velocity, number of vehicles moving in the same direction, time distance between the vehicles, length of traffic jams, time of access to the traffic, etc. Very useful parameters are also the types (i.e. motor car, delivery van, lorry, trailer, etc.) and the number of moving vehicles belonging to the specified type. The type of vehicle is very important for the access control (areas closed to some vehicle types, limits of speed and weight for different vehicle types), statistical purposes and weighing of vehicles in motion.
Recently, a variety of methods have been proposed to identify vehicle class, weight and speed. For instance, vehicle parameters are derived from traffic-induced signals of Inductive Loop Detectors (ILDs) [1-4]. These detectors are imbedded in the roadway surface where heavy traffic and construction can cause damage to detectors. They have high failure rates due to poor maintenance [5]. High maintenance costs have caused agencies to seek replacement technologies. Based on recent advances in computer vision, image analysis techniques have also been applied to develop vehicle classification systems [6-10]. Under adverse real-world conditions, techniques based on image analysis are often affected by weather conditions such as rain and fog, low resolution/contrast and this often leads to low classification accuracy [11]. There are a number of other sensing technologies relevant to vehicle classification. These technologies include infrared, ultrasonic, radar, microwave, and video detectors. They each have their own strengths and weaknesses.
Pattern recognition is concerned with how to divide the classification of objects into categories. For the last 30 years, pattern recognition has been used with increasing success in a number of areas such as medicine, weather forecasting, automated industrial inspection and transportation. Vehicle classification is a type of pattern recognition, so many effective pattern recognition methods and combinations of multiple sensors have been used for classification and identification of vehicles in a particular vehicle classification system which have achieved satisfactory results, such as artificial neural networks [1, 4, 6, 11], Bayesian network data fusion [12], fuzzy data fusion [13], etc.
In this paper a novel pavement strain-based vehicle classification approach is developed, which may be used to provide pavement structural monitoring information. Using this approach, vehicles passing over a pavement deck can be classified purely from strain-response readings taken from the structure over which truck is traveling. A vehicle classification is usually determined according to particular countries' criteria with regards to a specific vehicle design feature: number of axles, distance between axles, etc. A vehicle's category parameters can be extracted from multiple strain response curves when a vehicle crosses the instrument-containing pavement. The accuracy of any classification scheme depends on the accuracy of the equipment used to capture the values of the discrimination variables and the accuracy of the corresponding classification algorithm(s). In order to be obtain vehicle classification parameters, a number of sensors will be used and the classification parameters of each are repeatedly measured by multiple sensors placed at different measurement locations.
In China, the types of vehicle vary from state to another, depending on the prevailing economic and social activities. At a state level, different activities also use of different vehicles in different areas, which leads to a lack of a uniform standard for the feature parameters of vehicles (axle number, axle distance), so establishing some fixed thresholds for classifying vehicles is very difficult, and at the same time the classification accuracy is not high. On the other hand, these patterns from different types of vehicles seemed to have a lot of overlap between them; this necessitates the use of pattern recognition and classification techniques to distinguish between vehicle groups. Good separation is that which results in minimum classification errors. In our research, a support vector machines (SVMs) machine learning method was employed to process feature vectors extracted from multiple strain time histories to obtain the vehicles' classification information. The method is still new and believed to be stronger in classification problems than neural networks, especially in their principles of problem generalization. The SVM uses structural risk minimization (SRM) that minimizes the upper bound on the expected risk and is said to be superior to neural network's empirical risk minimization (ERM) [14,15].
The main aim of this research was to investigate the feasibility of developing a novel sensor system based on multiple embedded strain gauges installed in the pavement to classify moving vehicles. In the following sections, the pavement test bench based on multiple strain gauges and data collection equipment is introduced, followed by results from a preliminary analysis. Thereafter, based on pavement dynamic strain response time histories subjected to the moving load vehicle classification approach are presented in detail, including information on the analysis framework, key feature extraction method, one-against-one multi-class SVMs development, analysis results and discussions.

2. Instrumental Pavement Test Bench and Strain-Vehicle Database

2.1 Description of the testing bench

An instrumental pavement panel (measuring 4.6 m in length and 4.5 m in width) was installed in 2004 along the roadway on k220+300 of Tong-Three state road, in Jiamusi of Heilongjiang province, Northeast China. This panel were built of concrete materials and instrumented to verify durability and potential for long-term usage. During 2005-2006, periodic measurements acquired pavement strain data as the basis for development of our strain-based vehicle classification.
The testing bench includes five embedded concrete strain gauge sensors located symmetrically below the surface of the pavement slab (Figure 1), used to measure the rigid pavement strain response. In this figure W =4.50 m is the pavement slab width; L =4.6 m is the pavement slab length; v is the direction of traffic; sensors S1, S2, S4, S5 are 25 cm under the edge of pavement; s=4.1 m is the distance between S1 and S4. This layout can be used to measure all the vehicles travelling on the right side of the road. Through the data acquisition system (this PCI-data acquisition board supporting 12 bit 16 single-ended or 8 differential analog inputs maximum up to 1 MHz sampling rate, made by ADVANTECH CO. LTD PCI-1712 Multifunction DAQ, refer to Figure 2), live data from the strain gauges and the artificial observations are synchronized and stored in a temporary buffer. The acquisition program, written in C++, provided for different sampling rates and times. A sampling rate of 2 kHz for 2.5 sec for each channel was found to be most appropriate for each vehicle.
As all the strain gauges are located on a westbound lane (Figure 1), this study only focuses on records of vehicles traveling in this direction. Furthermore, for this relative simple structural system, the strains recorded by the different gauges were found to be highly correlated. In view of this fact, only the strain data of sensors 1, 3, 4 were utilized in this study.

2.1 Pavement strain measurement

Strain sensors measure the expansion and contraction of pavement material due to mechanical stress. Like all transducers, these sensors rely on indirect measurement for determining strains. The strain gauge is embedded into the pavement surface (refer to Figure 3). Then, the strain εSUR experienced by the pavement is transferred directly to the gauge, which responds with a linear change in electrical resistance.

2.2 Embedded Strain Gauge Sensor

Two common gauge sensor types are the electrical resistance strain gauge and the fiber optic strain gauge. For selecting a electrical resistance strain gauge the main considerations are that they are relatively inexpensive, can achieve overall accuracy of better than ±0.1%, they are available in a short gauge length, are only moderately affected by temperature changes, have small physical size and low mass, and are highly sensitive. Resistance strain gauges can be used to measure both static and dynamic strains. Figures 4 and 5 show a structural sketch of the embedded concrete dynamic strain sensor and photos, respectively. The embedment strain gauges are designed for direct embedding in concrete. The dimensions of the strip strain sensors are 12.5 cm length, 1.2 cm width and 1 cm thickness. They are uniaxial embeddable strain gauges, with self-temperature-compensation, and a resistance of 350±0.5 Ω. The gauge factor is 2.0 and the modulus of elasticity is 30,000 Mpa. Its standard range is -1500 με (compressive strain) and +400 με (tensile strain). The frequency scope could be extending from 0 Hz to 3 kHz which is used for measuring dynamic strain response in the concrete pavement subjected to moving vehicular loads. It is extra rugged to resist bending, and with large flanges to provide a greater engagement area (Figure 5).
Since the embedded concrete strain gauge is tightly bonded with epoxy to the rigid pavement, the strain will be transferred to strain gauge. The piezoresistive material of the strain gauge will produce a relative resistance change. A Wheatstone bridge is usually employed to transfer the relative resistance change of piezoresistive gauges into voltage output. The dynamic strain meter will provide incentive power for Measurement Bridge and amplify the output voltage Vout. When Vout is measured, the quantity of strain may be calculated from Equation (1):
ε SUR = V out K k 0
where εSUR is the longitudinal strain of the surface rigid pavement, Vout is the output of the dynamic strain meter, K is the sensitivity of the strain meter (here K=200 με/1.2V), k0 is the amplifier factor of the strain gauge (here k0 =2).

2.3 Characteristics of Traffic-Induced Pavement Strain Response

Herein, the nature and characteristics of the recorded traffic-induced strain time histories are first investigated. Figure 6 shows a representative record corresponding to a 4-axle semi-trailer truck (1F+2M+1R, Table 1). When a vehicle axle crosses the pavement slab, the strain response increases significantly. In Figure 6, four major peaks can be seen in the time history, each corresponding to an axle. In addition, the space of these peaks is consistent with the distance between the axles. For instance, the 2nd peak is far from the 3rd peak as these axles are widely spaced, and the magnitude of each strain peak primarily depends on the axle weight. In the linear range of rigid pavement response, the stress vehicle-induced is proportional to the strain. i.e. εSUR = σ/E, where εSUR is the longitudinal strain induced by the vehicle wheel load, σ is the stress and E is elastic constant called the Young's modulus. In our research, five embedded concrete strain gauges are employed to obtain traffic-induced dynamic pavement strain response time histories.
As summarized in Sun et al., different categorization schemes have been adopted in recent vehicle classification studies [28]. According to a Federal Highway Administration (FHWA) standard, vehicles are categorized into the six classes: passenger cars, motorcycles, buses, other 2-axle 4-tire vehicles, single-unit 2-axle 6 tire or more trucks, and combination trucks. In our research, a similar categorization scheme was defined to sort vehicles into the five classes (Table 1): small vehicles, medium trucks, buses/large trucks, 3-axle trucks, and combination trucks. The small vehicle class includes motorcycles if the corresponding record strain signal exceeds the ambient noise level.

3. Preliminary Experiments

3.1 Description of the experiments

The experimental plan for field testing covered a wide range of pavement responses as a function of magnitude of load (empty, intermediate and fully loaded) and speeds (5, 20, 30, 40 and 60 km/h). A two axle truck of known weight was used in the study. The accuracy of the corresponding vehicle classification parameters will directly affected the accuracy of the classification. For the purposes of examining capture the accuracy of vehicle parameters, a two axle truck of known axle distance and vehicle length was used in the research. In order to examine measurement repeatability, every experiment with the same velocity and the location across the pavement was replicated 10 times.

3.2 The calculation of vehicle parameters

To get other vehicle parameters, an accurate speed should be measured first. Figure 7 shows the real response curves of sensors S1, S3, S4 when a 2-axle truck crosses them. A Matlab program is applied to each strain time history to automatically identify the basic parameters: N (Number of peak each curve) corresponding to number of axle, v (speed of vehicle passing over sensor), WB (distance between adjacent axle). Since the distance between sensors S1 and S3 or S3 and S4 is known, and this distance is far greater than the tire's contact length, for a 2-axle vehicle, the speed can be calculated by the formula below:
v = ( v 1 + v 2 ) / 2 where v 1 = s / 2 t 31 1 ; v 2 = s / 2 t 43 1 .
where s/2 is the distance between S1 and S3 or S3 and S4 (see Figure 6); v1 is the velocity measured by S1 and S3; v2 is the velocity measured by S3 and S4; t 31 1 and t 43 1 is the time interval between the first peak of sensor S1 and S3, S3 and S4, respectively (see Figure 7.). To obtain a more accurate speed, we take twice the average speed of measurements [refer to Equation (2)].
With the measured speed, the wheelbase (axle distance) of the vehicle can be measured using the following equation:
W B = ( W B 1 + W B 2 + W B 3 ) / 3 where W B 1 = v τ 11 ; W B 2 = v τ 31 ; W B 3 = v τ 41 .
In the case of a vehicle with more than two-axles there would be a larger number of wheelbase distances. Specifically there are N-1 wheelbase distances. Using the wheelbases and the number of axles, the vehicle can be classified using any standard classification scheme such as Table 1.

3.3 Variables that affect a vehicle classification

There are many factors affecting the operation of a classification. The Department of Transportation classifies a vehicle depending on the number of axles and the space between them. In order to measure the wheelbase some knowledge of the speed is necessary as indicated by the Equation (2). The accuracy measure time is very important in the estimation of speed and wheelbase. A potential source of error in the estimation of speed is the uncertainty in measurements. A small error in measurement of distance s/2 between sensor S1 and S3 or S3 and S4 will introduce a larger error in the estimation of a vehicle's speed.
Errors in wheelbase estimation can be caused by the vehicle when its speed is not constant while the vehicle drives over the sensor. Acceleration of the vehicle will cause the wheelbase to seem smaller, while deceleration will cause the wheelbase to seem larger according to the classification system. Usually a classification system is placed at certain points in the highway with traffic signs that state that the speed should remain constant.

3.4. Two-axle truck experimental results and discussion

Classification depends mostly on the accuracy in estimating the speed. Since the speedometer of a car is not sufficient accurate, it cannot be used as a speed reference. In order to acquire the system of measurement accuracy, the measurement results were compared with the results from a radar gun (Figure 8). CSR-68 Speed Radar Gun measuring accuracy is 0.1 km/h. A two-axle truck was used in this study.
Table 2 gives the experimental wheelbase measurement results. The wheelbase of the two-axle test truck was 6.67 m. The following data was obtained this vehicle with different speeds repeated 10 times. Again s/2 is the distance between the sensors S1 and S3 or S3 and S4 (cf. Figure 6). When the speed is known, the wheelbase may be calculated with Equation (3). The precision of the calculations is four decimal places; however in the Table below, to save space only the first two decimal places are shown.
The data in Table 2 show that the system is able to accurately estimate the speed and the wheelbase using multiple strain gauge sensors. Thus the use of multiple strain gauge sensors placed at different locations (refer to Figure 6) can decrease the uncertainties in measuring speed and wheelbase.

4. Description of SVM fusion classification

A classification based on a fixed wheelbase and number of wheelbases often leads to low accuracy. Since the patterns seemed to have a lot of overlap between them, this necessitated the use of SVMs pattern recognition and classification techniques to distinguish between vehicle groups.

4.2. Support vector classification

The SVM is a pattern recognition technique that is reported to be an excellent universal statistical learning machine with superior classification abilities [18]. The SVM classifies by mapping the data from an input space with an appropriate kernel function into a high dimensional feature space (hypothesis space) where a linear decision rule can be found based on observing the principle of maximizing the margin [19]. The SVM generates a function from labeled training data and predicts the output type of an unclassified novel input. The solution provided by SVM is said to always be optimal because of the absence of local extremes [20]. The method is structured in such a way that only a part of the training data, called support vectors, is used during training, hence avoiding computational complexity and providing better generalization. A detailed description of the general concept of SVMs is given by Burges [29], and Schölkopf and Smola [30]. This literature was the base for the choice of support vector machines for classification by researchers. A brief summary of SVM theory is given below.

4.2.1. Principle of Support vector Classification

Since an N-class decision problem can be decomposed into a set of two-class problems we will firstly concentrate here on a two-class problem.
If a given training set is denoted as(xi, yi)i =1,…,n, where xi is an input vector, yi is its label and n is the number of training data, then a SVMs classification model in a high-dimension feature space can be represented as follows:
f ( x ) = sgn ( < w , ϕ ( x ) > + b )
where, w is a weight vector, b is a bias, ϕ(x) is a nonlinear mapping from the input variable into a high dimension feature space <, > denotes the dot product.
The optimal classification hyperplane can be obtained by the following primal formulation:
min w , b , ζ 1 2 | | w | | 2 + C i = 1 n ζ i s . t y i ( < w , ϕ ( x ) > + b ) 1 ζ i i = 1 , , n ζ i 0 i = 1 , , n
where 1 2 | | w | | 2 controls the complexity of the model, ζi is a slack variable measuring the error on xi, C is a regularization parameter, when determines the trade off between the empirical error and the complexity of the model. A Lagrangian function corresponding to (5) can be described as follows by introducing Lagrange multiplies ai ≥ 0 and γi ≥ 0.
L = 1 2 | | w | | 2 + C i = 1 n ζ i i = 1 n α i [ y i ( < w , ϕ ( x i ) > + b ) 1 + ζ i ] i = 1 n γ i ζ i
Based on Karush-Kuhn-Tucker (KKT) optimality conditions:
w L = 0 b L = 0 ζ i L = 0
We can get the following relations:
w = i = 1 n α i y i ϕ ( x i ) i = 1 n α i y i = 0
And the box constraints:
0 α i C , i = 1 , , n .
Plugging the latter equations in the Lagrangian, we derive the formulation of the dual problem. Hence the SVM classification becomes a Quadratic Programming (QP) problem:
min a i , i = 1 , , n 1 2 i = 1 n j = 1 n α i α j y i y j k ( x i , y j ) j = 1 n α i
s . t . i = 1 n α i y i 0 α i C , i = 1 , n .
From the solution of the dual, only the data that corresponding to non-zero values of αi work, and they are called the support vector (SV). Equation 4 can then be represented as in Equation 9 and the decision function that separates training vectors into classes in the input space is:
d ( x ) = sgn ( f ( x ) ) = sgn ( i = 1 n α i y i k ( x i , x ) + b )
where k(xi,x)=< ϕ(xi), ϕ(x)> is the kernel function [21].

4.2.2. Support vector classification training

The algorithm used by the SVM scans the training data and chooses from the hypothesis space a function that can fit the data with minimum error. The learning problem is solved through trading off between training error and the complexity of the hypothesis space. Training is conducted through finding the optimal classification function by solving the following regularization problem:
min f H 1 n i = 1 n V ( f ( x i ) , y i ) + λ f k 2
The term λ is the regularization parameter and f k 2 is the norm of the hypothesis space [19]. The function V is called the loss function which measures how good the prediction function f(x) is in comparison to the training truth yi. The loss function used for classification purposes is a non-negative function of the form V(f(xi),yi) = max{(1–yif(xi)),0}. The penalty to errors is used to control regularization, hence using the definition of C = 1 2 n λ, the training problem becomes:
min f H 1 n C i = 1 n V ( f ( x i ) , y i ) + 1 2 f k 2

4.2.3 Multi-class Support Vector Machines

SVMs were originally designed for binary classification. From Table 1 we can see vehicle classification is a multi-class problem with more than two types. For the binary pattern recognition problem, the support vector was introduced in Section 4.2. In this section, we will review two popular SVM multi-class methods: One-Against-All (OAA) [22] and One-Against-One (OAO) [23].
The earliest used implementation for SVMs multi-class classification is probably the OAA method [24]. It constructs k SVMs models where k is the number of classes. The ith SVMs is trained with all of the examples in the ith class with positive labels, and all other examples with negative labels. Thus given l training data (xi, yi)i=1,…l, where xRn and yi ∈ {1,…,k is the class of labels (i.e., vehicle classes), the ith SVMs solves the following problem:
min w i , b i , ζ i 1 2 w i 2 + C j = 1 l ζ j i ( w i ) T ϕ ( x j ) + b i 1 ζ j i , if y j = i , ( w i ) T ϕ ( x j ) + b i 1 ζ j i , if y j i , ζ j i 0 , j = 1 , l .
where the training data xi are mapped to a higher dimensional space by the function ϕ and C is the penalty parameter.
Minimizing 1 2 w i 2 means that we would like to maximize 2/ ǁwiǁ2, the margin between two groups of data. When data are not linear separable, there is a penalty term C j = 1 l ζ j i which can reduce the number of training errors. The basic concept behind SVMs is to search for a balance between the regularization term 1 2 w i 2 and the training errors.
After solving (12), there are k decision functions:
( w 1 ) T ϕ ( x ) + b 1 , ( w k ) T ϕ ( x ) + b k .
We say x is in the class which has the largest value of the decision function:
class of d ( x ) arg max x i = 1 , , k ( ( w i ) T ϕ ( x ) + b i ) .
In practice we solve the dual problem of (12), whose number of variables is the same as the number of data in (12). Hence k l-variable quadratic programming problems are solved.
Another major method is called OAO method. This method constructs k ( k 1 ) 2 classifiers where each one is trained on data from two classes. For training data from the ith and the jth classes, we solve the following binary classification problem:
min w i j , b i j , ζ i j 1 2 / w i j 2 + C m = 1 l ζ m i j ( w i j ) T ϕ ( x m ) + b i j 1 ζ m i j , if y m = i , ( w i j ) T ϕ ( x m ) + b i j 1 ζ m i j , if y m = j , ζ m i j 0.
There are different methods for doing the future testing after all k ( k 1 ) 2 classifiers are constructed. After some tests, we decided to use the following voting strategy suggested in [25]: if the sign (wij)Tϕ(xm) + bij says x is in the ith class, then the vote for the ith class is added by one. Otherwise, the jth is increased by one. Then we predict x is in the class with the largest vote. The voting approach described above is also called the “Max Wins” strategy. In case those two classes have identical votes, though it may not be a good strategy, now we simply select the one with the smaller index.
In practice we solve the dual of (14) whose number of variables is the same as the number of data in two classes. Hence if on average each class has l / k data points, we have to solve k(k−1) / 2 quadratic programming problems where each of them has about 2l / k variables. The decision function for multi-class classification is:
d ( x ) = arg max j { i = 1 m | ( w i j ) T ϕ ( x m ) + b i j | }
That is, the class attribute of x is determined by sum of maximal distances to the optimal classification hyperplane.
There is a lot of software that uses SVMs to solve classification problems. In this research, the SVM-KM toolbox [26], which is a library of MATLAB routines, was used to handle multi-class problems. The use of SVM requires the preparation of two kinds of datasets, the training data with which to train a classification system and the test set on which to run the trained model. The collected data came from our developed test bench in situ (refer to Section 4.1).
The training of the SVM model basically involved the process of extracting the support vectors from the training data set. The support vectors defined the structure of boundary hyperplanes and the class hyperplanes after which the training data became useless. The so determined hyperplanes were stored for use of classification of novel data that had close similarity to the training data.

4.2.4 Kernel selection

Prior to the training of the support vector machine model, the selection of an appropriate kernel function with which to map the data from the input space to the high dimension feature space, was of utmost importance. Lin et al. [27] proposed the radial basis function (RBF) to be the first choice for support vector machines practitioner, the reasons behind this claim were, first, the linear kernel is a special case of RBF; secondly, the sigmoid kernel behaves like RBF for certain parameters; thirdly the polynomial kernel has many parameters which makes their selection complex and fourthly and last, the RBF has less numerical difficulties. For these reasons, the RBF whose equation is shown below was used in this research:
K ( x i , x j ) = exp ( x i x j 2 2 σ 2 ) , σ > 0

4.2.5 Cross-validation

The two parameters for the RBF model, the penalty C and the kernel function parameter σ have the best values for only one single problem. The values of these parameters should be known beforehand for a given problem prior to model training. The method used in the SVM-KM toolbox to find the values of these parameters is cross-validation. This cross-validation involves splitting the training data into subsets, leaving out one subset and training the classifier with different values of C and σ on the remaining subsets and measure the performance on the subset initially left out. The values of C and σ giving the highest cross-validation rates are chosen for training the classifier on the whole data set. If the best values of the parameters for a problem are not found, the obtained classifier will suffer from over fitting.

4.2.6 Model training process

The C and α values obtained in the cross–validation were used for the training of the support vector machines. The training task involved the determination of the number of support vectors from each class, these vectors were used for the determination of weight vector, and the class and margin hyperplanes. When the training was over, a model file was created; this model file contained all the determined support vectors derived from the training data which were used in the classification of test data. After the determination of support vectors was done, the training data was not used anymore.

4.2.7 Data Fusion Classification Strategies

During the past decades, the data fusion problem has been well researched. However, it is still an ongoing research area because of the promotion from advances in other fields. The synergistic use of overlapping and complementary data sources provides information that is otherwise not available from individual sources. Furthermore, multiple data sources can provide more robust performance due to the inherent redundancy. Therefore data fusion techniques of combining data from several data sources can yield higher accuracy and robustness than that achieved by single data source. The main architecture for fusion is centralized and distributed. In this paper, centralized and distributed architectures based on multiclass SVMs are introduced. The fusion schemes make full use of multiclass SVMs characteristic. The centralized fusion scheme is illustrated as the fusion Scheme A. the distributed fusion schemes is illustrated as the fusion Scheme B.
In fusion Scheme A (see Figure 9), the features of all the sensor data sources are extracted and combined to form a single input space. Then the multiclass SVMs is trained and tested to create a decision maker.
In the fusion Scheme B (Figure 10), the features of every sensor data source are extracted and used to form an input space, respectively. Then the sub-Multiclass SVMs final classification decision is created. Here, the class attribute set is k= {1, 2, 3, 4, 5}, using the majority vote strategy, the final decision is:
d ( x ) = arg max { V 1 , V 2 , , V k } , V j = i = 1 N δ i j , δ i j = { 1 , d i ( x ) = j 0 , d i ( x ) j ( i = 1 , , N ; j = 1 , , k )
where d (x)is the final classification decision function, Vj is comprised of the obtained votes of class j and di (x) is the output of the ith Multiclass SVMs trained using the data from the ith sensor data source.

5. Experimental Results

Since freeways carry large volumes of vehicles, our data collection was installed on freeways to increase a possibility of having many vehicle types in the dataset, whilst using multiple strain gauge sensors for the measurements, thereby improving the accuracy of the classification parameters [17].
Data from five strain gauge sensors were collected in situ when a vehicle crossed the instrumental pavement. The panel was adopted as a test bed for periodic continuous monitoring. In 2005, during the 10 day data-collection experiment, through the data acquisition system, live data of pavement strain response induced by traffic was collected and stored. To thoroughly utilize the useful information, two features (number of axle and wheelbase) are extracted from the pavement strain time series (see Section 3.2). 602 different vehicle records were obtained. The records included a wide variety of vehicle types, ranging from 2-axle passenger cars to 6-axle semi-trailer trucks. Each instance is composed of 6 condition attributes (two features from each time series) and a class attribute (five states). In distributed schemes, the two features from one sensor data, added the class attribute, form an individual dataset [see (17)]. Therefore, three datasets from the corresponding three data sources are constructed. Following the classification scheme of Table 1, 602 records (Table 3) was manually labeled by artificial observation. The labeled records of each vehicle class are divided into a training set (50% of the records), validation set (25% of the records), and test set (25% of the records) of randomly selected records. Herein, this dataset was employed to develop and verify a strain-based vehicle classification approach.
For S 1 { N 1 1 , W B 1 1 , W B 2 1 , W B 3 1 , W B 4 1 , W B 5 1 , label } , For S 3 { N 2 3 , W B 1 3 , W B 2 3 , W B 3 3 , W B 4 3 , W B 5 3 , label } , For S 4 { N 3 4 , W B 1 4 , W B 2 4 , W B 3 4 , W B 4 4 , W B 5 4 , l a b e l } , For S 1 , S 3 , S 4 combination { N 1 1 , W B 1 1 , W B 2 1 , W B 3 1 , W B 4 1 , W B 5 1 , N 2 3 , W B 1 3 , W B 2 3 , W B 3 3 , W B 4 3 , W B 5 3 , N 3 4 , W B 1 4 , W B 2 4 , W B 3 4 , W B 4 4 , W B 5 4 , l a b e l } ,
where the labels are numbered 1,2,3,4 and 5 to represent five classes vehicle (refer to Table 1.). All of the datasets contain a separate test set. A description of the data we used is given in Table 3.
The experiments in this paper are conducted on single sensor data and 3-sensor (S1, S3, S4) combination data sets, using the two different classifier algorithms (OAO and OAA). Two multiclass SVM algorithms are applied twice: on the individual sensor S3 data and 3-combination combination data. In both algorithms we used Gaussian kernels. To determine the values of C and σ we used cross validation on the training set. In all the experiments we set the value of ζ to 0.01. A summary of the results is shown in Table 4. The average classification accuracies of two types of Multiclass SVMs are given in columns D1, D3 and D4. Scheme I is the centralized fusion strategy. Scheme II is distributed strategies of data fusion based on majority-vote strategy. Table 4 shows that all the average classification accuracies of combined classifiers in fusion strategies outperform those of the classifiers not using fusion strategies. If the centralized data fusion strategy is used, the accuracy of the two types of Multiclass SVMs methods differs little (Column S.I in Table 4). However, in the distrusted data fusion Scheme II, the One-Against-One method has higher accuracy than One-Against-All method (Column S.II in Table 4).

6. Conclusions

A novel vehicle classification technique has been developed based on multiple pavement strains caused by moving traffic loads. Pavement strains are often used for monitoring the health of the road structure; here we regard it as a method of classification of vehicles. The main advantages of vehicle classification based on pavement strain are as follows: firstly, due to the installation of strain gauge sensors underneath the pavement, the sensor will not be affected by the impact load and has better durability; secondly, sensors would not be affected by bad weather; thirdly, the system may monitor day and night, especially at night so overloaded vehicles cannot escape monitoring; fourthly, since sensors are low-cost, it can provide a cheaper vehicle classification. It is an invasive measurement method, often installed my modifying the road, which is its main drawback.
A prototype system for measurement of pavement strain induced by moving vehicle traffic was developed and built to collect data for verifying and testing the feasibility and performance of the system. The estimation of the speed is very important, as it often affects the accuracy of the estimated vehicle wheelbase. In this research multiple strain sensors were used to measure speed from the average, so to a certain extent, estimation speed errors caused by acceleration and deceleration of vehicles will be eliminated. As an added benefit, multiple sensors can also expand the scope of measurements, providing more vehicles feature parameters.
The overlap of vehicle classification feature parameters belonging to different classes suggested the need to use a pattern recognition technique for separating vehicles into different groups. As an application of machine learning the support vector machines (SVMs) multi-class model was used for this purpose. To improve classification accuracy and robustness centralized and distributed fusion schemes based on two popular SVMs multi-class algorithms were used as fusion multiple sensor data. Comparison of experimental results shows the OAA and OAO methods with distributed fusion strategies are more suitable for practical use. In the paper, only two features are extracted from the pavement strain data and used to train SVM classifier. To improve classification accuracies and robustness, more features (such as vehicle length, axle load etc.) from strain time series to train SVM classifier will be studied.

Acknowledgments

The help of the School of Science and Engineering on Communication of Harbin Institute of Technology is gratefully acknowledged. They funded the instrumentation used in this work, and provided the test vehicles, the test site, and assistance during the test runs. Support of this research was provided by China Heilongjiang Provincial Communications Department Fund, Grant No. HJZ_2004_12. This support is gratefully acknowledged.

References and Notes

  1. Bajaj, P.; Sharma, P.; Deshmukh, A. Vehicle Classification for Single Loop Detector with Neural Genetic Controller: A Design Approach. Intelligent Transportation Systems Conference, Seattle, Washington, USA, Sept. 30-Oct. 3, 2007; pp. 721–725.
  2. Gajda, J.; oka, R.; Stencel, M.; Wajda, A.; Zeglen, T. A vehicle classification based on inductive loop detectors. Instrumentation and Measurement Technology Conference, Budapest, Hungary, May 21-23, 2001; pp. 460–464.
  3. Ki, Y.K.; Baik, D.K. Vehicle-Classification Algorithm for Single-Loop Detectors Using Neural Networks. Vehicular Technology Conference, Melbourne, Australia, May 7-10, 2006; pp. 1704–711.
  4. Lin, P.Q.; Xu, J.M. Adaptive Vehicle Classification Based on Information Gain and Multi-branch BP Neural Networks. Intelligent Control and Automation, Dalian, China, June 21-23, 2006; pp. 8687–8691.
  5. Hussain, K.F.; Moussa, G.S. Automatic vehicle classification system using range sensor. ITCC'05 2005, 2, 107–112. [Google Scholar]
  6. Ha, D.M.; Lee, J.-M.; Kim, Y.-D. Neural-edge-based vehicle detection and traffic parameter extraction. Imag. Vis. Comp. 2004, 22, 899–907. [Google Scholar]
  7. Stefano Messelodi, S.; Modena, C.M.; Cattoni, G. Vision-based bicycle/motorcycle classification. Patt. Rec. Lett. 2007, 28, 1719–1726. [Google Scholar]
  8. Fernández-Caballero, A.; Gómez, F.J.; López-López, J. Road-traffic monitoring by knowledge-driven static and dynamic image analysis. Expert Syst. Appl. 2008, 35, 701–719. [Google Scholar]
  9. Klausner, A.; Tengg, A.; Rinner, B. Vehicle Classification on Multi-Sensor Smart Cameras Using Feature-and Decision-Fusion. ICDSC '07 2007, 67–74. [Google Scholar]
  10. Gupte, S.; Masoud, O.; Papanikolopoulos, P. Vision-based vehicle classification. ITS 2000, 46–51. [Google Scholar]
  11. Goyal, A.; Verma, B. A Neural Network based Approach for the Vehicle Classification. Computational Intelligence in Image and Signal Processing 2007, 226–231. [Google Scholar]
  12. Junghans, M.; Jentschel, H.-J. Qualification of traffic data by Bayesian network data fusion. Inf. Fusion 2007, 1–7. [Google Scholar]
  13. Sroka, R. Data fusion methods based on fuzzy measures in vehicle classification process. Instrumentation and Measurement Technology Conference, Como, Italy, May 18-20, 2004; pp. 2234–2239.
  14. Gunn, S.R.; Brown, M.; Bossley, K.M. Network performance assessment for Neuro-fuzzy data modeling. Intell. Data Anal. 1997, 1208, 313–323. [Google Scholar]
  15. Castro, L.N.; Iyoda, E.M.; Zuben, F.V.; Gudwin, R. Feedforward Neural Network Initialization: an evolutionary approach. SBRN, Belo Hoizonte, Brazil, Dec 9-11, 1998; pp. 43–48.
  16. Zhang, W.B.; Wang, Q.; Ma, S.L.; Li, X.K. Field experimental study on measurement and analysis strain on the rigid pavement slab subjected to moving vehicle loads. The First International Conference of Transportation Engineering, Chengdu, China, Jul 22-24, 2007; pp. 2741–2746.
  17. Zhang, W.B.; Wang, Q.; Ma, S.L.; Li, X.K. A novel multisensor system for moving vehicle classification. J. TianJin University 2008, 41, 194–198. [Google Scholar]
  18. Brailvosky, V.L.; Barzilay, O.; Shahave, R. On global, local and mixed neighborhood kennels for support vector machines. Patt. Rec. Lett. 1999, 20, 1183–1190. [Google Scholar]
  19. Madevska-Bogdanova, A.; Nikolik, D.; Curfs, L. Probabilistic SVM output for pattern recognition using analytical geometry. Neurocomputing 2004, 62, 293–303. [Google Scholar]
  20. Paysan, P. Stereovision based vehicle classification using support vector machines. Thesis, Massachusetts Institute of Technology, Cambridge University, Cambridge, Massachusetts, 2004. [Google Scholar]
  21. Schölkopf, B.; Burges, D.J.C.; Somla, A.J. Advances in Kernel Methods; The Massachusetts Institute of Technology Press: Cambridge, Massachusetts, 1999. [Google Scholar]
  22. Rifkin, R.; Klautau, A. In Defense of One-Vs-All Classification. J. Mach. Learn. Res. 2004, 5, 101–141. [Google Scholar]
  23. Guermeur, Y.; Elisseeff, A.; Zelus, D. A comparative study of multi-class support vector machines in the unifying framework of large margin classifiers. Appl. Stochastic Model. Bus. Ind. 2005, 21, 199–214. [Google Scholar]
  24. Bredensteiner, E. J.; Bennett, K. P. Multicategory classification by support vector machines. Comput. Optimiz. Appl. 1999, 53–79. [Google Scholar]
  25. Friedman, J. Another approach to polychotomous classification; Technical report. Department of Statistics, Stanford University, 1996. Available at http://www-stat.stanford.edu/reports/friedman/poly.ps.z.
  26. The SVM-KM toolbox. http://asi.insa-rouen.fr/enseignants/arakotom/toolbox/index.html.
  27. Hsu, C.; Chang, C.; Lin, C. A practical guide to support vector classification. Available at http://www.csie.ntu.tw/∼cjlin/libsvm/index.html (10/14/2005).
  28. Sun, C.; Ritchie, S.G.; Oh, S. Inductive classifying artificial network for vehicle type categorization. Comput.-Aided Civ. Inf. Eng. 2003, 161–172. [Google Scholar]
  29. Burges, C.J.C. A tutorial on support vector machines for pattern recognition. Data Min. Knowl. Discov. 1998, 2, 121–167. [Google Scholar]
  30. Schölkopf, B.; Smola, A. Learning With Kernels.; MIT Press: Cambridge, MA, 2002. [Google Scholar]
Figure 1. The layout plan of the strain gauge sensors.
Figure 1. The layout plan of the strain gauge sensors.
Sensors 08 06952f1
Figure 2. Photos of: (a) the data collection device (b) the data collection card and interface board.
Figure 2. Photos of: (a) the data collection device (b) the data collection card and interface board.
Sensors 08 06952f2
Figure 3. Illustration of the strain caused by moving wheel loads.
Figure 3. Illustration of the strain caused by moving wheel loads.
Sensors 08 06952f3
Figure 4. The structural sketch of the embedded strain gauge sensor.
Figure 4. The structural sketch of the embedded strain gauge sensor.
Sensors 08 06952f4
Figure 5. Photos of the embedded strain gauge sensor.
Figure 5. Photos of the embedded strain gauge sensor.
Sensors 08 06952f5
Figure 6. Characteristics of a four axle traffic-induced strain response time history.
Figure 6. Characteristics of a four axle traffic-induced strain response time history.
Sensors 08 06952f6
Figure 7. Vehicle classification using multiple embedded strain gauge sensors.
Figure 7. Vehicle classification using multiple embedded strain gauge sensors.
Sensors 08 06952f7
Figure 8. CSR-68 Speed radar gun.
Figure 8. CSR-68 Speed radar gun.
Sensors 08 06952f8
Figure 9. Data fusion architecture Scheme A.
Figure 9. Data fusion architecture Scheme A.
Sensors 08 06952f9
Figure 10. Data fusion architecture Scheme B.
Figure 10. Data fusion architecture Scheme B.
Sensors 08 06952f10
Table 1. Definition of Vehicle Classes.
Table 1. Definition of Vehicle Classes.
Vehicle typeNotationNumber of axlesDistribution of axles1Vehicle typesFHWA vehicle categories
Small vehiclesC121F+1RPassenger car, Minivan, van, SUV, pickup truckPassenger car, other 2-axle 4-tire vehicles
Medium trucksC221F+1RMedium single-unit 2-axle truckOther 2-axle 4-tire vehicles
Buses/Large trucksC321F+1RBuses, large single-unit 2-axle trucksBus, single-unit 2-axle, 6-tire or more truck
3-axle trucksC431F+2RSingle-unit 3-axle truckssingle-unit 2-axle 6-tire or more trucks
Combination trucksC53-61F+1M+1R
1F+2M+1R
1F+1M+2R
1F+2M+2R
1F+2M+3R
Semi-trailer, truck, trucks with trailercombination trucks
Note: 1.F-Front, M-Middle, R-Rear. (For instance: 1F+1R indicate a vehicle with one front axle and one rear axle.)
Table 2. The wheel base measurement using the proposed system.
Table 2. The wheel base measurement using the proposed system.
Numberτ11(sec)τ31(sec)τ41(sec)v1(m/s)v2(m/s)v(m/s)WB(m)Actual WB (m)WB Error
10.540.520.5312.2412.4412.346.586.67-1.3%
20.420.420.4215.2715.1915.236.456.67-3.2%
30.450.450.4514.0114.2114.116.396.67-4.1%
40.510.520.5213.3413.1413.246.826.670.7%
50.430.430.4315.1114.9115.016.496.67-2.6%
60.590.590.5911.2811.3411.316.726.67-0.7%
70.510.510.5112.7912.9612.856.596.67-1.1%
80.540.540.5412.8712.6712.776.886.671.7%
90.610.610.6110.7510.9510.856.616.67-0.8%
100.520.520.5212.7512.9512.856.696.670.3%
Table 3. Numbers of Labeled Records.
Table 3. Numbers of Labeled Records.
Vehicle typeTotalTraining setValidation setTest set
C172361818
C268341717
C3174884242
C484422121
C52041025151
Table 4. Comparison of Classification Results.
Table 4. Comparison of Classification Results.
Multiclass SVMsSingle sensor data sourceS.IS.II

D1D2D3
One-Against-All (%)89.592.390.994.695.5
One-Against-One (%)91.491.891.294.496.4
Note: D1, D2, D3: Sensor S1 data, Sensor S3 data, Sensor S4 data; S.I, S.II: Scheme I, Scheme II.

Share and Cite

MDPI and ACS Style

Zhang, W.; Wang, Q.; Suo, C. A Novel Vehicle Classification Using Embedded Strain Gauge Sensors. Sensors 2008, 8, 6952-6971. https://doi.org/10.3390/s8116952

AMA Style

Zhang W, Wang Q, Suo C. A Novel Vehicle Classification Using Embedded Strain Gauge Sensors. Sensors. 2008; 8(11):6952-6971. https://doi.org/10.3390/s8116952

Chicago/Turabian Style

Zhang, Wenbin, Qi Wang, and Chunguang Suo. 2008. "A Novel Vehicle Classification Using Embedded Strain Gauge Sensors" Sensors 8, no. 11: 6952-6971. https://doi.org/10.3390/s8116952

APA Style

Zhang, W., Wang, Q., & Suo, C. (2008). A Novel Vehicle Classification Using Embedded Strain Gauge Sensors. Sensors, 8(11), 6952-6971. https://doi.org/10.3390/s8116952

Article Metrics

Back to TopTop