Next Article in Journal
Novel Calibration Approach for Monitoring Aerosol Hydrogen Isotopes Using Laser-Induced Breakdown Spectroscopy for Molten Salt Reactor Off-Gas Streams
Previous Article in Journal
Validity of AI-Based Gait Analysis for Simultaneous Measurement of Bilateral Lower Limb Kinematics Using a Single Video Camera
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrating Surrounding Vehicle Information for Vehicle Trajectory Representation and Abnormal Lane-Change Behavior Detection

1
Intelligent Transportation Systems Research Center, Wuhan University of Technology, Wuhan 430063, China
2
Shandong Hi-Speed Group Innovation Research Institute, Jinan 250014, China
3
Engineering Research Center of Transportation Information and Safety, Ministry of Education, Wuhan 430063, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(24), 9800; https://doi.org/10.3390/s23249800
Submission received: 17 November 2023 / Revised: 8 December 2023 / Accepted: 8 December 2023 / Published: 13 December 2023

Abstract

:
The detection of abnormal lane-changing behavior in road vehicles has applications in traffic management and law enforcement. The primary approach to achieving this detection involves utilizing sensor data to characterize vehicle trajectories, extract distinctive parameters, and establish a detection model. Abnormal lane-changing behaviors can lead to unsafe interactions with surrounding vehicles, thereby increasing traffic risks. Therefore, solely focusing on individual vehicle perspectives and neglecting the influence of surrounding vehicles in abnormal lane-changing behavior detection has limitations. To address this, this study proposes a framework for abnormal lane-changing behavior detection. Initially, the study introduces a novel approach for representing vehicle trajectories that integrates information from surrounding vehicles. This facilitates the extraction of feature parameters considering the interactions between vehicles and distinguishing between different phases of lane-changing. The Light Gradient Boosting Machine (LGBM) algorithm is then employed to construct an abnormal lane-changing behavior detection model. The results indicate that this framework exhibits high detection accuracy, with the integration of surrounding vehicle information making a significant contribution to the detection outcomes.

1. Introduction

Lane-changing (LC) is one of the fundamental and indispensable driving behaviors of vehicles. When drivers aim to achieve more favorable driving conditions or perform diverging or merging operations, lane-changing behavior occurs [1]. During a lane-changing maneuver, drivers must monitor the traffic conditions in adjacent lanes and make decisions based on relative speed and distance; therefore, the driver’s field of vision involves extensive scanning of visual hazards from the front, sides, and even rear. They need to sift through information and make quick judgments. If the driver’s perception and operation are accurate and appropriate during a lane-changing behavior, it will be completed safe. However, in reality, the driver’s cognitive load significantly increases during the lane-changing process, making driving operations more error-prone and leading to driving risks [2,3,4,5]. Studies have confirmed that unsafe lane-changing behavior is one of the main causes of traffic instabilities, such as decreases in road capacity and speed, thus drawing increasing attention to lane-changing research in recent years [6,7,8].
Human factors are the primary causes of unsafe driving behaviors. Abnormal driving behaviors resulting from psychological states and attitudes can easily lead to traffic risks and even accidents [9,10,11]. Therefore, detecting and identifying abnormal driving behaviors, including abnormal lane-changing behavior, and subsequently regulating drivers’ driving habits through supervision, training, and education can reduce the occurrence of hazardous behaviors, thereby lowering the risk of road travel [12].
In recent years, artificial intelligence (AI) technology has accelerated the development of Intelligent Transportation Systems (ITS), with Machine Learning (ML) and Deep Learning (DL) techniques being widely applied in the detection of abnormal driving behaviors [13,14]. According to statistics in the field of driving behavior detection, ML algorithms are the most extensively used, accounting for 62.66%, followed by DL algorithms at 32.91% and statistical methods at 4.43% [15]. The use of ML and DL techniques has been proven as effective for driving behavior detection, leading to a better understanding of the causes of aberrant driving and an enhancement of road safety.
Regarding data sources, research on driving behavior mainly focuses on two aspects: objective and subjective measurement. Subjective measurement is primarily in the form of questionnaires, with The Manchester Driver Behavior Questionnaire (DBQ) being the most representative [16,17]. However, questionnaire surveys may introduce human subjectivity, leading to biases. Objective measurement can obtain quantitative data, and it comes in three main forms: driving simulator studies (DSS), field driving studies (FDS), and naturalistic driving studies (NDS) [18,19,20]. DSS collects simulated driving data from drivers on simulators. NDS uses unobtrusive measurement methods to record detailed information about the driver, vehicle, and surrounding environment; commonly used sensors include GPS, accelerometers, smartphone sensors, and OBDII sensors, among others [21,22,23]. However, in these two testing methods, drivers are consciously engaged in experimental testing, leading to deviations from actual driving performance. FDS uses equipment such as cameras and radar to monitor driver behavior at fixed points, encoding and recording driving behavior data; in comparison, it can obtain more objective and larger-scale data [24,25]. With the continuous development of Vehicle-to-Infrastructure (V2I) technology, road perception units, represented by high-definition cameras, can promptly acquire motion trajectory information, such as the position and speed of road vehicles. Through advanced communication technology, these units provide real-time road information to travelers and road regulatory authorities. This not only serves as a data source for the study of driving behavior but also establishes the foundational conditions for the application of abnormal behavior detection in road supervision [26,27,28,29].
In summary, video data and ML algorithms are feasible for detecting abnormal lane-changing behavior. Therefore, this study utilizes road video data and employs the approach of ML algorithms to conduct research on abnormal lane-changing behavior detection, aiming to serve the management and law enforcement of road traffic. Generally, there are two core tasks when using ML algorithms for abnormal lane-changing behavior detection: trajectory representation and behavior detection. Among these, trajectory representation serves as the foundation, extracting feature metrics from the trajectory to serve as input for the detection model. In many studies, the vehicle trajectory is directly represented through a coordinate vector and its variants [30]. Johnson et al. pioneered the description of object motion using a flow vector (x, y, dx, dy) [31]. Building upon this, subsequent research incorporated more information into the trajectory, such as speed, angle, position, etc., forming higher-dimensional coordinate vector representations [32,33]. Trajectory representation based on curve fitting has also been applied; common practice involves using low-order curves to approximate trajectories, such as using cubic B-spline curves for approximation [34]. However, methods based on curve fitting have obvious limitations, as the information extracted from the approximated trajectory shape is quite limited, making it difficult to discern underlying behavioral patterns. In recent years, deep learning methods have gradually gained widespread application in vehicle trajectory representation [35]. Lv et al. transformed trajectories into images and used convolutional neural networks to extract features from them [36]. Tang et al. employed hidden Markov models to learn and train trajectories in taxi route prediction [37]. Liang et al. used an Long Short-Term Memory (LSTM) encoder–decoder model to represent each trajectory as a series of intersections and related movement directions, which were then fed into the network for future trajectory generation [38]. However, despite their powerful automatic feature representation learning capabilities, data-driven deep learning methods for trajectory representation tend to have poorer interpretability. Apart from the aforementioned methods, there has also been some research on using natural language and graphic images to represent trajectories [39].
Despite some research having been conducted on vehicle trajectory representation, challenges persist [40]. As shown in Figure 1a, the existing trajectory representations described above mainly focus on describing the motion information of individual subject vehicles (SV), such as speed and position. However, during the actual lane-changing process, if the subject vehicle (SV) is too close to the vehicles behind it in the target lane (TVTL), leading to malicious lane-changing behavior, as illustrated in Figure 1b, the existing trajectory representation methods cannot accurately describe it. Therefore, existing trajectory representations suffer from the issue of being centered around a single subject vehicle, resulting in generated trajectories lacking interaction information between vehicles or targets.
The direct purpose of acquiring vehicle trajectories is to extract feature metrics and apply them to detection models. Many studies have focused on using physical quantities and their variants such as speed and displacement as the features of vehicle trajectories for subject vehicles. Regarding lane-changing behavior, studies in the past have concentrated on key factors like lane-changing frequency, speed, acceleration, and headway distance [41]. Yao et al. further refined these physical quantities by using their statistical measures, such as mean, extremum, and percentiles, rendering the trajectory features more nuanced [42]. In recent years, interactions between vehicles have also been considered in the selection of feature metrics, such as gaps and relative speeds with surrounding vehicles [43,44]. This has been applied to the study of lane-changing behavior, primarily throughout the entire lane-changing process. Thus, there is a lack of targeted research on feature extraction for different stages of lane-changing.
In terms of behavior detection, various machine learning algorithms such as neural network models [45,46], Support Vector Machine (SVM) [47,48,49], Dynamic Bayesian Networks [50,51], and Light Gradient Boosting Machine (LGBM) have been widely employed [52]. Among these, neural network models, due to their complexity, can lead to increased training times and potential entrapment in local or global optima [53]. Support Vector Machine (SVM), as a powerful tool for data classification, has extensive applications in pattern recognition research [54], but it is computationally complex and performs poorly when the sample sizes of various classes in the dataset are imbalanced. Dynamic Bayesian Networks are similarly computationally complex and handling high-dimensional data can be challenging. The LGBM algorithm, introduced in 2017, exhibits advantages over traditional algorithms, including low memory usage, higher accuracy, and the ability to handle high-dimensional and large-scale data [55].
In summary, despite the extensive research in trajectory representation and behavior detection methods, the existing studies exhibit the following gaps:
(1) The majority of existing trajectory representation studies are based on the subject vehicle, with limited consideration for surrounding vehicles. A trajectory representation method that integrates information from surrounding vehicles has not been proposed. Moreover, the selected trajectory feature indicators are mostly applied to the entire lane-changing process, lacking targeted proposals or selections for different stages of lane-changing.
(2) Compared to traditional machine learning models, the Light Gradient Boosting Machine (Light GBM) algorithm is better suited for the classification and detection of large-scale, high-dimensional data.
To address these gaps, we first propose a trajectory representation method that integrates information from surrounding vehicles. Subsequently, we define feature indicators that characterize different stages of lane-changing using this method. We summarize and categorize three typical instances of abnormal lane-changing behavior, followed by the application of the Light Gradient Boosting Machine (Light GBM) algorithm for the detection of abnormal lane-changing behavior.
The organization of this paper is as follows: The Section 2 outlines the data sources, trajectory representation method, feature extraction approach, and the establishment of the model for detecting abnormal lane-changing behavior. The Section 3 presents the experimental results and discussions. Finally, the Section 4 provides the conclusion and outlines future work.

2. Methodology

2.1. Research Framework

This study leverages roadside video data to propose a trajectory representation method that integrates information from surrounding vehicles. It combines machine learning techniques for detecting abnormal lane-changing behavior. This process encompasses three key tasks:
(1) Trajectory representation and feature extraction—achieving a trajectory representation that integrates information from surrounding vehicles. This involves considering the mutual interactions between vehicles and extracting features at different stages of lane-changing. Subsequently, lane-changing events are further extracted.
(2) Labeling of abnormal lane-changing behavior: three typical instances of abnormal lane-changing behavior are defined. A dataset of abnormal lane-changing behavior is established for training the model.
(3) Model establishment for abnormal lane-changing detection: the framework is designed to detect abnormal lane-changing behavior, leveraging the dataset established in the previous step.
The research framework is illustrated in Figure 2.

2.2. Data Source

This study selected the Luoshi Road elevated expressway in Wuhan City as the observation site. Six observation points were established by six experimenters at elevated writing platforms on the roadside, equipped with fixed motion cameras. The observation scope covered the section from Bayi Road to Mafangshan Tunnel, with a total effective road length of 1.41 km, including various types of sections such as weaving zones, merging, and diverging areas. After careful screening and inspection of the videos captured at the six points, valid videos from 14:30 to 18:00 were retained.
At Point 3, the length of the observed road segment was 205 m, with a dual-lane upstream section and a downstream section with four lanes, including merging and diverging areas. Due to the higher frequency of lane-changing events in the downstream section at this point, as well as the diversity of lane-changing types, this study chose the data from the downstream section at Point 3 to construct the vehicle trajectory dataset for further research. Figure 3 illustrates the vehicle detection process after perspective transformation of the captured images at Point 3. This study employed a Cartesian coordinate system to describe the trajectories, with the Y-coordinate defined as parallel to the direction of the lanes and the X-coordinate defined as perpendicular to the lane direction. The origin of the coordinates was set at the lower-left corner of the road section at Point 1, with the coordinate unit in terms of pixel values from the images. After applying perspective transformation to the dataset, performing vehicle object detection, and associating data, a total of 11,168 valid vehicle trajectories were extracted, providing motion information including vehicle position, speed, acceleration, etc. (at intervals of 0.1 s).

2.3. Trajectory Representation and Feature Extraction

Definition 1 (Trajectory Point).
The trajectory of a vehicle at time i is represented as Ti = (Si,ti), where si is a sequence of pairs representing the precise spatial locations of the vehicle’s trajectory points and ti is the timestamp indicating the time index of the trajectory point.
Definition 2 (Trajectory).
In geographical space, a time series of position sampling points with a certain time interval that can approximate the actual movement process of a mobile object is referred to as a trajectory. It can be expressed as a finite time series. In this paper, the trajectory represents the vehicle’s driving process on the road:
T = ( T 1 , T 2 , T 3 , , T i , , T n ) , i 1 , n
Definition 3 (Lane ID).
The detection range of the roadside perception unit is fixed, and the lane position is fixed. Therefore, the current lane ID of the target vehicle can be identified based on the vehicle’s position (e.g., centroid position). The Lane ID defined in this paper is the lane ID where the target vehicle is located.
Definition 4 (Surrounding Vehicles).
By utilizing the roadside perception unit, the positions of all vehicles within the perception range can be extracted. Based on this, the trajectory representation method proposed in this study integrates information from surrounding vehicles. The definition of surrounding vehicles is shown in Table 1. This method is universally applicable, suitable for both the following process and lane-changing process of vehicles. Furthermore, this study focuses on the research of abnormal behaviors during the vehicle lane-changing process and provides specific definitions and descriptions for surrounding vehicles in the lane-changing process (refer to Table 1). Corresponding illustrations are shown in Figure 4, where Figure 4a is applied to trajectory representation and Figure 4b is applied to lane change feature extraction.
Definition 5 (Trajectory Representation).
In this paper, considering the integration of information regarding the positions of surrounding vehicles and the lane information, a novel method for representing vehicle trajectories is proposed. Specifically, it employs (SSV, SLPV, SLTV, SPV, STV, SRPV, SRTV) to represent the positional information of the target vehicle and surrounding vehicles, where ti represents the timestamp and r indicates the lane ID. The specific form is as follows:
T i = s S V , s L L V , s L T V , s L V , s T V , s R L V , s R T V , t i , r T = ( T 1 , T 2 , T 3 , T i , T n ) , i 1 , n
Definition 6 (Lane-Change Feature Extraction).
The proposed trajectory representation method allows for the extraction of characteristic variables that represent the interactions between vehicles during lane-changing processes. These features can be directly computed from the trajectory representation results. For instance, variables like v and a can be obtained from the instantaneous positions of both the target vehicle and surrounding vehicles. The occurrence time of a lane change can be determined by checking if the lane ID has changed.
The longitudinal velocities of vehicles SV, PV, PVTL, TVTL, PVS, and TVS are defined as v, v1, v2, v3, v4, and v5, respectively. The relative speed between vehicles is defined as follows:
v r 1 = v 1 v v r 2 = v 2 v v r 3 = v v 3 v r 4 = v 4 v v r 5 = v v 5 ,
Lane-changing behavior is a continuous process, typically lasting several seconds, encompassing the driver’s decision-making, preparation, and the execution and completion of the lane change. Information such as the positions and speeds of surrounding vehicles before a lane change affects the driver’s decision and preparation. Additionally, the states of preceding and following vehicles after a lane change influence the progression and duration of the driver’s lane-changing action. Furthermore, the driver’s driving habits and style also impact the behavioral performance of the entire lane-changing process. Therefore, this study primarily considers four factors in selecting feature indicators:
(1) Multi-vehicle variables, obtained through trajectory representation calculations that integrate information from surrounding vehicles.
(2) Single-vehicle variables, taking into account the influence of driver habits and style on the entire lane-changing process.
(3) Before-lane-change features, considering the impact of surrounding vehicles on the driver’s lane change preparation and decision-making before a lane change.
(4) After-lane-change features, considering the partial impact of surrounding vehicles on the driver’s completion of the lane change and the subsequent lane-keeping process.
For each LC event, a total of 156 candidate feature indicators were selected. Table 2 provides detailed descriptions of these candidate features.

2.4. Abnormal LC Behavior Labeling

In this study, lane-changing trajectories were extracted from trajectory data obtained through video analysis. Both normal and abnormal lane-changing behaviors were annotated based on original video footage. The specific steps are outlined below:
Step 1: Constructing trajectories by integrating surrounding vehicle information. Surrounding vehicle data, including the position and lane ID of surrounding vehicles as defined in Table 1, were extracted. These were then represented in the form of trajectories as proposed in Section 2.3. If a vehicle was not present in a specific position around the SV vehicle, the relative positional information was set as infinity.
Step 2: Labeling of lane-changing events. For each trajectory, lane changes were determined based on lane IDs. If a change in lane ID was identified, the trajectory was marked as a lane-changing trajectory. Furthermore, the target lane of the vehicle was determined based on the change in lane ID.
Step 3: Extraction of lane-changing segments. Previous studies have indicated that the duration of lane-changing events does not exceed 5 s, with lane-changing processes occurring between 2 to 6 s constituting nearly 95% of cases. Additionally, the preparation time for lane changes takes approximately 2 s [56]. As this study primarily focuses on abnormal behaviors occurring before and after lane changes rather than the intention to change lanes, a reference point tLC was established based on the time of lane ID change. A fixed time duration of 6 s was employed to extract different lengths of time windows both before and after tLC (3 s before lane change +3 s after, 2 s before +4 s after, and 4 s before +2 s after). This facilitated the acquisition of three sets of lane-changing segments, allowing for the assessment of which window setting yielded the most effective detection of abnormal lane-changing behavior.
Step 4: Removal of Incomplete Segments. For each lane-changing trajectory, if the length before or after timestamp tLC does not meet the requirements of the specified time window, that lane-changing segment is deleted. This ensures uniformity in the length of extracted lane-changing segments (6 s) within each set, with consistent lengths before and after the lane change.
Step 5: Computation of Feature Parameters. Using the candidate feature parameters provided in Table 2, feature parameters are computed for each lane-changing segment.
Step 6: Labeling Abnormal Trajectories. In this study, based on the following criteria, three types of lane-changing behaviors are classified as abnormal lane changes:
(1) “Cutting off behavior” (COB): In this study, the term “Cutting off behavior” is defined as a situation where a vehicle changes lanes with insufficient distance from the trailing vehicle in the target lane, leading to interference with the trailing vehicle’s movement. Figure 5 illustrates this concept, with a schematic on the left and an actual case on the right. The white vehicle within the red box changes lanes in close proximity to the red car, resulting in an emergency brake by the red car and a brief congestion behind it.
(2) “Tailgating behavior” (TB): In this study, “Tailgating behavior” is defined as a scenario where a vehicle in the lane-changing process maintains a close distance to the leading vehicle in the current lane. Figure 6 demonstrates this, with the white vehicle within the red box initiating and completing a lane change in close proximity to the red car in the adjacent lane. This behavior reduces the driver’s reaction time in emergency situations and increases the risk of collisions.
(3) “Cutting in behavior” (CIB): In this study, “Cutting in behavior” is defined as a situation where a vehicle changes lanes with insufficient distance from the leading vehicle in the target lane. Figure 7 provides a visual representation, showing a white vehicle within the red box forcefully completing a lane change without allowing adequate safety distance for the leading and trailing vehicles in the target lane, resulting in driving risks.
The annotation process was completed using a combination of manual labeling and trajectory data screening. The purpose of data screening was to extract potential segments of abnormal lane-changing behavior, thereby reducing the labor intensity of manual verification. The screening criteria were as follows:
(1) For COB behavior, it was required that there be a preceding vehicle in the target lane behind the subject vehicle. For TB behavior, the presence of a preceding vehicle in front was required. For CIB behavior, the presence of a vehicle in the front of the target lane was necessary.
(2) Surrounding vehicles were defined as those with a longitudinal distance less than 75 m from the subject vehicle [57].
(3) Both the subject vehicle and surrounding vehicles were required to have speeds greater than 1 m/s, ensuring that both vehicles were always in motion.
To enhance the accuracy of the labeling results, each video was processed by two individuals simultaneously. In cases of discordant opinions in the labeling results, the workers engaged in discussions to reach a consensus. Furthermore, the manual verification process adhered to the principle of accepting a high abnormality rate (above 10%) to ensure that valid events were not overlooked.
Due to the use of three different window settings, the number of segments varied under each setting. For instance, if a lane change occurred in the initial phase of a trajectory, resulting in a window before the lane change of less than 3 s, it would not be included in the “3 s before LC + 3 s after LC” segment set. Accordingly, the number of extracted lane-changing events for the three different window settings were 9313, 9102, and 9146. These events were then subjected to feature parameter computation and abnormal lane-changing behavior detection.
Each lane-changing segment could exhibit characteristics of more than one type of abnormal lane-changing behavior. As a result, the total count of normal and abnormal events exceeded the total count of lane-changing segments. Based on the different window settings, three distinct sets of segments were ultimately formed, named as Window Settings 1, Window Settings 2, and Window Settings 3. Table 3 summarizes the statistical results for these three datasets.

2.5. Abnormal Behavior Detection

In this study, the Light Gradient Boosting Machine (LGBM) algorithm was employed for detecting abnormal lane-changing behavior. Proposed by Microsoft in 2017, LGBM is a gradient boosting algorithm characterized by low memory usage, high efficiency, and strong interpretability [58]. The fundamental concept of the LGBM algorithm involves discretizing continuous floating-point feature values into k integers. It then constructs a histogram of width k, accumulating statistics based on the discretized values in the histogram. Ultimately, it identifies the optimal splitting point. Additionally, LGBM employs a leaf-wise growth strategy with depth limitations, replacing the traditional level-wise decision tree growth strategy. This enhancement improves accuracy while mitigating the risk of overfitting. The specific steps are outlined as follows:
Step 1: Train–Test Split. The samples were partitioned in a 7:3 ratio for training and testing, respectively. Seventy percent of the dataset was utilized for model training, followed by model evaluation on the remaining 30%.
Step 2: Normalization Process. Given that the units of the trajectory feature parameters used in this study varied, a min–max standardization method was employed to normalize the sample data, thereby eliminating disparities among features:
x = ( x x min ) / ( x max x min )
where x′ represents the normalized value, x denotes the original sample value, xmin stands for the minimum sample value, and xmax signifies the maximum sample value.
Step 3: Imbalanced Data Resampling. The imbalanced distribution of samples across different categories of lane-changing behaviors can impact the performance of the detection model. To mitigate this, the Synthetic Minority Oversampling Technique (SMOTE) method was applied to oversample the small sample data related to abnormal behaviors, thus rectifying the issue of sample imbalance.
Step 4: Hyperparameter Tuning. The Differential Evolution Algorithm (DE) was employed for fine-tuning the model to obtain the optimal hyperparameters for the LGBM model. These optimal hyperparameters were subsequently utilized for retraining the training dataset. Table 4 shows the results of LGBM model structure optimization based on the differential evolution algorithm.
Step 5: Abnormal Behavior Detection. The retrained LGBM model was applied to the test set to detect abnormal lane-changing behavior. Based on the test set’s results, a confusion matrix was obtained.
Step 6: Model Performance Evaluation. The model’s performance was evaluated using four metrics: precision, recall, accuracy, and F1 score.

3. Results and Discussion

3.1. Results of Abnormal LC Recognition

3.1.1. Comparison of Model Performance with Three Different Window Settings

Modeling was conducted using the three sets of segments outlined in Section 2.4. These sets were partitioned into training and testing data in a 7:3 ratio. This resulted in respective test set sizes of 2794, 2731, and 2744. The LGBM model was implemented using the scikit-learn and LGBM libraries in Python 3.11.
Since each lane-changing segment could potentially involve multiple types of abnormal lane-changing behavior, the model established in this study conducted separate detection for the three types of abnormal lane-changing behavior. Specifically, it produced “yes” or “no” outcomes for each type.
Figure 8 illustrates the confusion matrices for the detection of abnormal lane-changing behaviors. Figure 8a–c represents the results of detecting COB, TB, and CIB using Window Settings 1 for training and testing. Figure 8d–i depicts the results for abnormal lane-changing behavior detection using Window Settings 2 and Window Settings 3, respectively. Larger values in the top-left corner of each confusion matrix indicate better performance in detecting “normal” behavior. Conversely, larger values in the bottom-right corner indicate superior performance in detecting “abnormal” behavior. Specifically, the top-left corner of each confusion matrix represents “True positive”, indicating the number of samples with “normal” behavior detected correctly. As shown in Figure 8, the True positive values for Window Settings 1 are all above 2600, with the highest for CIB behavior at 2695 and the lowest for TB behavior at 2619. The True positive values for Window Settings 2 and Window Settings 3 are lower than those under Window Settings 1 for the same behavior type, reflecting the superior detection performance of Window Settings 1 in normal behavior. The True positive values for COB and TB behaviors are both below 2600, and for CIB behaviors, they are 2613 and 2649, respectively. The bottom-right corner of the confusion matrix represents “True negative”, indicating the number of samples with “abnormal” behavior detected correctly. For COB behavior, Window Settings 2 has the highest True negative values, followed by Window Settings 1. For TB behavior, Window Settings 3 has the highest True negative values, followed by Window Settings 1. For CIB behavior, Window Settings 2 has the highest True negative values, followed by Window Settings 1. It can be seen that in terms of abnormal behavior detection, Window Settings 1 also shows good detection stability. In conclusion, since the True positive and True negative values under three window settings are relatively high, it indicates that the LGBM model used in this study performs well in the detection of abnormal lane-changing behaviors.
Quantitative analysis of the data in Table 5 reveals the following:
(1) Under all three window settings, the model’s accuracy is consistently above 0.9, and there is minimal variation in the detection results for each abnormal behavior across the three window settings, indicating good overall detection performance.
(2) In Window Settings 2, the precision for CIB detection is 0.826, lower than the 0.896 in Window Settings 1 and significantly lower than the 0.923 in Window Settings 1. In Window Settings 3, the precision for COB detection is 0.732, much lower than the 0.866 in Window Settings 1 and the 0.871 in Window Settings 2. This suggests that Window Settings 2 and Window Settings 3 exhibit shortcomings in the detection performance of CIB and COB behaviors, while Window Settings 1 demonstrates better detection stability.
(3) Under all three window settings, the model’s recall is consistently above 0.9, except for Window Settings 3, where the recall for COB behavior is lower than the other two settings. For the other two behavior types, there is minimal variation in the detection results across the three window settings.
(4) Window Settings 1 achieves F1 scores above 0.9 for all three abnormal behaviors, while Window Settings 2 shows F1 scores below 0.9 for CIB behavior, and Window Settings 3 exhibits F1 scores below 0.9 for COB behavior. Since F1 score best reflects the model’s comprehensive performance, Window Settings 1 demonstrates the best overall performance.
In summary, the models under all three window settings exhibit good detection levels, but Window Settings 1 show better detection stability and overall performance. Therefore, this study selects Window Settings 1, i.e., the “3 s before LC + 3 s after LC” window setting, as a representative choice for future investigations.

3.1.2. Performance Comparison between LGBM Model and Three Other Typical Models

To further validate the performance of the proposed LGBM model, three widely recognized and commonly used methods were employed for comparison using the same segment dataset (Window Settings 1) and features. Table 6 presents the comparative results of the LGBM model with SVM, KNN, and XGBoost models in terms of two metrics. The following is shown in the table:
For the detection of “COB” behavior, the LGBM model exhibits a precision of 0.866 and an F1 score of 0.916, slightly lower than the XGBoost model’s 0.892 and 0.920, respectively, but higher than the SVM model and significantly higher than the KNN model.
For “TB” behavior detection, the LGBM model shows a precision of 0.865 and an F1 score of 0.916, slightly higher than the XGBoost model’s 0.858 and 0.899, and significantly higher than the SVM and KNN models.
In the case of “CIB” behavior detection, the LGBM model achieves a precision of 0.892 and an F1 score of 0.930, significantly higher than the XGBoost model’s 0.865 and 0.892, higher than the SVM model, and significantly higher than the KNN model.
Additionally, in terms of model training time, both the LGBM and XGBoost models demonstrate significantly shorter training times compared to the SVM and KNN models. Specifically, the XGBoost model has the shortest training time at 3.83 s, while the LGBM model’s training time is 4.07 s, 1.06 times that of the XGBoost model.
In summary, the following conclusions can be drawn: (a) The detection accuracy of the LGBM model is higher than that of the SVM model and significantly surpasses that of the KNN model; (b) The LGBM model exhibits slightly lower detection accuracy for “COB”, higher detection accuracy for “TB”, and significantly higher detection accuracy for “CIB” compared to the XGBoost model; (c) Both the LGBM and XGBoost models demonstrate relatively short training times, with the LGBM model’s training time slightly higher than the XGBoost model but significantly lower than the SVM and KNN models; (d) Considering the detection results of all four models, the LGBM model shows the optimal performance for detecting abnormal lane-changing behavior.

3.2. SHAP-Based Model Interpretation

The SHAP algorithm, proposed by Lundberg and Lee in 2017, can be used to understand the underlying mechanisms of machine learning models [59]. The core concept of SHAP values is to calculate the mean marginal contribution of a feature to the model’s prediction, thereby providing explanation and quantification of the model. Figure 9 displays the overall importance of each feature. Figure 9a–c show the top ten features contributing to the detection of the three types of COB, TB, and CIB abnormal behaviors, respectively. The horizontal axis represents the average SHAP value of the feature, while the vertical axis represents different features.
For COB behavior, the features with the highest contributions are the space headway between the subject vehicle and the trailing vehicle in the target lane after the lane change (d3a), the relative speed between the subject vehicle and the preceding vehicle in the target lane before the lane change (vr2b), and the relative speed between the subject vehicle and the trailing vehicle in the target lane after the lane change (vr3a). For TB behavior, the features with the highest contributions are the relative speed between the subject vehicle and the preceding vehicle in the same lane before the lane change (vr1b), the total longitudinal speed of the subject vehicle before the lane change (vb), and the space headway between the subject vehicle and the preceding vehicle in the same lane before the lane change (d1b). For CIB behavior, the features with the highest contributions are the relative speed between the subject vehicle and the preceding vehicle in the target lane after the lane change (vr2a), the relative speed between the subject vehicle and the preceding vehicle in the target lane before the lane change (vr2b), and the space headway between the subject vehicle and the preceding vehicle in the target lane after the lane change (d2a).
For the detection of each type of abnormal behavior, the features with the highest contributions follow these patterns: features with high contributions are mainly multi-vehicle variables, with the space headway and relative speed between vehicles being the most prominent; compared to entire lane change variables, before-lane-change variables and after-lane-change variables contribute significantly more; compared to other statistical values, the contributions of feature extremes (max and min values) are higher.
In summary, the following conclusions can be drawn:
(1) Features considering the relationships between vehicles, i.e., multi-vehicle variables, contribute far more to the detection of abnormal lane-changing behavior compared to single-vehicle variables. Multi-vehicle variables are calculated based on the proposed trajectory representation method that integrates surrounding vehicle information, underscoring the method’s importance in detecting abnormal lane-changing behavior;
(2) The distinct features before and after lane changes proposed in this study play a particularly significant role in the model;
(3) The contributions of space headway and relative speed extremes between vehicles are higher compared to other features.
Figure 10 displays the specific impact of each sample feature on the detection model results within the SHAP interpretation framework. Figure 10a–c show the results for the three types of COB, TB, and CIB abnormal lane-changing behaviors, respectively. The vertical axis represents different features, while the horizontal axis represents the SHAP value of each sample feature. The depth of color corresponds to the color bar on the right, reflecting the magnitude of the sample feature value. For example, in Figure 10a, for the first three features, when the individual feature values are small, they exhibit large SHAP values, indicating a negative correlation. In Figure 10b, the second feature, maximum value of vb, shows the opposite pattern: when the individual feature value is large, the SHAP value is also large, indicating a positive correlation.
Taking vr1b and vb, which contribute the most to TB, as an example, Figure 11 further illustrates the mapping relationship of individual feature variables with SHAP values. Figure 11a,b represent the minimum value of vr1b and maximum value of vb features, respectively. The horizontal axis represents the size of the feature variable’s value, while the vertical axis represents the SHAP value. The red dots in the figure indicate a significant contribution to the detection of abnormal behavior, while the blue dots indicate a significant contribution to the detection of normal behavior. When the minimum value of vr1b is greater than −3, the SHAP value is small, but when it is greater than −3, the contribution to both abnormal and normal behavior detection is significant. Conversely, for the maximum value of vb, the SHAP value is large when the feature value is greater than 12, and small when it is less than 12.

3.3. Discussion

In this paper, we innovatively defined three typical abnormal lane-changing behaviors: “Cutting off Behavior”, “Tailgating Behavior”, and “Cutting in Behavior”. While previous studies on abnormal driving behaviors have made significant progress, particularly in the use of driving simulators and naturalistic driving approaches, they have primarily focused on driver-operated behaviors, such as those induced by drivers’ psychology and style, for instance, “Talking on a cell phone while driving”, “Sounding the horn to express anger”, and the like [60]. Such behaviors are challenging for roadside cameras to capture, making them less applicable to road safety supervision. Guo et al. collect driving behavior data, extracting dangerous behaviors like “Sharp acceleration” and “Sharp left turn”, which were then applied to collision prediction [61]. However, the defined risky behaviors were limited to individual vehicles and did not consider interactions between vehicles. Wang et al. took into account the relationships between the changing vehicle and surrounding vehicles, extracting “Cutting in Behavior” for abnormal driving behavior [57]. Building upon this foundation, our study further defined three typical abnormal lane-changing behaviors.
Simultaneously, concerning the characterization of features related to abnormal lane-changing behavior, this paper also considers multi-vehicle factors, proposing “Single-vehicle variables” and “Multi-vehicle variables”. Additionally, the study accounts for different stages of lane-changing, introducing “Entire lane change variables”, “Before-lane change variables”, and “After-lane change variables”. Typically, parameters characterizing dangerous or abnormal driving behaviors focus on individual vehicle speed and acceleration, particularly acceleration [62]. Previous research has also suggested considering feature variables that capture the interactions between vehicles. For example, Xue et al. considered the gap and speed differences between vehicles [56], and Chen et al. expanded this to include acceleration differences between vehicles [63]. What sets this study apart is the use of “Multi-vehicle variables”, calculated based on the trajectory representation method proposed in this paper. Furthermore, the introduced features accounting for the before- and after-lane-change process, namely “Before-lane change variables” and “After-lane change variables”, make a significant contribution to the recognition of abnormal lane-changing behavior, playing a crucial role in the detection model presented in this paper.
In terms of detection models, machine learning models are crucial for conducting research on driving behaviors. Chen et al. employed the SVM method to identify hazardous driving behaviors such as Abrupt Double Lane Change (ALC), Retrograde Driving (RD), and Illegal U-Turn (IT), applying it to a video surveillance system [30]. Du et al. established a model using the KNN method and completed the lane change detection task using the NGSIM dataset [64]. Zhang et al. introduced a lane change prediction framework using the XGBoost model for lane-change behavior prediction [65]. In contrast, the LGBM model adopted in this study, proposed in recent years, has been demonstrated to exhibit high accuracy in the research on abnormal lane-changing behaviors presented in this paper.
The detection of abnormal driving is a significant contribution to the development of intelligent transportation. From the perspective of the driver, almost all vehicles collect speed data, and ADAS systems also issue warnings when the distance between two vehicles is too close. However, the results obtained in this study indicate that the relative speeds and distances between the front and rear vehicles in the same lane as well as the target lane can effectively characterize different forms of abnormal lane-changing behavior during lane changes. Therefore, the trajectory representation method and feature metrics proposed in this study provide a new approach for detecting abnormal driving behavior, which can be applied to autonomous driving.
Detecting abnormal driving behavior contributes significantly to the development of intelligent transportation. Therefore, from an applied perspective, this study can serve the domains of automated driving and traffic regulation in intelligent traffic environments. On one hand, from the driver’s perspective, nearly all vehicles collect speed data, and Advanced Driver Assistance Systems (ADAS) issue warnings when the following distance is too short. However, the results obtained in this study indicate that during lane-changing maneuvers, the relative speed and distance between the ego vehicle, leading vehicle in the current lane, and surrounding vehicles in the target lane can effectively characterize different forms of abnormal lane-changing behavior. Thus, the proposed trajectory representation method and associated features offer a new approach for abnormal driving behavior detection applicable to automated driving.
On the other hand, from a traffic regulation standpoint, this study focuses on lane-changing behavior, identifying three typical abnormal lane-changing behaviors through the detection model rather than merely distinguishing between normal and abnormal driving behavior. Furthermore, this research extends beyond significant violations or lapses, such as disregarding the speed limit [60], and starts from driving performance, identifying abnormal lane-changing behavior at the level of driving trajectories. This provides a new avenue for traffic authorities to detect and regulate dangerous driving behavior through roadside devices. Additionally, since this study is based on vehicle trajectory data, its applicability is not limited to roadside video data. Various methods for obtaining vehicle trajectories, including laser radar, millimeter-wave radar, unmanned aerial vehicle aerial photography, and sensor integration technologies, can all support the conduct and application of this research.

4. Conclusions

This paper proposes a trajectory representation method that integrates information about the positions of surrounding vehicles, diverging from the conventional trajectory representation based on an individual vehicle’s positional-time series; this research innovatively incorporates the relative positions between vehicles into the trajectory representation. Subsequently, this method is applied to roadside video data from expressways in Wuhan City. Utilizing the constructed driving trajectories, we introduce feature variables that take into account the interactions between vehicles and differentiate between before- and after-lane-change states. We then employ the LGBM model algorithm to detect abnormal lane-changing behavior and use the SHAP algorithm to interpret the influence of feature variables on the detection results. The results indicate that the LGBM model used in this paper outperforms other models in terms of detection performance, and the trajectory representation method and feature parameters proposed in this study make a significant contribution to the detection results.
In future work, we plan to conduct tests in different geographical locations using various sensor technologies to further enrich this study, thereby comprehensively improving the detection performance.

Author Contributions

The study was conceptualized by N.L. and D.X., while M.L. and X.Y. were responsible for project administration and data processing. D.X. managed algorithm design, model construction, and drafted the manuscript. All authors actively participated in discussions, contributed to the review, and made revisions to the article. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Open Project of Shandong Key Laboratory of Smart Transportation (Preparation) (2021SDKLST008) and the National Natural Science Foundation of China (No. 52072290).

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors thank Wuhan University of Technology for support. Chunqing Liu and Congcheng Chai executed the experiments and performed data preprocessing.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ali, Y.; Zheng, Z.; Bliemer, M.C.J. Calibrating lane-changing models: Two data-related issues and a general method to extract appropriate data. Transp. Res. Part C Emerg. Technol. 2023, 152, 104182. [Google Scholar] [CrossRef]
  2. Antin, J.F.; Wotring, B.; Perez, M.A.; Glaser, D. Investigating lane change behaviors and difficulties for senior drivers using naturalistic driving data. J. Saf. Res. 2020, 74, 81–87. [Google Scholar] [CrossRef] [PubMed]
  3. Guo, H.Y.; Xie, K.; Keyvan-Ekbatani, M. Modeling driver’s evasive behavior during safety–critical lane changes:Two-dimensional time-to-collision and deep reinforcement learning. Accid. Anal. Prev. 2023, 186, 107063. [Google Scholar] [CrossRef] [PubMed]
  4. Muslim, H. Design and Evaluation of Lane-Change Collision Avoidance Systems in Semi-Automated Driving. IEEE Trans. Veh. Technol. 2023, 72, 7082–7094. [Google Scholar] [CrossRef]
  5. Wu, H.; Wu, C.; Lyu, N.; Li, J. Does a faster takeover necessarily mean it is better? A study on the influence of urgency and takeover-request lead time on takeover performance and safety. Accid. Anal. Prev. 2022, 171, 106647. [Google Scholar] [CrossRef] [PubMed]
  6. Chen, D.; Ahn, S. Capacity-drop at extended bottlenecks: Merge, diverge, and weave. Transp. Res. Part B Methodol. 2018, 108, 1–20. [Google Scholar] [CrossRef]
  7. Atagoziev, M.; Schmidt, E.G.; Schmidt, K.W. Lane change scheduling for connected and autonomous vehicles. Transp. Res. Part C Emerg. Technol. 2023, 147, 103985. [Google Scholar] [CrossRef]
  8. Zheng, Z.D. Recent developments and research needs in modeling lane changing. Transp. Res. Part B Methodol. 2014, 60, 16–32. [Google Scholar] [CrossRef]
  9. Wang, X.S.; Xu, X.Y. Assessing the relationship between self-reported driving behaviors and driver risk using a naturalistic driving study. Accid. Anal. Prev. 2019, 128, 8–16. [Google Scholar] [CrossRef]
  10. Fei, G.; Zhang, X.; Yang, Y.; Yao, H.; Yang, J.; Li, X.; Gao, L.; Zhou, Y.; Ming, W.; Stallones, L.; et al. Driving anger among motor vehicle drivers in China: A cross-sectional survey. Traffic Inj. Prev. 2019, 20, 570–574. [Google Scholar] [CrossRef]
  11. Deffenbacher, J.L.; Oetting, E.R.; Lynch, R.S. Development of a drivinganger scale. Psychol. Rep. 1994, 74, 83–91. [Google Scholar] [CrossRef] [PubMed]
  12. Wang, X.S.; Xing, Y.L.; Luo, L.; Yu, R.J. Evaluating the effectiveness of Behavior-Based Safety education methods for commercial vehicle drivers. Accid. Anal. Prev. 2018, 117, 114–120. [Google Scholar] [CrossRef] [PubMed]
  13. Peng, Y.C.; Li, C.Y.; Wang, K.; Gao, Z.; Yu, R.J. Examining imbalanced classification algorithms in predicting real-time traffic crash risk. Accid. Anal. Prev. 2020, 144, 105610. [Google Scholar] [CrossRef] [PubMed]
  14. Martinez, C.M.; Heucke, M.; Wang, F.Y.; Gao, B.; Cao, D. Driving Style Recognition for Intelligent Vehicle Control and Advanced Driver Assistance: A Survey. IEEE Trans. Intell. Transp. Syst. 2018, 19, 666–676. [Google Scholar] [CrossRef]
  15. Bouhsissin, S.; Sael, N.; Benabbou, F. Driver Behavior Classification: A Systematic Literature Review. IEEE Access 2023, 11, 14128–14153. [Google Scholar] [CrossRef]
  16. Min, K.; Ando, A. Analysis on characteristics of dangerous driving events via recorded data of drive-recorder. Transp. Res. Procedia 2020, 48, 1342–1363. [Google Scholar] [CrossRef]
  17. Ali, Y.; Haque, M.M.; Zheng, Z.D.; Bliemer, M.C.J. Stop or go decisions at the onset of yellow light in a connected environment: A hybrid approach of decision tree and panel mixed logit model. Anal. Methods Accid. Res. 2021, 31, 100165. [Google Scholar] [CrossRef]
  18. Wang, W.S.; Xi, J.Q.; Chong, A.; Li, L. Driving Style Classification Using a Semisupervised Support Vector Machine. IEEE Trans. Hum.-Mach. Syst. 2017, 47, 650–660. [Google Scholar] [CrossRef]
  19. Bejani, M.M.; Ghatee, M. A context aware system for driving style evaluation by an ensemble learning on smartphone sensors data. Transp. Res. Part C Emerg. Technol. 2018, 89, 303–320. [Google Scholar] [CrossRef]
  20. Vignali, V.; Bichicchi, A.; Simone, A.; Lantieri, C.; Dondi, G.; Costa, M. Road sign vision and driver behaviour in work zones. Transp. Res. Part F Traffic Psychol. Behav. 2019, 60, 474–484. [Google Scholar] [CrossRef]
  21. Wang, R.; Xie, F.; Zhao, J.; Zhang, B.; Sun, R.; Yang, J. Smartphone Sensors-Based Abnormal Driving Behaviors Detection: Serial-Feature Network. IEEE Sens. J. 2021, 21, 15719–15728. [Google Scholar] [CrossRef]
  22. Wang, Q.L.; Liu, Y.J.; Liu, J.W.; Gu, Y.L.; Kamijo, S. Critical areas detection and vehicle speed estimation system towards intersection-related driving behavior analysis. In Proceedings of the 2018 IEEE International Conference on Consumer Electronics (ICCE), Hue Imperial City, Vietnam, 18–20 July 2018. [Google Scholar]
  23. Shahverdy, M.; Fathy, M.; Berangi, R.; Sabokrou, M. Driver behavior detection and classification using deep convolutional neural networks. Expert Syst. Appl. 2020, 149, 113240. [Google Scholar] [CrossRef]
  24. Pathivada, B.K.; Perumal, V. Modeling driver behavior in dilemma zone under mixed traffic conditions. Transp. Res. Procedia 2017, 27, 961–968. [Google Scholar] [CrossRef]
  25. Feng, Y.X.; Pickering, S.; Chappell, E.; Iravani, P.; Brace, C. Driving Style Analysis by Classifying Real-World Data with Support Vector Clustering. In Proceedings of the 2018 3rd IEEE International Conference on Intelligent Transportation Engineering (ICITE), Singapore, 3–5 September 2018. [Google Scholar] [CrossRef]
  26. Zhao, C.; Liao, F.X.; Li, X.H.; Du, Y.C. Macroscopic modeling and dynamic control of on-street cruising-for-parking of autonomous vehicles in a multi-region urban road network. Transp. Res. Part C Emerg. Technol. 2021, 128, 103176. [Google Scholar] [CrossRef]
  27. Sharma, A.; Zheng, Z.; Bhaskar, A. A pattern recognition algorithm for assessing trajectory completeness. Transp. Res. Part C Emerg. Technol. 2018, 96, 432–457. [Google Scholar] [CrossRef]
  28. Liu, Z.Q.; Weng, J.; Guo, J.J.; Ma, J.F.; Huang, F.; Sun, H.; Cheng, Y.D. PPTM: A Privacy-Preserving Trust Management Scheme for Emergency Message Dissemination in Space–Air–Ground-Integrated Vehicular Networks. IEEE Internet Things J. 2022, 9, 5943–5956. [Google Scholar] [CrossRef]
  29. Guo, J.J.; Liu, Z.Q.; Tian, S.Y.; Huang, F.; Li, J.X.; Li, X.H.; Igorevich, K.K.; Ma, J.F. TFL-DT: A Trust Evaluation Scheme for Federated Learning in Digital Twin for Mobile Networks. IEEE J. Sel. Areas Commun. 2023, 41, 3548–3560. [Google Scholar] [CrossRef]
  30. Chen, Z.J.; Wu, C.Z.; Huang, Z.; Lyu, N.C.; Hu, Z.Z.; Zhong, M.; Cheng, Y.; Ran, B. Dangerous driving behavior detection using video-extracted vehicle trajectory histograms. J. Intell. Transp. Syst. 2017, 21, 409–421. [Google Scholar] [CrossRef]
  31. Johnson, N.; Hogg, D. Learning the distribution of object trajectories for event recognition. Image Vis. Comput. 1996, 14, 609–615. [Google Scholar] [CrossRef]
  32. Ahmed, S.A.; Dogra, D.P.; Kar, S.; Roy, P.P. Trajectory-Based Surveillance Analysis: A Survey. IEEE Trans. Circuts Syst. Video Technol. 2019, 29, 7. [Google Scholar] [CrossRef]
  33. Goncalves, V.P.M.; Silva, L.P.; Nunes, F.L.S.; Ferreira, J.E.; Araujo, L.V. Concept drift adaptation in video surveillance: A systematic review. Multimed. Tools Appl. 2023, 15, 1–41. [Google Scholar] [CrossRef]
  34. Chen, Z.J.; Cai, H.; Zhang, Y.S.; Wu, C.Z.; Mu, M.C.; Li, Z.X.; Sotelo, M.A. A novel sparse representation model for pedestrian abnormal trajectory understanding. Expert Syst. Appl. 2019, 138, 112753. [Google Scholar] [CrossRef]
  35. Wang, S.Z.; Cao, J.N.; Yu, S. Deep Learning for Spatio-Temporal Data Mining: A Survey. IEEE Trans. Knowl. Data Eng. 2022, 34, 3681–3700. [Google Scholar] [CrossRef]
  36. Lv, J.M.; Li, Q.; Sun, Q.H. T-CONV: A Convolutional Neural Network for Multi-scale Taxi Trajectory Prediction. In Proceedings of the 2018 IEEE International Conference on Big Data and Smart Computing (BigComp), Shanghai, China, 15–17 January 2018. [Google Scholar] [CrossRef]
  37. Tang, J.J.; Liang, J.; Zhang, S.; Huang, H.L.; Liu, F. Inferring driving trajectories based on probabilistic model from large scale taxi GPS data. Physica A 2018, 506, 566–577. [Google Scholar] [CrossRef]
  38. Liang, Y.; Zhao, Z. NetTraj: A Network-Based Vehicle Trajectory Prediction Model with Directional Representation and Spatiotemporal Attention Mechanisms. IEEE Trans. Intell. Transp. Syst. 2022, 23, 14470–14481. [Google Scholar] [CrossRef]
  39. Kuo, Y.L.; Huang, X.; Barbu, A.; McGill, S.G.; Katz, B.; Leonard, J.J.; Rosman, G. Trajectory Prediction with Linguistic Representations. In Proceedings of the 2022 IEEE International Conference on Robotics and Automation (ICRA2022), Philadelphia, PA, USA, 23–27 May 2022. [Google Scholar] [CrossRef]
  40. Wang, W.; Xia, F.; Nie, H.S.; Chen, Z.K.; Gong, Z.G.; Kong, X.J.; Wei, W. Vehicle Trajectory Clustering Based on Dynamic Representation Learning of Internet of Vehicles. IEEE Trans. Intell. Transp. Syst. 2021, 22, 3567–3576. [Google Scholar] [CrossRef]
  41. Ahmed, I.; Karr, A.F.; Rouphail, N.M.; Chasea, R.T.; Tanvir, S. Characterizing lane changing behavior and identifying extreme lane changing traits. Transp. Lett. 2023, 15, 450–464. [Google Scholar] [CrossRef]
  42. Yao, D.; Zhang, C.; Zhu, Z.H.; Huang, J.H.; Bi, J.P. Trajectory clustering via deep representation learning. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017. [Google Scholar] [CrossRef]
  43. Wissing, C.; Nattermann, T.; Glander, K.H.; Hass, C.; Bertram, T. Lane Change Prediction by Combining Movement and Situation based Probabilities. In Proceedings of the 20th World Congress of the International-Federation-of-Automatic-Control (IFAC), Toulouse, France, 9–14 July 2017. [Google Scholar] [CrossRef]
  44. Wang, J.H.; Zhang, Z.; Lu, G.Q. A Bayesian inference based adaptive lane change prediction model. Transp. Res. Part C Emerg. Technol. 2021, 132, 103363. [Google Scholar] [CrossRef]
  45. Lu, C.; Wang, H.J.; Lv, C.; Gong, J.W.; Xi, J.Q.; Cao, D.P. Learning Driver-Specific Behavior for Overtaking: A Combined Learning Framework. IEEE Trans. Veh. Technol. 2018, 67, 6788–6802. [Google Scholar] [CrossRef]
  46. Yi, Y.T.; Lu, C.; Wang, B.Y.; Cheng, L.; Li, Z.R.; Gong, J.W. Fusion of Gaze and Scene Information for Driving Behaviour Recognition: A Graph-Neural-Network-Based Framework. IEEE Trans. Intell. Transp. Syst. 2023, 24, 8109–8120. [Google Scholar] [CrossRef]
  47. Xie, S.L.; Imani, M.; Dougherty, E.R.; Braga-Neto, U.M. Nonstationary Linear Discriminant Analysis. In Proceedings of the 2017 51st Asilomar Conference on Signals, Systems, and Computers, Pacific Grove, CA, USA, 29 October–1 November 2018. [Google Scholar]
  48. Xun, Y.J.; Guo, W.; Liu, J.J. G-DriverAUT: A Growable Driver Authentication Scheme Based on Incremental Learning. IEEE Trans. Veh. Technol. 2023, 72, 5759–5768. [Google Scholar] [CrossRef]
  49. Wei, W.Y.; Fu, X.S.; Zhong, S.Y.; Ge, H.C. Driver’s mental workload classification using physiological, traffic flow and environmental factors. Transp. Res. Part F Psychol. Behav. 2023, 94, 151–169. [Google Scholar] [CrossRef]
  50. Imani, M.; Ghoreishi, S.F.; Braga-Neto, U.M. Bayesian Control of Large MDPs with Unknown Dynamics in Data-Poor Environments. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montréal, QC, Canada, 2–8 December 2018. [Google Scholar]
  51. Imani, M.; Ghoreishi, S.F.; Allaire, D.; Braga-Neto, U.M. MFBO-SSM: Multi-Fidelity Bayesian Optimization for Fast Inference in State-Space Models. In Proceedings of the 33rd AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. [Google Scholar]
  52. Ustuner, M.; Balik Sanli, F. Polarimetric target decompositions and light gradient boosting machine for crop classification: A comparative evaluation. Int. J. Geo-Inf. 2019, 8, 76. [Google Scholar] [CrossRef]
  53. Song, H.M.; Woo, J.Y.; Kim, H.K. In-vehicle network intrusion detection using deep convolutional neural network. Veh. Commun. 2020, 21, 100198. [Google Scholar] [CrossRef]
  54. Cao, J.J.; Leng, W.C.; Liu, K.; Liu, L.; He, Z.; Zhu, Y.H. Object-Based Mangrove Species Classification Using Unmanned Aerial Vehicle Hyperspectral Images and Digital Surface Models. Remote Sens. 2018, 10, 89. [Google Scholar] [CrossRef]
  55. Zhang, Z.S.; Wang, S.; Lin, N.; Wang, Z.P.; Liu, P. State of Health Estimation of Lithium-Ion Batteries in Electric Vehicles Based on Regional Capacity and LGBM. Sustainability 2023, 15, 2052. [Google Scholar] [CrossRef]
  56. Xue, Q.W.; Xing, Y.Y.; Lu, J. An integrated lane change prediction model incorporating traffic context based on trajectory data. Transp. Res. Part C Emerg. Technol. 2022, 141, 103738. [Google Scholar] [CrossRef]
  57. Wang, X.S.; Yang, M.M.; Hurwitz, D. Analysis of cut-in behavior based on naturalistic driving data. Accid. Anal. Prev. 2019, 124, 127–137. [Google Scholar] [CrossRef]
  58. Shangguan, Q.Q.; Fu, T.; Wang, J.H.; Fang, S.E.; Fu, L.P. A proactive lane-changing risk prediction framework considering driving intention recognition and different lane-changing patterns. Accid. Anal. Prev. 2022, 164, 106500. [Google Scholar] [CrossRef]
  59. Bragança, H.; Colonna, J.G.; Oliveira, H.A.B.F.; Souto, E. How Validation Methodology Influences Human Activity Recognition Mobile Systems. Sensors 2022, 22, 2360. [Google Scholar] [CrossRef]
  60. Jiao, Y.; Wang, X.; Hurwitz, D.; Hu, G.; Xu, X.; Zhao, X. Revision of the driver behavior questionnaire for Chinese drivers’ aberrant driving behaviors using naturalistic driving data. Accid. Anal. Prev. 2023, 187, 107065. [Google Scholar] [CrossRef] [PubMed]
  61. Guo, M.; Zhao, X.; Yao, Y.; Yan, P.; Su, Y.; Bi, C.; Wu, D. A study of freeway crash risk prediction and interpretation based on risky driving behavior and traffic flow data. Accid. Anal. Prev. 2021, 160, 106328. [Google Scholar] [CrossRef] [PubMed]
  62. Zhu, Y.; Jiang, M.; Yamamoto, T.; Ding, N.; Shinkai, H.; Aoki, H.; Shimazaki, K. A Framework for Combining Lateral and Longitudinal Acceleration to Assess Driving Styles Using Unsupervised Approach. IEEE Trans. Intell. Transp. Syst. 2023, 1–19. [Google Scholar] [CrossRef]
  63. Chen, T.; Shi, X.P.; Wong, Y.D. Key feature selection and risk prediction for lane-changing behaviors based on vehicles’ trajectory data. Accid. Anal. Prev. 2019, 129, 156–169. [Google Scholar] [CrossRef]
  64. Du, L.Y.; Chen, W.; Pei, Z.H.; Zheng, H.J.; Fu, S.Z.; Chen, K.; Wu, D. Learning-based lane-change behaviour detection for intelligent and connected vehicles. Comput. Intell. Neurosci. 2020, 2020, 8848363. [Google Scholar] [CrossRef]
  65. Zhang, Y.; Shi, X.P.; Zhang, S.; Abraham, A. A XGBoost-Based Lane Change Prediction on Time Series Data Using Feature Engineering for Autopilot Vehicles. IEEE Trans. Intell. Transp. Syst. 2022, 23, 19187–19200. [Google Scholar] [CrossRef]
Figure 1. Comparison of two trajectory representation methods. (a) A trajectory representation method that only considers the target vehicle. (b) A trajectory representation method that considers the surrounding vehicles.
Figure 1. Comparison of two trajectory representation methods. (a) A trajectory representation method that only considers the target vehicle. (b) A trajectory representation method that considers the surrounding vehicles.
Sensors 23 09800 g001
Figure 2. Research framework.
Figure 2. Research framework.
Sensors 23 09800 g002
Figure 3. Vehicle detection at Point 3.
Figure 3. Vehicle detection at Point 3.
Sensors 23 09800 g003
Figure 4. Illustration of surrounding vehicles. (a) trajectory representation. (b) lane change feature extraction.
Figure 4. Illustration of surrounding vehicles. (a) trajectory representation. (b) lane change feature extraction.
Sensors 23 09800 g004
Figure 5. Schematic and example of “Cutting off behavior”.
Figure 5. Schematic and example of “Cutting off behavior”.
Sensors 23 09800 g005
Figure 6. Schematic and Example of “Tailgating behavior”.
Figure 6. Schematic and Example of “Tailgating behavior”.
Sensors 23 09800 g006
Figure 7. Schematic and Example of “Cutting in behavior”.
Figure 7. Schematic and Example of “Cutting in behavior”.
Sensors 23 09800 g007
Figure 8. Confusion matrix of abnormal LC behavior detection using LGBM model. (ac) represents the results of detecting COB, TB, and CIB using Window Settings 1, (df) represents the results of Window Settings 2, (gi) represents the results of Window Settings 3.
Figure 8. Confusion matrix of abnormal LC behavior detection using LGBM model. (ac) represents the results of detecting COB, TB, and CIB using Window Settings 1, (df) represents the results of Window Settings 2, (gi) represents the results of Window Settings 3.
Sensors 23 09800 g008
Figure 9. Overall importance of each feature variable. (a) shows the top ten features contributing to the detection of the three types of COB, (b) shows the top ten features contributing to the detection of the three types of TB, (c) shows the top ten features contributing to the detection of the three types of CIB.
Figure 9. Overall importance of each feature variable. (a) shows the top ten features contributing to the detection of the three types of COB, (b) shows the top ten features contributing to the detection of the three types of TB, (c) shows the top ten features contributing to the detection of the three types of CIB.
Sensors 23 09800 g009
Figure 10. Distribution of SHAP values for each feature variable. (a) shows the results for COB, (b) shows the results for TB, (c) shows the results for CIB.
Figure 10. Distribution of SHAP values for each feature variable. (a) shows the results for COB, (b) shows the results for TB, (c) shows the results for CIB.
Sensors 23 09800 g010aSensors 23 09800 g010b
Figure 11. SHAP value mapping of the minimum value of vr1b and maximum value of vb. (a) shows the results for vr1b, (b) shows the results for vb.
Figure 11. SHAP value mapping of the minimum value of vr1b and maximum value of vb. (a) shows the results for vr1b, (b) shows the results for vb.
Sensors 23 09800 g011
Table 1. Definition and description of surrounding vehicles for trajectory representation and lane change feature extraction.
Table 1. Definition and description of surrounding vehicles for trajectory representation and lane change feature extraction.
Trajectory Representation 1Lane Change Feature Extraction 2
Surrounding VehiclesDescriptionLeft LCRight LCDescription
Subject Vehicle,
SV
Subject vehicle in the sceneSVSVThe subject vehicle, which changes lanes.
Preceding Vehicle,
PV
Nearest forward vehicle in the same lane as the subject vehiclePVPVThe preceding vehicle of SV in the original lane.
Trailing Vehicle,
TV
Nearest behind vehicle in the same lane as the subject vehicleTVTVThe trailing vehicle of SV in the original lane.
Left Preceding Vehicle,
LPV
Nearest forward vehicle in the lane left of the subject vehiclePVTLPVSThe preceding vehicle of SV in the target lane when performing left lane change/the preceding vehicle of SV in the lane on the other side of the original lane when performing right lane change.
Left Trailing Vehicle,
LTV
Nearest behind vehicle in the lane left of the subject vehicleTVTLTVSThe trailing vehicle of SV in the target lane when performing left lane change/the trailing vehicle of SV in the lane on the other side of the original lane when performing right lane change.
Right Preceding Vehicle,
RPV
Nearest forward vehicle in the lane right of the subject vehiclePVSPVTLThe preceding vehicle of SV in the lane on the other side of the original lane when performing right lane change/the preceding vehicle of SV in the target lane when performing left lane change.
Right Trailing Vehicle,
RTV
Nearest behind vehicle in the lane right of the subject vehicleTVSTVTLThe trailing vehicle of SV in the lane on the other side of the original lane when performing right lane change/the trailing vehicle of SV in the target lane when performing left lane change.
1 is the definition and description of surrounding vehicles in general situations, used for trajectory representation. 2 is a special definition for the lane-change feature extraction process.
Table 2. Explanation of the features during the lane change.
Table 2. Explanation of the features during the lane change.
Entire lane change variables
vThe longitudinal velocity of SV during the lane change (m/s)
aThe longitudinal acceleration of SV during the lane change (m/s2)
vlatThe lateral velocity of SV during the lane change (m/s)
alatThe lateral acceleration of SV during the lane change (m/s2)
Before lane change: Single-vehicle variables
vbThe longitudinal velocity of SV during the lane change (m/s)
abThe longitudinal acceleration of SV during the lane change (m/s2)
vlatbThe lateral velocity of SV during the lane change (m/s)
alatbThe lateral acceleration of SV during the lane change (m/s2)
Before lane change: Multi-vehicle variables
d1bThe space headway between PV and SV (m)
d2bThe space headway between PVTL and SV (m)
d3bThe space headway between TVTL and SV (m)
d4bThe space headway between PVS and SV (m)
d5bThe space headway between TVS and SV (m)
vr1bThe relative speed of PV and SV (m/s)
vr2bThe relative speed of PVTL and SV (m/s)
vr3bThe relative speed of TVTL and SV (m/s)
vr4bThe relative speed of PVS and SV (m/s)
vr5The relative speed of TVS and SV (m/s)
After lane change: Single-vehicle variables
vaThe longitudinal velocity of SV during the lane change (m/s)
aaThe longitudinal acceleration of SV during the lane change (m/s2)
vlataThe lateral velocity of SV during the lane change (m/s)
alataThe lateral acceleration of SV during the lane change (m/s2)
After lane change: Multi-vehicle variables
d2aThe space headway between PVTL and SV (m)
d3aThe space headway between TVTL and SV (m)
vr2aThe relative speed of PVTL and SV (m/s)
vr3aThe relative speed of TVTL and SV (m/s)
1. For each characteristic indicator, calculate a set of statistical values consisting of mean, maximum, minimum, 0.25 quantiles, 0.5 quantiles (=median), and 0.75 quantiles. 2. Each lane change event is divided into two time periods: the before-lane-change period and the after-lane-change period.
Table 3. The statistics of LC segments.
Table 3. The statistics of LC segments.
Window Settings 1Window Settings 2Window Settings 3
3 s before LC + 3 s after LC2 s before LC + 4 s after LC4 s before LC + 2 s after LC
LC events931391029146
Normal834080858145
AbnormalCOB521569507
TB574513578
CIB337357303
Table 4. Optimal Results of LGBM Based on DE algorithm.
Table 4. Optimal Results of LGBM Based on DE algorithm.
HyperparametersRangeOptimization Results
Learning rate(0.01,0.1)0.059
Min data in leaf(10,100)67
Bagging fraction(0.5,1)0.965
Colsample bytree(0.5,1)0.983
Max depth(5,20)11
Lambda l1(0,0.5)0.0086
Lambda l2(0,0.5)0.3626
Min gain to split(0,0.5)0.2007
Scale pos weight(1,10)7
Table 5. Comparison of model performance with three different window settings.
Table 5. Comparison of model performance with three different window settings.
Window Settings 1Window Settings 2Window Settings 3
COBTBCIBCOBTBCIBCOBTBCIB
Precision0.8660.8650.8960.8710.9660.8260.7320.8900.923
Accuracy0.9910.9900.9950.9900.9960.9920.9820.9910.996
Recall0.9710.9740.9660.9740.9530.9690.9040.9760.955
F1 score0.9160.9160.9300.9200.9600.8920.8090.9310.939
Table 6. Performance comparison between LGBM model and three other typical models.
Table 6. Performance comparison between LGBM model and three other typical models.
COBTBCIBTraining Time/s
PrecisionF1 ScorePrecisionF1 ScorePrecisionF1 Score
LGBM0.8660.9160.8650.9160.8960.9304.07
SVM0.8330.8640.8240.8290.8080.84312.73
KNN0.7060.7930.7480.8060.6590.7339.48
XGBoost0.8920.9200.8580.8990.8650.8923.83
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, D.; Liu, M.; Yao, X.; Lyu, N. Integrating Surrounding Vehicle Information for Vehicle Trajectory Representation and Abnormal Lane-Change Behavior Detection. Sensors 2023, 23, 9800. https://doi.org/10.3390/s23249800

AMA Style

Xu D, Liu M, Yao X, Lyu N. Integrating Surrounding Vehicle Information for Vehicle Trajectory Representation and Abnormal Lane-Change Behavior Detection. Sensors. 2023; 23(24):9800. https://doi.org/10.3390/s23249800

Chicago/Turabian Style

Xu, Da, Mengfei Liu, Xinpeng Yao, and Nengchao Lyu. 2023. "Integrating Surrounding Vehicle Information for Vehicle Trajectory Representation and Abnormal Lane-Change Behavior Detection" Sensors 23, no. 24: 9800. https://doi.org/10.3390/s23249800

APA Style

Xu, D., Liu, M., Yao, X., & Lyu, N. (2023). Integrating Surrounding Vehicle Information for Vehicle Trajectory Representation and Abnormal Lane-Change Behavior Detection. Sensors, 23(24), 9800. https://doi.org/10.3390/s23249800

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop