Next Article in Journal
Granular Media Friction Pad for Robot Shoes—Hexagon Patterning Enhances Friction on Wet Surfaces
Next Article in Special Issue
On the Design of a Decision Support System for Robotic Equipment Adoption in Construction Processes
Previous Article in Journal
Ensemble Learning for Threat Classification in Network Intrusion Detection on a Security Monitoring System for Renewable Energy
Previous Article in Special Issue
Long Short-Term Memory Network-Based Metaheuristic for Effective Electric Energy Consumption Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real Driving Cycle-Based State of Charge Prediction for EV Batteries Using Deep Learning Methods

Department of Computer Science, Hanyang University, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(23), 11285; https://doi.org/10.3390/app112311285
Submission received: 23 October 2021 / Revised: 20 November 2021 / Accepted: 22 November 2021 / Published: 29 November 2021
(This article belongs to the Special Issue Artificial Intelligence Applications and Innovation)

Abstract

:
An accurate prediction of the State of Charge (SOC) of an Electric Vehicle (EV) battery is important when determining the driving range of an EV. However, the majority of the studies in this field have either been focused on the standard driving cycle (SDC) or the internal parameters of the battery itself to predict the SOC results. Due to the significant difference between the real driving cycle (RDC) and SDC, a proper method of predicting the SOC results with RDCs is required. In this paper, RDCs and deep learning methods are used to accurately estimate the SOC of an EV battery. RDC data for an actual driving route have been directly collected by an On-Board Diagnostics (OBD)-II dongle connected to the author’s vehicle. The Global Positioning System (GPS) data of the traffic lights en route are used to segment each instance of the driving cycles where the Dynamic Time Warping (DTW) algorithm is adopted, to obtain the most similar patterns among the driving cycles. Finally, the acceleration values are predicted from deep learning models, and the SOC trajectory for the next trip will be obtained by a Functional Mock-Up Interface (FMI)-based EV simulation environment where the predicted accelerations are fed into the simulation model by each time step. As a result of the experiments, it was confirmed that the Temporal Attention Long–Short-Term Memory (TA-LSTM) model predicts the SOC more accurately than others.

1. Introduction

Since Electric Vehicles (EVs) do not emit CO 2 , they have a great potential to prevent air pollution [1]. In addition, it is attractive for drivers to use EVs, because the cost of charging the battery of an electric vehicle is much cheaper than the cost of refueling [2]. In recent years, the performance of electric vehicle batteries has greatly improved and, consequently, there are EVs capable of driving more than 500 km when fully charged. However, there are few battery-charging stations compared to conventional gas stations, and batteries take a long time to charge [3]. Therefore, when driving an electric vehicle, it is very important to predict the driving distance through the state-of-charge (SOC) of the battery.
Accurate battery SOC measurement is an important feature in the BMS of electric vehicles, which can be implemented with microcontrollers (MCUs). Recently, as pre-trained neural network models have been installed and used in automotive MCUs, it has become possible to apply AI to the Battery Management System (BMS) [4]. In addition, among recent studies, a method to more accurately measure SOC using a cloud data center that provides high-capacity and high-performance calculations for big-data-based, data-driven Deep Learning (DL), and interworking with the vehicle’s BMS, was also proposed [5].
Future applications for fully Connected and Autonomous Vehicles (CAVs) include an AR-based navigation system, a video-conferencing application, real-time image-processing and inferencing solutions implemented by a neural network model equipped on board [6,7]. Since these applications require a high level of computation power and low power consumption, various hardware implementations such as a Graphics Processing Unit (GPU), field-programmable gate array (FPGA), and Application-Specific Integrated Circuits (ASIC) are employed. Specifically, a Tesla Full Self-Driving (FSD) computer, built on ASIC, meets 50 Tera floating point operations per second (TFLOPS) of the Neural Network model, and uses 100 W or less power consumption during computation [8]. To benefit from these state-of-the-art embedded systems adapted to the full CAVs, it is important to both estimate the current SOC and predict the future SOC to aid the drivers in decision-making.
It is important to accurately measure the current SOC, but it is also necessary to predict the future SOC based on the current driving data. An accurate prediction of the SOC can help determine the possible driving distance and charging time. There are two main methods to predict the SOC of an EV battery. The first one is predicting the SOC through simulation based on the vehicle dynamics. This approach was implemented by [9], which used standard driving cycles (SDC) as input of the simulation model. The second method consists of predicting the SOC through Machine Learning (ML) techniques using the battery data collected while driving the electric vehicle [10]; for example, using information from the battery itself (voltage, current, temperature) and ML algorithms to predict the future SOC.
Recently, there have been studies to predict SOC through complex mathematical formulas, considering the actual driving cycle of electric vehicles, but ML was not used.
In this paper, we propose an algorithm that accurately predicts the SOC of an EV battery by acquiring actual driving information through OBD-II, dividing and preprocessing this driving information by section based on GPS information and DTW, training a neural network with the dataset and running simulations using Functional Mock-Up Units (FMU).

2. Related Work

This section classifies the related studies into two major categories: modeling and simulation, and SOC estimation and prediction in the EV domain. Since there have been significant research efforts in these fields, we are focused on work that has been conducted rather recently, and discuss the limitations of these works.
First of all, in the field of modeling and simulation, extensive research has been carried out to improve the accuracy of EV energy consumption estimation with high-fidelity simulation models [11,12,13,14,15,16,17]. By using simulation environments such as MATLAB/Simulink and DACCOSIM, each physical component of an EV is mathematically modeled to form a subsystem block, which is composed of primitive blocks. The simulation is performed by the subsystem blocks mapping the input signals to the output values, which are passed to the subsequent blocks over time during the simulation. Recently, distributed co-simulation based on Functional Mock-up Interface (FMI), which provides a standardized interface to ease the sharing of different models [18], has been widely used. Techniques to reduce the simulation time while maintaining the simulation accuracy have also been developed. S. Hong proposed a redundancy reduction algorithm (RRA) to increase the simulation speed by adaptively adjusting the simulation step size, increasing it when a specific pattern of the cycle is detected as repeated, and decreasing it when a zero-crossing point is detected [19]. However, this is mainly applicable to the SDCs, especially the New European Driving Cycle (NEDC), where many repetitive parts are known ahead of the simulation.
These studies heavily rely on the SDCs to validate the methods. However, as mentioned above, SDCs have inherent limitations, as they are not real driving cycles (RDCs). Even though a variety of SDCs have been developed to emulate urban, rural, and highway environments, they do not reflect personalized factors, i.e., individual drivers’ habits, which are highly complicated as they dynamically change over time [20,21]. Moreover, they can hardly incorporate factors such as traffic conditions, traffic signals, and time of driving, which are unique to each driving instance [22]. Therefore, in order to provide more personalized simulation results, both the mathematically defined models and data-driven methods that learn the drivers’ patterns from the previous RDCs are required.
Next, there have been a myriad of studies in the area of SOC estimation and prediction. In the application of the pure EV, accurate prediction of the battery SOC consumption is highly important, since this would directly lead to an accurate prediction of the EV range [23]. According to the literature [24,25,26], methodological approaches to the SOC estimation can be divided into two: ML-based and non-ML-based methods. In brief, Coulomb counting and open-circuit voltage (OCV) methods have been used as conventional methods, while, recently, ML models consisting of deep neural networks (DNNs) have been intensely applied [24,25,26]. In the ML-based data-driven approach, the SOC values are predicted by automatically adjusting parameters to minimize the resultant error by learning from the pattern of the input features such as current, voltage, and temperature [27,28,29]. In this way, the models are fitted to the specific real-world measurement data, which enables predictions with enhanced accuracy under similar conditions.
However, few studies have been conducted on the prediction of future SOC trends through simulation. Most of the research has been focused on the internal parameters of the battery management system (BMS) [30] to estimate the SOC value. That is, research on the relationship between the EV subsystems in an integrated view to improve the performance of the simulation, as well as reducing the overhead of the tedious model-based simulation, has been insufficient. If an individual driver’s OBD data can be repeatedly collected on the road, one can simulate the SOC patterns for the next cycle for the same segment of the route with the velocity and acceleration data as inputs to an EV model. In this way, an approximate trajectory of the SOC consumption can be obtained, which can, in turn, help the driver be aware of the possible range and adjust her driving behavior from the perspective of an SOC.

3. Overall Procedure for Real Driving Cycle-Based SOC Estimation

This paper proposes a method to predict SOC using a real-world driving cycle. The proposed SOC prediction method using real-world driving cycle is as shown in the Figure 1.
The tested driving routes run from a fixed site to the other, i.e., from the homeplace of one of the authors to a parking lot in the Hanyang University campus. The route consists of several intersections between the start point and the destination point. The driving data were collected onboard while the author was driving, using an application named Torque Pro running on an Android device connected to a Bluetooth dongle, which supports the OBD-II specifications. The OBD data were comprised of various second-by-second temporal data, such as GPS latitude, longitude, velocities, etc., and were saved in a Comma-Separated-Value (CSV) format. The route is highlighted, as shown in Figure 2. The route was extracted and drawn on the map using the Google API with the GPS latitude and longitude information from the OBD data.
Since the overall driving cycles are highly dependent on the operations of en-route traffic lights, to enhance the accuracy of the predicted driving cycles, we first divided each of the end-to-end routes into sub-routes by the traffic lights (as can be seen from Figure 2 as circular markers). The segments obtained from different instances of the route were then compared to one another. To find the most similar intervals across all of the instances of the route, a Dynamic Time Warping (DTW) algorithm was employed. After grouping the segments of the route by their similarity, different Machine Learning (ML) algorithms were used to learn from them and predict future driving cycles.
To validate the performance of the prediction of the SoC consumption, we generated the eventual SoC values by an Electric Vehicle (EV) simulation model, implemented based on Functional Mock-up Units (FMU). The FMU-based EV simulation model reads as inputs the velocities, calculates the electric current from power consumption and regeneration by acceleration and deceleration, respectively, and writes as outputs the resultant SoC values to a file at each second. As expected, the test results showed that the SoC values could be predicted with high accuracy, as long as the driving cycle was predicted with high accuracy.
The paper is structured as follows: first, we proposed the preprocessing method using the DTW algorithm to group the driving intervals by similarity. Secondly, we proposed a deep learning algorithm that accurately predicts the speed after learning from similar driving cycles. The speed was also predicted using the existing Seasonal Autoregressive Integrated Moving Average (SARIMA) model. Finally, to evaluate the models, the predicted speeds by both algoritms were used as input to a simulation model, which computes the SOC of the driving cycle.

4. Methodology

4.1. Dynamic Time Warping (DTW) for Similarity of Real Driving Cycles

First proposed by [31] in the speech recognition field, the DTW is an algorithm that measures similarity between time series, which may vary in length. The technique performs non-linear warpings so that the time series are stretched or shrinked in order to find the optimal alignment between them. It has a wide range of applications, such as data mining, gesture recognition, robotics, speech-processing, manufacturing and medicine [32].
We propose a method to find similar driving routes based on the DTW algorithm, using the acceleration of the vehicle as input. The reason why acceleration is used is that the SOC is closely related to power consumption, which, in turn, is closely related to the acceleration of the vehicle. There were more than 30 driving-cycle-related data collected and, due to the large noise and uncertainty existent in the dataset, only similar segments were used for the prediction. Therefore, the goal of the algorithm is to find the most similar acceleration time series segments to use them as input to the ML models.
The algorithm that was used can be described in the following manner:
  • First of all, the base time series which we want to use to find other similar time series must be defined. For this present work, we defined the route shown in the map of the Figure 2, represented by the time series in the Figure 3, as the base route to which the remaining 30 time series will be compared.
  • Secondly, the base time series may further be split into an arbitrarily number of base segments. Considering the Figure 2 and Figure 3, the markers from A to F and the dotted black vertical lines represent situations in which the velocity of the car reached zero or, in other words, the start and the end points of each of the four segments (A to C, C to D, D to E, E to F). Note that the point B was not taken into account, because we regarded it as noise due to its proximity to point A, and that the each of the base segments are shown in greater detail in the subfigures of Figure 4.
  • Thirdly, for each of the four base segments, we set the lower bound and the upper bound times for the similar patter- finder algorithm by subtracting and adding an ϵ arbitrary small amount of time. For example, the second base segment was extracted from second 69 to 135 of the base time series. If ϵ = 50 , then the lower-bound and upper-bound times for a similar pattern-finder algorithm will be, respectively, 19 s and 185 s, as shown in Table 1.
  • With the lower and upper bounds of the algorithm defined, we split the period between the lower and upper bounds into smaller search periods, spaced by an arbitrary time interval. Intuitively, the lower the interval is, the higher the accuracy of the algorithm. These smaller search periods are shown in Table 1 for an interval of five seconds between each period.
  • Next, we looped through the search periods.
  • Inside the search-periods-loop, we loop through the remaining 30 time series, which will be compared to the base time series.
  • Next, we extract the corresponding search period of the time series to be used in the comparison.
  • Next, we obtain the distance between the extracted time series and the base segment by DTW.
  • Finally, after the two loops are finished, we rank the all the extracted time series according to their cost, computed by DTW, and obtain the 10 periods that have the lowest cost, which will form the training and test datasets for the machine learning models.
A summary of the algorithm is shown in Algorithm 1.
Algorithm 1 Similar time series finding algorithm
 1:
Set b a s e _ t i m e _ s e r i e s
 2:
Split b a s e _ t i m e _ s e r i e s into b a s e _ s e g m e n t s
 3:
for b s in b a s e _ s e g m e n t s  do
 4:
    Set l o w e r _ b o u n d and u p p e r _ b o u n d s
 5:
    Set i n t e r v a l
 6:
    Compute s e a r c h _ p e r i o d s
 7:
    Initialize empty list c o s t s
 8:
    for  p e r i o d in s e a r c h _ p e r i o d s  do
 9:
        for  t s in t i m e _ s e r i e s _ l i s t  do
10:
           Extract p e r i o d from t s
11:
            c o s t D T W   ( e x t r a c t e d _ t s _ p e r i o d , b s )
12:
           Append c o s t to c o s t s
13:
        end for
14:
    end for
15:
     g e t _ n _ s e r i e s _ l e a s t _ c o s t   ( c o s t s , n u m _ s e r i e s )
16:
end for

4.2. Deep Learning-Based SOC Prediction

In this section, our intention is to predict the future speed from the past speed of the EV; hence, we consider this task as a univariate time series prediction task. Given the observed time series data x = [ x 1 , x 2 , , x T ] R T , the task is to predict the future value x T + 1 R . Formally, we intend to predict x ^ T + 1 R through a function f, as follows:
x ^ T + 1 = f ( x 1 , x 2 , , x T ) ,
where f ( · ) is a linear or nonlinear function that needs to be learned.
Our main contribution is presenting a model for predicting speed, which is based on an LSTM with a temporal attention mechanism. The architecture of the prediction model is shown in Figure 5. First, we use an LSTM layer to encode the information from the input sequence into a feature representation. The final prediction is then made by utilizing the temporal attention mechanism over the output features of the LSTM layer.
Recurrent Neural Network (RNN) is one of the family of neural networks specialized in processing sequential data. If a general feed-forward neural network approach is used to process fixed-length data, the network has separate parameters for each input feature, so it needs to learn them separately at each time position. In contrast, an RNN shares weights across multiple time steps. This sharing of weights is important, as it allows for the generalization of unseen sequences and the sharing of statistical strength across time steps. Long short-term memory (LSTM) and gated recurrent unit (GRU), which are variants of RNN, are commonly used for handling sequential data such as language modeling [33,34] and time series prediction [35,36]. LSTM outperforms RNN and GRU for many sequence prediction tasks because of its gating systems, so we chose LSTM for this task. The forward process of an RNN cell is defined as:
f t = σ ( W f [ x t , h t 1 ] + b f )
i t = σ ( W i [ x t , h t 1 ] + b i )
c ˜ t = tanh ( W c [ x t , h t 1 ] + b c )
c t = f t · c t 1 + i t · c ˜ t
o t = σ ( W o [ x t , h t 1 ] + b o )
h t = o t · tanh ( c t )
where W f , W i , W c , W o R m × ( m + 1 ) are weight matrices, and b f , b i , b c , b o R m are bias to be learned. f t is the forget gate, which determines how much previous information is forgotten, i t is the input gate, which determines how much new information is added, and o t is the output gate, which, along with the current cell state c t , determines the output of this cell.
Recently, researchers have proposed multiple attention mechanisms for time series tasks and achieved better results than LSTM and GRU [37,38,39]. Inspired by the dynamic spatial–temporal attention mechanism [39], we employ the temporal attention mechanism to predict the acceleration of EVs in this work, since our task is univariate time series forecasting, which does not require consideration of spatial attention. Therefore, the temporal attention mechanism assigns a weight to each hidden state by correlating the output at each time step with the output at the last time step. Intuitively, for the prediction of the next time step, the current state is very important, but we cannot ignore the state of previous time steps. A general RNN-based prediction model either uses a fully connected layer to connect the hidden states of all time steps, giving a relatively accurate view of the hidden states of all time steps, or uses the last hidden state for prediction. However, we believe that temporal attention chooses the balanced option of adaptively attributing all time steps an attention score according to the relevance of different time steps to the state of the last time step, and finally connecting the fused context vector with the last hidden state for prediction, which, in turn, enhances the contribution of the current state to the prediction of the next time step.
First, the attention score is computed based on the relevance between the current state and the last state.
Attention Score ( h t , h T ) = h t T W s h T , 1 t T
where W s R m × m is learnable weights, which can be trained jointly with the LSTM layer to adaptively learn the correlation of hidden state of each time step with the last time step.
Then, for each time step, the attention score is converted into probabilistic form using the Softmax function, and the attention weight for this timepoint is obtained by
β t = exp ( Attention Score ( h t , h T ) ) j = 1 T exp ( Attention Score ( h j , h T ) )
where the attention weight β t R demonstrates the importance of hidden state h t for prediction and the Softmax function is applied to ensure all β sum to 1.
Next, the context vector is obtained by aggregating the hidden states of the RNN layer:
c = t = 1 T β t h t
where c is a weighted sum of all hidden states and can be considered as an adaptive selection of relevant hidden states among all time steps.
Finally, the prediction x ^ T + 1 can be obtained through the linear combination of context vector and the last hidden state:
x ^ T + 1 = W l T [ c , h T ] + b o
where W l R 2 m and b o R are learnable parameters.
The proposed model is differentiable, so the learnable parameters can be updated by back propagation. The mean squared error (MSE) is applied as loss function:
L o s s ( x T + 1 , x ^ T + 1 ) = 1 N i = 1 N ( x T + 1 i x ^ T + 1 i ) 2
where N is the number of samples.

5. Experimental Results

Before applying the proposed method to the RDCs, we checked the prediction results of an SDC using Seasonal Auto Regressive Integrated Moving Average (SARIMA), a conventional time-series forecasting model. This model is an extended version of ARIMA, supporting the modeling of the seasonal components of the time-series data. We used NEDC as the SDC and predicted the velocity after training the model. The mean absolute percentage error (MAPE) is used as criteria for evaluating the performance.
M A P E = 1 N t = 1 N x t x ^ t x t
where x t and x ^ t are ground truth and predicted value, respectively.
As can be seen from Table 2, the SDC can be predicted accurately enough without using the DL model. However, the accuracy for the RDC is much lower than the SDC. Therefore, we will use the proposed DL method to improve the predictions for the RDC.
To validate our proposed model using RDC, we collected real EV driving data through OBD II. Our dataset of EV speed was obtained from actual driving and divided into four road segments: location A to location C, location C to location D, location D to location E, and location E to location F. Each route has 11 driving cycles, and we took the seen driving cycles of each route as the training set and the rest as the test set. We first verified the performance of our speed predictions. by analyzing the performance of our model using the grid search method with different combinations of hyperparameters. The purpose was to find the best combination of hyperparameters and then verify the effect of each component on the results. Once we obtained the best model, we predicted the speed/acceleration along with the SOC values and compared them with some commonly used methods.

5.1. Speed/Acceleration Prediction

5.1.1. Performance with Different Hyperparameters

First, we integrated all the training sets in the dataset and adopted a five-fold cross validation (CV) method to find the best combination of hyperparameters. We implemented our model using the pytorch library and trained and tested it on an Nvidia GTX 1080 Ti GPU. Adam optimizer was used to train the models. The hyperparameters we need to consider here are the time step length T and the hidden state size m of the LSTM layer. For the proposed model, we set T as varying among [ 5 , 10 , 15 ] and m as varying among [ 16 , 32 , 64 , 128 ] . The average RMSE results for the 5-fold CV are tabulated in Table 3. First, we describe the effect of the time step length T. It is obvious from Table 3 that the average RMSE results for T = 10 and T = 15 are notably larger than for the case of T = 5 . Therefore, for speed prediction, the best choice of time step length is 5. Then, for the case T = 5 , the average RMSE value is the smallest when the hidden state size m equals to 32. In deep learning, too few parameters can lead to underfitting problems while too many parameters can lead to overfitting problems. Therefore, we believe that m = 16 belongs to underfitting, m = 64 and m = 128 belong to overfitting. Thus, the optimal combination of hyperparameters for our model for this dataset is { T = 5 , m = 32 } .

5.1.2. Performance with Different Models

First, to verify that the LSTM in TA-LSTM is the preferred encoding layer, we compared LSTM and other RNN variants on four road segments of the dataset. RNN and GRU are commonly used for time series forecasting, and all three models follow the five-fold CV approach introduced in the previous section to identify the best combination of hyperparameters. All three models obtained the minimum average RMSE at { T = 5 , m = 32 } . The first seven driving cycles of each road segment were used as the training set and the remaining four driving cycles were used as the test set, and each road segment was trained and tested separately. In addition, 20% of the data in the training set were randomly selected as the validation set during the training process, and the best-performing model in the validation set was saved for testing. The best test results for each model after training multiple times are shown in Table 4, as can be seen in Table 4, LSTM performs the best in all four road sections. Therefore, we choose LSTM as the coding layer in this task.
Then, to verify the effectiveness of temporal attention, we compared the LSTM with TA-LSTM. The training process and the allocation of training and test sets are the same as in the previous experiment. The results of the comparison are depicted in Figure 6. The comparison shows that the addition of temporal attention yields better results.
Since temporal attention can be combined with various variants of RNN, we compared the performance of different encoding layers combined with temporal attention. From Table 4 and Table 5, it can be seen that adding temporal attention to all three different encoding layers can improve the prediction accuracy. TA-LSTM has the best performance. Therefore, we chose TA-LSTM for the task of predicting speed. The comparison between the predicted and true values of each driving cycle for each road section is shown in Figure 7, Figure 8, Figure 9 and Figure 10. As can be observed from these figures, our model can fit the real data very well in most cases.

5.2. SOC Prediction Using FMI-Based EV Simulation

After predicting the speed of the EV, we subsequently utilized the predicted results and our simulator to predict the SOC.
We used FMI-based EV simulation with the previously predicted velocity and acceleration as inputs to predict the EV’s SOC. Our EV Simulation model is shown in Figure 11. As can be seen in the figure, given the inputs, the FMU2 to FMU6 are responsible for computing the SOC using the Formulas (14)–(38), whose variables are detailed in Table 6. Additionally, the lookup2d function in Formulas (29)–(31) performs the same operation as lookup2d in MATLAB/Simulink. The necessary efficiency values of the electric machine were obtained through the efficiency curves given in [9].
F t = F r r + F a d + F h c + F l a + F w a
F r r = μ r r m g
F a d = 1 2 ρ A C d v 2
F h c = m g sin ( α )
F l a = m a
F w a = 0.05 × F l a
T t = F t × r w
P t = F t × v
ω w = v r w
S w = 30 π × ω w
T s = T t η g × G ( P t > 0 )
T s r = η g × T t G ( P t < 0 )
S s = S s r = G × S w
P s = T s × S s × π 30
P s r = T s r × S s × π 30
Efficiency   ( η ) = lookup 2 d   ( Torque   ( Nm ) , Speed   ( rpm ) )
η m = lookup 2 d ( T s , S s )
η r = lookup 2 d ( T s r , S s r )
P b m = P s η m
P b r = η r × P s r
P b c ( t ) = P b m ( t ) + P a u x , F t ( t ) > 0 P b r ( t ) + P a u x , F t ( t ) < 0
I B ( t ) = E B 0 2 × R B i E B 0 2 × R B i 2 P b c R B i
Q ( t ) = 0 t I B d t
C = C 0 ( 1 K C t t b u ) ( 1 + a l p h a C ( T c u r r T r e f ) )
SOC ( t ) = C Q ( t ) C
To verify the effectiveness of our speed prediction model for the final SOC prediction, we compared the proposed model results with the speed and SOC prediction obtained with the Seasonal Autoregressive Integrated Moving Average (SARIMA) model. The final SOC prediction results are shown in Table 7. As can be seen from the table, the RMSE of TA-LSTM drops 56.69%, 84.97%, 82.34%, and 91.62% compared to SARIMA in the AtoC, CtoD, DtoE, and EtoF sections, respectively.
The predicted SOC results made by TA-LSTM and SARIMA for the driving cycle of each road section are shown in Table 8 and Figure 12, Figure 13, Figure 14 and Figure 15. From Table 8, we can see that TA-LSTM performed better than SARIMA in every driving cycle. In addition, the RMSE results of TA-LSTM for all driving cycles in the CtoD and DtoE sections are less than 0.0004 . Although the RMSE for the AtoC and EtoF section is slightly worse than the first three sections, the RMSE results are less than 0.0006 . The comparison between the real SOC and the predicted values shows that the predicted results of our model are very close to the real values. This indicates that our proposed method is able to perform SOC prediction properly.
In addition, in order to simulate the battery model by reflecting more realistic situations, we checked the change in SOC by varying the ambient temperature and battery usage time. The real driving cycle, the parameters of the battery model, and the ambient temperature and battery operating time values for the simulation are shown in Table 9.
From the results in Figure 16, it can be confirmed that the battery SOC decreased faster because the battery capacity decreases as the temperature decreases.
From the results of Figure 17, it can be confirmed that the battery SOC decreased more quickly because the battery capacity decreases as the battery usage time increases.

6. Conclusions

In our paper, we proposed a method for predicting the real driving-cycle-based SOC for EV batteries using deep-learning-based algorithms. For the RDC data, the authors used the data collected through the OBD II standard while driving from home to university in an actual vehicle. The proposed method first classifies detailed routes based on the section with traffic lights using GPS information, and finds the most similar patterns through DTW among the driving cycles of each section. Next, we predicted these sections by learning speed using TA-LSTM based on temporal attention, and predicted SOC based on this predicted speed. From the experimental results, it was confirmed that the case of using TA-LSTM could most accurately predict the speed, and that this could be very close to the actual values when used as a basis to predict the SOC. Therefore, if the proposed method is applied to the Real Driving Cycle, in which the same section is repeatedly driven, an accurate SOC prediction can be expected.
All the algorithms we propose can be implemented as an in-vehicle embedded system. It has been confirmed that DL models can be pre-trained and executed in an MCU. FMU simulation for the battery SOC can also be executed in an embedded system, such as an MCU or a sensor node [4,40]. If a high-capacity memory is required due to an increase in the demands of storing a massive amount of data or improving the calculation speed, a prediction can be performed in conjunction with a cloud data center or an edge server through the vehicular network, such as the Vehicle to Infrastructure (V2I) network.
The idea of the paper is to propose an algorithm that accurately predicts the SOC value of the next time step. However, predicting the SOC after multiple time steps may also be useful to the driver. Therefore, in a future work, the algorithm could be adapted so that it may predict, for example, the SOC after the completion of the entire driving cycle.

Author Contributions

Methodology, S.H., H.H. and D.K.; Project administration, S.H.; Software, S.H., H.H., D.K. and S.C.; Supervision, I.J.; Writing–original draft, S.H. and H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported partly by the Institute for Information & Communications Technology Promotion (IITP) grant funded by the Korea government (MSIP) (No. 2020-0-00107, Development of the technology to automate the recommendations for big data analytic models that define data characteristics and problems), and partly by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2019R1I1A1A01058964).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

ASICApplication Specific Integrated Circuits
BMSBattery Management System
CAVsConnected and Autonomous Vehicles
CSVComma Separated Value
CVCross Validation
DLDeep Learning
DNNsDeep Neural Networks
DTWDynamic Time Warping
EVElectric Vehicle
FGPAField-programmable gate array
FMIFunctional Mock-up Interface
FMUFunctional Mock-up Units
FSDFull Self-Driving
GPSGlobal Positioning System
GPUGraphics Processing Unit
GRUGated recurrent unit
MAPEMean Absolute Percentage Error
MCUMicrocontroller
MLMachine Learning
MSEMean Squared Error
NEDCNew European Driving Cycle
OBDOn Board Diagnostics
OCVOpen circuit voltage
RDCReal Driving Cycle
RMSERoot Mean Squared Error
RNNRecurrent Neural Network
RRARedundancy reduction algorithm
SARIMASeasonal Autoregressive Integrated Moving Average
SDCStandard Driving Cycle
SOCState of Charge
TA-LSTMTemporal Attention Long Short-Term Memory
TFLOPSTera floating point operations per second
V2IVehicle to Infrastructure
Variables
α Angle of the driving surface, rad
η g Gearbox efficiency
η m Efficiency of power consumed (motoring mode)
η r Efficiency of power generated (regenerative braking mode, W
α c Linear temperature coefficient of the capacity C, K 1
μ r r Coefficient of rolling resistance
ρ Density of the air
AFrontal area of the vehicle, m 2
aAceleration of the vehicle, m/s 2
CBattery capacity, Ah
C d Aerodynamic drag coefficient
E b 0 Open-circuit voltage of the battery, V
F t Traction force of the vehicle, N
F a d Aerodynamic drag, N
F h c Hill climbing force, N
F l a Force required to give linear acceleration, N
F r r Rolling resistance force of the wheels, N
F w a Force required to give angular acceleration to the rotating motor, N
GGear ratio of differential
gGravity acceleration, m/s 2
I B Battery current
mVehicle mass, kg
P s Shaft power of electric machine (motoring mode), W
P t Traction power, W
P a u x Power consumed by auxiliary loads, W
P b c Total power consumed in the EV, W
P b m Power consumed by electric machine (motoring mode), W
P b r Power consumed by electric machine (regenerative braking mode), W
P s r Shaft power of electric machine (regenerative braking mode), W
QTotal charge of the battery
r w Wheel radius, m
R B i Internal resistance of the battery, Ω
S s Shaft angular velocity of electric machine (motoring mode), rpm
S W Angular velocity of the wheels, rpm
S s r Shaft angular velocity of electric machine (regenerative braking mode), rpm
T s Shaft torque of electric machine (motoring mode), Nm
T t Traction torque, Nm
T r e f Reference temperature, K
T s r Shaft torque of electric machine (regenerative braking mode, Nm)
vVelocity of the vehicle, m/s
w W Angular velocity of the wheels, rad/s

References

  1. Cao, J.; Emadi, A. A New Battery/UltraCapacitor Hybrid Energy Storage System for Electric, Hybrid, and Plug-In Hybrid Electric Vehicles. IEEE Trans. Power Electron. 2012, 27, 122–132. [Google Scholar] [CrossRef]
  2. Sivak, M.; Schoettle, B. Relative Costs of Driving Electric and Gasoline Vehicles in the Individual US States, University of Michigan. Report No. SWT-2018-1. 2018. Available online: http://websites.umich.edu/~umtriswt/PDF/SWT-2018-1.pdf (accessed on 19 November 2021).
  3. Moghaddam, Z.; Ahmad, I.; Habibi, D.; Phung, Q.V. Smart Charging Strategy for Electric Vehicle Charging Stations. IEEE Trans. Transp. Electrif. 2018, 4, 76–88. [Google Scholar] [CrossRef]
  4. STMicroelectronics. Artificial Intelligence (AI) Plugin for Automotive SPC5 MCUs. Available online: https://www.st.com/en/development-tools/spc5-studio-AI.html (accessed on 19 November 2021).
  5. Li, S.; He, H.; Li, J.; Wang, H. Big data driven Deep Learning algorithm based Lithium-ion battery SoC estimation method: A hybrid mode of C-BMS and V-BMS. In Proceedings of the Applied Energy Symposium: MIT A+B, Boston, MA, USA, 22–24 December 2019. [Google Scholar]
  6. Lu, S.; Shi, W. The Emergence of Vehicle Computing. IEEE Internet Comput. 2021, 25, 18–22. [Google Scholar] [CrossRef]
  7. Liu, L.; Lu, S.; Zhong, R.; Wu, B.; Yao, Y.; Zhang, Q.; Shi, W. Computing Systems for Autonomous Driving: State of the Art and Challenges. IEEE Internet Things J. 2021, 8, 6469–6486. [Google Scholar] [CrossRef]
  8. Findelair, A. Tomorrow’s Car Silicon Brain, How Is It Made? Towards Data Science. Available online: https://towardsdatascience.com/tomorrows-car-silicon-brain-how-is-it-made-9090e1f06c9d (accessed on 23 March 2021).
  9. Bhatt, A. Planning and application of Electric Vehicle with MATLAB®/Simulink®. In Proceedings of the 2016 IEEE International Conference on Power Electronics, Drives and Energy Systems (PEDES), Trivandrum, India, 14–17 December 2016; pp. 1–6. [Google Scholar]
  10. Wei, M.; Ye, M.; Li, J.B.; Wang, Q.; Xu, X. State of Charge Estimation of Lithium-Ion Batteries Using LSTM and NARX Neural Networks. IEEE Access 2020, 8, 189236–189245. [Google Scholar] [CrossRef]
  11. Amrhein, M.; Krein, P.T. Dynamic simulation for analysis of hybrid electric vehicle system and subsystem interactions, including power electronics. IEEE Trans. Veh. Technol. 2005, 54, 825–836. [Google Scholar] [CrossRef]
  12. Lv, C.; Zhang, J.; Li, Y.; Yuan, Y. Mechanism analysis and evaluation methodology of regenerative braking contribution to energy efficiency improvement of electrified vehicles. Energy Convers. Manag. 2015, 92, 469–482. [Google Scholar] [CrossRef]
  13. Fiori, C.; Ahn, K.; Rakha, H.A. Power-based electric vehicle energy consumption model: Model development and validation. Appl. Energy 2016, 168, 257–268. [Google Scholar] [CrossRef]
  14. Genikomsakis, K.N.; Mitrentsis, G. A computationally efficient simulation model for estimating energy consumption of electric vehicles in the context of route planning applications. Transp. Res. Part D Transp. Environ. 2017, 50, 98–118. [Google Scholar] [CrossRef]
  15. Liu, K.; Wang, J.; Yamamoto, T.; Morikawa, T. Exploring the interactive effects of ambient temperature and vehicle auxiliary loads on electric vehicle energy consumption. Appl. Energy 2018, 227, 324–331. [Google Scholar] [CrossRef]
  16. Iora, P.; Tribioli, L. Effect of Ambient Temperature on Electric Vehicles’ Energy Consumption and Range: Model Definition and Sensitivity Analysis Based on Nissan Leaf Data. World Electr. Veh. J. 2019, 10, 2. [Google Scholar] [CrossRef] [Green Version]
  17. Luin, B.; Petelin, S.; Al-Mansour, F. Microsimulation of electric vehicle energy consumption. Energy 2019, 174, 24–32. [Google Scholar] [CrossRef]
  18. Blochwitz, T.; Otter, M.; Akesson, J.; Arnold, M.; Clauß, C.; Elmqvist, H.; Friedrich, M.; Junghanns, A.; Mauß, J.; Neumerkel, D.; et al. Functional Mockup Interface 2.0: The Standard for Tool independent Exchange of Simulation Models. In Proceedings of the 9th International MODELICA Conference, Munich, Germany, 3–5 September 2012; pp. 173–184. [Google Scholar]
  19. Hong, S.; Lim, D.; Joe, I.; Kim, W. F-DCS: FMI-Based Distributed CPS Simulation Framework with a Redundancy Reduction Algorithm. Sensors 2020, 20, 252. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Bär, T.; Nienhüser, D.; Kohlhaas, R.; Zöllner, J.M. Probabilistic driving style determination by means of a situation based analysis of the vehicle data. In Proceedings of the 14th International IEEE Conference on Intelligent Transportation Systems (ITSC), Washington, DC, USA, 5–7 October 2011; pp. 1698–1703. [Google Scholar] [CrossRef]
  21. Ellison, A.B.; Greaves, S.P.; Bliemer, M.C.J. Driver behaviour profiles for road safety analysis. Accid. Anal. Prev. 2015, 76, 118–132. [Google Scholar] [CrossRef]
  22. Wu, X.; He, X.; Yu, G.; Harmandayan, A.; Wang, Y. Energy-Optimal Speed Control for Electric Vehicles on Signalized Arterials. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2786–2796. [Google Scholar] [CrossRef]
  23. Varga, B.O.; Sagoian, A.; Mariasiu, F. Prediction of Electric Vehicle Range: A Comprehensive Review of Current Issues and Challenges. Energies 2019, 12, 946. [Google Scholar] [CrossRef] [Green Version]
  24. How, D.N.T.; Hannan, M.A.; Hossain Lipu, M.S.; Ker, P.J. State of Charge Estimation for Lithium-Ion Batteries Using Model-Based and Data-Driven Methods: A Review. IEEE Access 2019, 7, 136116–136136. [Google Scholar] [CrossRef]
  25. Xiong, R.; Cao, J.; Yu, Q.; He, H.; Sun, F. Critical Review on the Battery State of Charge Estimation Methods for Electric Vehicles. IEEE Access 2018, 6, 1832–1843. [Google Scholar] [CrossRef]
  26. Zhang, R.; Xia, B.; Li, B.; Cao, L.; Lai, Y.; Zheng, W.; Wang, H.; Wang, W. State of the Art of Lithium-Ion Battery SOC Estimation for Electrical Vehicles. Energies 2018, 11, 1820. [Google Scholar] [CrossRef] [Green Version]
  27. Yang, F.; Song, X.; Xu, F.; Tsui, K. State-of-Charge Estimation of Lithium-Ion Batteries via Long Short-Term Memory Network. IEEE Access 2019, 7, 53792–53799. [Google Scholar] [CrossRef]
  28. Chemali, E.; Kollmeyer, P.J.; Preindl, M.; Emadi, A. State-of-charge estimation of Li-ion batteries using deep neural networks: A machine learning approach. J. Power Sources 2018, 400, 242–255. [Google Scholar] [CrossRef]
  29. Babaeiyazdi, I.; Rezaei-Zare, A.; Shokrzadeh, S. State of charge prediction of EV Li-ion batteries using EIS: A machine learning approach. Energy 2021, 223, 120116. [Google Scholar] [CrossRef]
  30. Xing, Y.; Ma, E.W.M.; Tsui, K.L.; Pecht, M. Battery Management Systems in Electric and Hybrid Vehicles. Energies 2011, 4, 1840–1857. [Google Scholar] [CrossRef]
  31. Sakoe, H.; Chiba, S. Dynamic programming algorithm optimization for spoken word recognition. IEEE Trans. Acoust. Speech Signal Process. 1978, 26, 43–49. [Google Scholar] [CrossRef] [Green Version]
  32. Keogh, E.J.; Pazzani, M.J. Derivative Dynamic Time Warping. In Proceedings of the 2001 SIAM International Conference on Data Mining, Chicago, IL, USA, 5–7 April 2001. [Google Scholar]
  33. Sundermeyer, M.; Schluter, R.; Ney, H. LSTM Neural Networks for Language Modeling. In Proceedings of the Thirteenth Annual Conference of the International Speech COMMUNICATION association, Portland, OR, USA, 9–13 September 2012. [Google Scholar]
  34. Sutskever, I.; Martens, J.; Hinton, G. Generating Text with Recurrent Neural Networks. In Proceedings of the 28th International Conference on Machine Learning, Bellevue, WA, USA, 28 June–2 July 2011. [Google Scholar]
  35. Fu, R.; Zhang, Z.; Li, L. Using LSTM and GRU neural network methods for traffic flow prediction. In Proceedings of the 31st Youth Academic Annual Conference of Chinese Association of Automation (YAC), Wuhan, China, 11 November 2016. [Google Scholar]
  36. Cui, S.; Joe, I. Collision prediction for a low power wide area network using deep learning methods. J. Commun. Netw. 2020, 22, 205–214. [Google Scholar] [CrossRef]
  37. Qin, Y.; Song, D.; Cheng, H.; Cheng, W.; Jiang, G.; Cottrell, G.W. A dual-stage attention-based recurrent neural network for time series prediction. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  38. Fan, C.; Zhang, Y.; Pan, Y.; Li, X.; Zhang, C.; Yuan, R.; Wu, D.; Wang, W.; Pei, J.; Huang, H. Multi-horizon time series forecasting with temporal attention learning. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 25 July 2019. [Google Scholar]
  39. Cui, S.; Joe, I. A Dynamic Spatial-Temporal Attention-Based GRU Model With Healthy Features for State-of-Health Estimation of Lithium-Ion Batteries. IEEE Access 2021, 9, 27374–27388. [Google Scholar] [CrossRef]
  40. Bertsch, C.; Neudorfer, J.; Ahle, E.; Armugham, S.; Ramachandran, K.; Thuy, A. FMI for Physical Models on Automotive Embedded Targets. In Proceedings of the 11th International Modelica Conference, Versailles, France, 21–23 September 2015. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Procedure of Real Driving cycle based SOC estimation.
Figure 1. Procedure of Real Driving cycle based SOC estimation.
Applsci 11 11285 g001
Figure 2. Real Driving Cycle with Google MAP.
Figure 2. Real Driving Cycle with Google MAP.
Applsci 11 11285 g002
Figure 3. Speed and acceleration over time of the base time series. (a) Speed × time of the base time series. (b) Acceleration × time of the base time series.
Figure 3. Speed and acceleration over time of the base time series. (a) Speed × time of the base time series. (b) Acceleration × time of the base time series.
Applsci 11 11285 g003
Figure 4. Speed and acceleration over time of the four base segments. (a) Speed × Time (A to C segment); (b) Acceleration × Time (A to C segment); (c) Speed × Time (C to D segment); (d) Acceleration × Time (C to D segment); (e) Speed × Time (D to E segment); (f) Acceleration × Time (D to E segment); (g) Speed × Time (E to F segment); (h) Acceleration × Time (E to F segment).
Figure 4. Speed and acceleration over time of the four base segments. (a) Speed × Time (A to C segment); (b) Acceleration × Time (A to C segment); (c) Speed × Time (C to D segment); (d) Acceleration × Time (C to D segment); (e) Speed × Time (D to E segment); (f) Acceleration × Time (D to E segment); (g) Speed × Time (E to F segment); (h) Acceleration × Time (E to F segment).
Applsci 11 11285 g004
Figure 5. Architecture of TA-LSTM.
Figure 5. Architecture of TA-LSTM.
Applsci 11 11285 g005
Figure 6. The effect of temporal attention on RMSE results.
Figure 6. The effect of temporal attention on RMSE results.
Applsci 11 11285 g006
Figure 7. Speed prediction of route A to C: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 7. Speed prediction of route A to C: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g007
Figure 8. Speed prediction of route C to D: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 8. Speed prediction of route C to D: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g008
Figure 9. Speed prediction of route D to E: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 9. Speed prediction of route D to E: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g009
Figure 10. Speed prediction of route E to F: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 10. Speed prediction of route E to F: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g010
Figure 11. FMI-based model for EV power consumption.
Figure 11. FMI-based model for EV power consumption.
Applsci 11 11285 g011
Figure 12. SOC prediction of route A to C: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 12. SOC prediction of route A to C: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g012
Figure 13. SOC prediction of route C to D: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 13. SOC prediction of route C to D: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g013
Figure 14. SOC prediction of route D to E: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 14. SOC prediction of route D to E: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g014
Figure 15. SOC prediction of route E to F: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Figure 15. SOC prediction of route E to F: (a) cycle 8; (b) cycle 9; (c) cycle 10; (d) cycle 11.
Applsci 11 11285 g015
Figure 16. Temperature effect on the battery SOC.
Figure 16. Temperature effect on the battery SOC.
Applsci 11 11285 g016
Figure 17. Battery aging effect on the battery SOC.
Figure 17. Battery aging effect on the battery SOC.
Applsci 11 11285 g017
Table 1. Table showing the start and end time of the segments and the search periods.
Table 1. Table showing the start and end time of the segments and the search periods.
First Base
Segment
Second Base
Segment
Third Base
Segment
Fourth Base
Segment
Start Point (s)069135224
End Point (s)69135224404
Search Start
Point (s)
01985174
Search End
Point (s)
119185274454
examles of Search periods for interval = 5 s (start, end)(0, 69),
(5, 74),

(45, 114),
(50, 119)
(19, 85),
(24, 90),

(114, 180),
(119, 185)
(85, 174),
(90, 179),

(180, 269),
(185, 274)
(174, 354),
(179, 359),

(269, 449),
(274, 454)
Table 2. MAPE velocity results of SARIMA for the SDC (NEDC) and RDC.
Table 2. MAPE velocity results of SARIMA for the SDC (NEDC) and RDC.
ModelSDC (NEDC)RDC (Four Road Segments)
AtoCCtoDDtoEEtoF
SARIMA0.0140.060.060.120.35
Table 3. RMSE results of 5-fold CV.
Table 3. RMSE results of 5-fold CV.
HyperparametersAverage RMSE
Tm
5160.4908
320.4880
640.4926
1280.4926
10160.4988
320.4998
640.5013
1280.4982
15160.5007
320.5055
640.5025
1280.5059
Table 4. RMSE results of different models.
Table 4. RMSE results of different models.
ModelsAtoCCtoDDtoEEtoF
RNN0.44850.37020.45490.4924
GRU0.43690.36890.45720.4873
LSTM0.42400.35560.45240.4864
Table 5. RMSE results of temporal attention with different models.
Table 5. RMSE results of temporal attention with different models.
ModelsAtoCCtoDDtoEEtoF
TA-RNN0.44120.36140.45150.4870
TA-GRU0.42950.35860.45270.4832
TA-LSTM0.41710.34560.44610.4821
Table 6. Parameters for FMU-based EV simulation (Default setting).
Table 6. Parameters for FMU-based EV simulation (Default setting).
ParametersDescriptionValues
mMass of the vehicle (kg)1000
r w Wheel radius (m)0.2736
gGravity acceleration (m/s 2 )9.81
ρ Air density1.2
AFront area of vehicle (m 2 )2.36
α Angle of driving surface (rad)0
μ r r Rolling resistance coefficient0.015
C d Aerodynamic drag coefficient0.3
η g Gearbox efficiency0.98
GGearbox ratio8.59
C 0 Initial capacity (C)720,000
R b i Internal resistance (Ω)0.008
E b 0 Open-circuit voltage (V)53.6
K c t Linear t dependency of the capacity C 6.084436 × 10 10
t b u battery usage time (s)0
T r e f Reference temperature ( C)20 C
T c u r r Ambient temperature ( C)20 C
a l p h a C Linear temperature coefficient of capacity (K 1 )0
Table 7. RMSE ( × 10 2 ) results of SOC prediction.
Table 7. RMSE ( × 10 2 ) results of SOC prediction.
ModelsAtoCCtoDDtoEEtoF
SARIMA0.08200.09980.15120.3938
TA-LSTM0.03560.01500.02670.0330
Table 8. RMSE ( × 10 2 ) results of SOC prediction of each driving cycle.
Table 8. RMSE ( × 10 2 ) results of SOC prediction of each driving cycle.
CycleAtoCCtoDDtoEEtoF
SARIMATA-LSTMSARIMATA-LSTMSARIMATA-LSTMSARIMATA-LSTM
80.08920.01610.12880.01700.12360.01950.25050.0176
90.07920.04990.04110.01390.11790.02370.51700.0252
100.06040.03680.14290.01630.14080.03810.26150.0288
110.09500.03100.03390.01220.20600.02140.47110.0509
Table 9. Parameters for the battery model simulation.
Table 9. Parameters for the battery model simulation.
ParametersDescriptionValues
Driving cycleReal driving cycleCycle 8
K c t Linear t dependency of the capacity C 6.084436 × 10 10
t b u battery usage time (s)0, 157,698,305 s (5 years),
315,396,610 s (10 years)
T r e f Reference temperature ( C)20 C
T c u r r Ambient temperature ( C)0, 20, 30 C
a l p h a C Linear temperature coefficient of capacity (K 1 )0.03 (0)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hong, S.; Hwang, H.; Kim, D.; Cui, S.; Joe, I. Real Driving Cycle-Based State of Charge Prediction for EV Batteries Using Deep Learning Methods. Appl. Sci. 2021, 11, 11285. https://doi.org/10.3390/app112311285

AMA Style

Hong S, Hwang H, Kim D, Cui S, Joe I. Real Driving Cycle-Based State of Charge Prediction for EV Batteries Using Deep Learning Methods. Applied Sciences. 2021; 11(23):11285. https://doi.org/10.3390/app112311285

Chicago/Turabian Style

Hong, Seokjoon, Hoyeon Hwang, Daniel Kim, Shengmin Cui, and Inwhee Joe. 2021. "Real Driving Cycle-Based State of Charge Prediction for EV Batteries Using Deep Learning Methods" Applied Sciences 11, no. 23: 11285. https://doi.org/10.3390/app112311285

APA Style

Hong, S., Hwang, H., Kim, D., Cui, S., & Joe, I. (2021). Real Driving Cycle-Based State of Charge Prediction for EV Batteries Using Deep Learning Methods. Applied Sciences, 11(23), 11285. https://doi.org/10.3390/app112311285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop