1. Introduction
Lithium-ion batteries (LIBs) are widely used in electric vehicles and energy storage systems due to their high energy density, low self-discharge rate, long service life, high reliability, and high safety. LIBs are easily affected by factors such as the charging and discharging cycle, temperature changes, fluctuations, and stress in different operating conditions, resulting in continuous battery degradation [
1]. Therefore, it is necessary to accurately and real-time estimate the capacity of lithium batteries to provide safer and more reliable energy for electric vehicles (EVs), smart grids, and other applications. This will enhance the stability and manageability of the energy system while maximizing the utilization of battery capacity. State of health (SOH) is a widely accepted indicator of battery life degradation, providing insight into the extent of degradation in a battery’s current maximum discharging capacity compared to its factory-rated capacity [
2]. SOH directly impacts battery performance and usable lifespan. Currently, three common methods are employed for SOH estimation: electrochemical mechanism-based estimation, equivalent circuit model (ECM)-based estimation, and data-driven methods. The electrochemical model (EM) is a mechanistic approach that simulates battery chemical reactions to characterize battery performance [
3,
4]. EM offers strong interpretability in describing external battery characteristics. However, EM relies on certain parameters that are challenging to measure or identify, such as electrode materials and electrolyte chemistry. Consequently, the model may not accurately capture battery aging behavior [
5]. In contrast to the EM, the ECM method offers an alternative approach that is not reliant on the internal chemistry of a battery. By using essential electrical components, such as resistance, capacitance, and inductance, the ECM method constructs a circuit model capable of describing the battery’s external characteristics. Through the amalgamation of this model with substantial state data and the application of circuit analysis, the ECM method proves to be a powerful tool for accurately identifying the health state of a battery [
6]. ECM is computationally efficient, making it the primary battery model employed for online capacity estimation in battery management systems (BMSs) for EVs. However, its accuracy is frequently hindered by the performance of model parameterization.
The data-driven approach offers a solution to the limitations of the aforementioned methods, as it does not heavily rely on parameters. This approach does not require an understanding of the internal principles of the battery or the construction of a physical battery model [
7]. This method relies on collected battery aging data to extract and analyze information related to battery capacity degradation, enabling the estimation of the SOH for LIBs [
8,
9]. Since pure data-driven methods for capacity estimation lack transparency and interpretability [
10], embedding physical or domain knowledge into the development of machine learning models has received more attention [
11,
12]. The use of a physics-informed neural network (PINN) is a classic method in this type of solution. It can fully integrate the physical characteristics of the problem into the loss function, thereby improving the generalization performance of data-driven models by utilizing data and mathematical operators based on underlying physical principles. Wen et al. [
13] proposed a semi-physical and semi-empirical dynamic model to capture the capacity degradation of batteries, and based on this, they introduced a PINN to fuse the prior information of the dynamic model with the information extracted from monitoring data. Hofmann et al. combined simulated data generated by the pseudo-2-dimensional (P2D) model with real-world data for data-driven model training to reduce the computational cost of traditional physical and chemical models [
14].
Data-driven methods can be broadly categorized into feature-based models and learning-based models. The accuracy of SOH estimation in feature-based models heavily relies on the extraction of relevant features. By employing correlation analysis, more features can be incorporated, leading to more accurate capacity estimation. One prominent feature-based method is incremental capacity curve analysis (ICA), which effectively quantifies battery capacity loss [
15,
16]. The ICA method is derived from the voltage differential capacity and extracts the incremental capacity (IC) curve from the discharging and charging phases. Zhang et al. [
17] extracted peak values from the discharge IC curve for SOH estimation, while Li et al. [
18] extracted the peak and its position from the charging IC curve as the health characteristics. Although IC analysis can explain the aging behavior inside the battery, it needs to be differentiated to filter the data, which increases the difficulty of data processing [
19]. In addition, the noise in the data point is easily enlarged, resulting in a decrease in estimated accuracy. For the feature selection level, Li et al. [
20] extracted the features of the IC curve while taking the proportion of the constant current (CC) mode, the equal voltage drop, and the sample entropy of the discharge voltage as health features. Wang et al. [
21] extracted three health features under CC-CV charging and CC discharge modes: the time corresponding to the same charging voltage interval, the time corresponding to the same discharge voltage interval, and the time corresponding to the same discharge temperature change.
The learning-based model aims to simplify or eliminate the feature extraction process by directly inputting a certain length of charging segment data as model input [
22], mainly using a convolutional neural network (CNN) for feature extraction. Yang et al. [
23] used the CNN to extract the SOH index and the change in capacity between two consecutive charge–discharge cycles (ΔSOH) and used the random forest algorithm to obtain the final capacity degradation estimation result. The correlation of features has a great influence on the accuracy of the model. Sun et al. [
24] introduced an attention mechanism, by giving important feature weights, so that the key features in the input sequence that have a greater impact on the output results.
However, the existing data-driven methods are mostly based on feature extraction under relatively ideal conditions in the laboratory (the peak value of the IC curve, the voltage interval corresponding to the peak value, etc.), or the complete external characteristic curve (voltage, current, etc.) is used as the input of the data-driven model. Although this can achieve good results, it has certain limitations in real vehicle applications. First of all, the EVs are usually not fully charged and discharged, and the charging behaviors are complex and random. Secondly, in actual use and battery experiments, there are problems such as long experimental time, high cost, and difficulty in obtaining labels [
25,
26]. Furthermore, the effectiveness of the data-driven method depends on a substantial volume of historical data, which can be challenging to obtain in BMSs of EVs in practical conditions. Therefore, there is significant research value in developing a capacity estimation model that is suitable for real-world scenarios and can accurately predict capacity degradation using a limited amount of data.
For this issue, research on capacity estimation based on incomplete fragment inputs has shown great research value. For example, Tian et al. [
27] used a small portion of the charging curve as input to estimate the complete charging curve and then used the predicted charging curve as an input feature for capacity estimation. In addition, Zhang et al. [
28] introduced shapelets to further partition the charging segments based on the extraction of short charging segments and used the MRMR algorithm to find the optimal shapelet as the model input, in order to greatly reduce the input volume of the model. For relatively fixed fragment inputs, the model can learn the mapping relationship between features and capacity degradation well.
Considering that the state of charge (SOC) at the beginning (end) of charging is not fixed in actual vehicle operation scenarios, this article uses fixed charging segments and random charging segments as inputs for research. In addition, this paper introduces a flexible and robust data-driven method capable of achieving accurate estimation of battery capacity degradation across diverse battery datasets, even with limited samples available. The overall framework of this method is illustrated in
Figure 1.
In this method, a learning-based capacity estimation model is first created using a gated convolutional and bi-directional long short-term memory hybrid neural network (GateCNN-BiLSTM). The CNN is used to capture the advanced spatial features of the battery sequence [
29], and the bi-directional long short-term memory (BiLSTM) neural network is used to extract the positive and negative context information of the battery sequence [
30]. The gating mechanism can suppress noise information and enhance the sensitivity of the network to important features, thereby improving the generalization ability of the network. The model can use easily measured data (such as voltage and current) recorded in any short charging time to estimate the current battery capacity state. Secondly, based on the learning-based model, combined with the capacity estimation framework of single-source domain and multi-source domain migration, the model also shows excellent performance under a very small amount of data.
In summary, the main contributions of this paper include the following three points:
The proposed model has a flexible input and output structure and can accurately estimate the current capacity state based on short charging data segments. This method does not depend on the laboratory aging test environment of the complete charge and discharge cycle and avoids the problem that the existing data-driven methods need to perform artificial feature extraction. Therefore, this method has wide applicability in actual vehicle charging scenarios.
Through the verification of different LIB aging datasets, it is proved that the learning-based model is robust to datasets of various battery types and charge–discharge conditions. In the case of random charging segment input, the minimum root mean square error (RMSE) and mean absolute error (MAE) of the capacity estimation results are 0.384% and 0.339%, respectively.
The transfer learning framework proposed in this study shows good results in the case of a very small amount of data, which significantly improves the training speed and stability. It overcomes the problem of the capacity label being difficult to obtain and the labeling cost being high in practical applications, and it further enhances the robustness and versatility of the method.
The remaining sections of this paper are organized as follows:
Section 2 presents the lithium battery dataset used in this study.
Section 3 introduces the basic estimation model for lithium battery capacity degradation estimation, focusing on the utilization of partial charging segments as input and discussing different output structures. In
Section 4, we propose a transfer learning framework that combines the basic capacity degradation estimation model with small-sample learning to enhance the model’s generalization capability in scenarios with limited samples.
Section 5 presents the experimental results and analysis. Finally,
Section 6 concludes the paper.
4. Capacity Estimation Method Based on Small-Sample Learning
4.1. Small-Sample Scenario Division
The blue curve in
Figure 9 represents the capacity degradation curve of the target domain. To verify the knowledge transfer performance of LIBs’ health status in the case of small samples, the corresponding small-sample scenarios in the target knowledge domain data are set. As shown in the red dots in
Figure 9, the sample data of the target domain only contain limited information and fail to fully carry the knowledge of the entire information space. This means that the GateCNN-BiLSTM model cannot accurately infer the knowledge that is not included in the complete information space. At this time, the model faces the challenge of incomplete knowledge and needs to be solved by transfer learning. Given this situation, the first 5%, 10%, 15%, and 20% charging cycle samples in the target domain are selected as the training set of the target domain to study the influence of different sample ratios on the capacity estimation results.
4.2. Capacity Estimation Method of Single-Source Domain Migration Based on Domain Adaptation
To improve the performance of the basic capacity estimation model when the battery training samples are scarce, a domain adaptive transfer learning algorithm based on multi-core maximum mean difference (MMD) is proposed. By transferring the relevant feature knowledge of other battery datasets to the capacity estimation model, the information deficiency caused by insufficient training data is compensated for. In addition, this method also breaks the limitation that the training set and the test set of the neural network must be identically distributed, and it further expands the application range of the algorithm.
The MMD-based transfer learning algorithm integrates the MMD metric into the basic model GateCNN-BiLSTM of transfer learning to form the GateCNN-BiLSTM-MMD model. When the basic model extracts the features of the source domain and the target domain, the MMD is used to reduce the distribution difference between the source domain and the target domain features, so that the basic model can learn the features of the source domain and the target domain at the same time and realize the transfer of the source domain feature knowledge to the target domain. The transfer learning framework based on MMD is shown in
Figure 10.
The main idea of the MMD algorithm is to map the samples of the two datasets to a high-dimensional space through a specific mapping function, that is, the reproducing kernel Hilbert space (RKHS) [
44]. In RKHS, the samples of the two datasets are separated, and the MMD value is the Euclidean distance between the average data points in the two datasets [
45].
MMD is widely used in transfer learning algorithms to quantify the distribution difference between the source domain and the target domain [
46,
47]. When MMD = 0, it means that the distribution of the two domains is the same. On the contrary, the larger the MMD value, the greater the distribution difference between the two domains. According to the main idea of MMD, its definition is as follows:
where
X and
Y represent the sample space of two different distributions;
and
represent the samples extracted from different sample spaces;
represents RKHS;
and
represent the number of samples;
represents the nonlinear function of data mapping to RKHS. However, it is difficult to find a suitable mapping function, so the kernel trick is used to deal with it. The common kernel trick is as follows:
Therefore, Equation (20) can be converted to:
The battery capacity estimation algorithm based on MMD transfer learning is shown in Algorithm 1.
Algorithm 1 Domain adaptation capacity estimation based on MMD transfer learning |
Input: Output: Capacity degradation rate Ratec |
Normalize and Divide into small-sample scenarios. That is, the first 20% of the samples are extracted as the target domain training set train, and the last 80% of the samples are used as the test set test. is used to form a source domain dataset for knowledge transfer. while loss > 0.0001 do Taking and train as inputs, the target domain feature T1 and the source domain feature T2 are obtained by GateCNN-BiLSTM. Calculate the maximum mean difference of the two features. The final loss value is calculated by superimposing the loss value obtained by each input variable. end while is obtained and result is normalized. The root mean square error (RMSE) and mean absolute error (MAE) are calculated. Return RMSE, MAE, model.
|
In this model, the battery task dataset with a limited number of samples is used as the target domain dataset, and the complete battery dataset with different distributions from the target domain is used as the source domain dataset. Based on the input data of the source domain and the target domain, the data characteristics of the source domain and the target domain are obtained through the basic GateCNN-BiLSTM model. To measure the distribution difference between the source domain and the target domain features, the MMD values of the source domain and the target domain features are calculated. In addition, while calculating the MMD value, the source domain and target domain features are used as the input of the BiLSTM model to obtain the output values of the source domain and the target domain. Then, based on the output values of the source domain and the target domain, the loss functions of the source domain and the target domain are calculated by using the source domain label and the target domain label . Finally, combining these two loss functions with the MMD value, the loss function of the GateCNN-BiLSTM-MMD model is obtained. During the model training process, by adjusting the weight of the CNN model, the feature distribution of the source domain will continue to approach the feature distribution of the target domain until the two distributions are the same. Finally, the source domain features that are distributed with the target domain are applied to the target domain model, and the migration of the source domain features to the target domain is realized.
Due to the differences in charging conditions and battery types, the capacity degradation mechanism of LIBs is not consistent, and the battery aging data do not meet the characteristics of an independent and identical distribution. In this paper, different charging conditions and different types of battery data are divided into different source domains, and the target domain is established for small-sample data of specific charging conditions and specific battery types. Finally, knowledge transfer is completed across domains based on transfer learning modeling. The knowledge domain division methods based on different charging conditions and battery types are shown in
Table 2 and
Table 3, respectively. The numbers after the symbol # represent different datasets.
As an index for measuring the difference in feature distribution, MMD gradually reduces the difference in feature distribution between the source domain and target domain by adjusting parameters in the iterative process of the model to achieve the goal of transfer learning. This process is realized by an iterative optimization model, which makes the model have better generalization performance and adaptability in the target domain.
4.3. Capacity Estimation Method of Multi-Source Domain Migration Based on Meta-Learning
For LIBs, a single-source domain usually can only provide limited battery capacity degradation information, and when the distribution difference between the source domain data and the target domain data is too large, the mobility will be low, and even negative migration will occur in extreme cases. This has great limitations for solving the lack of domain knowledge in the case of small samples. Therefore, this section will expand on the differences between charging conditions and battery types. Based on the theory of meta-learning, combined with the basic learner of GateCNN-BiLSTM for training, the capacity estimation modeling of LIBs is completed by fusing multi-source domain data and the knowledge transfer from multi-distributed sources.
4.3.1. Model-Agnostic Meta-Learning
Both meta-learning and transfer learning can be applied to small-sample learning. The difference between them is that transfer learning needs to transfer the corresponding knowledge from the source domain to the target domain, which is highly dependent on the source domain. If the data of the source domain and the target domain are completely different, the effectiveness of knowledge transfer will be affected [
48,
49]. The essence of meta-learning is to increase the generalization ability of the learner in multiple tasks. Through continuous learning and adaptation of each specific task, the model network has a certain learning ability. Even if the source task and the target task are very different, this learning ability can also be applied to new tasks [
50,
51].
The model-agnostic meta-learning (MAML) algorithm is adopted to learn the general initialization of model parameters in a set of related tasks and then fine-tune the initialization according to the data of specific tasks, thus allowing the model to quickly adapt to new tasks with only a small amount of additional data. The dataset in each task can be divided into a support set and a query set. The support set is used to update the parameters of a specific model. After the update, the model obtains the corresponding loss through the query set. Finally, the loss sum of all tasks is counted to provide gradient information for the update of the initial parameters of the model [
52].
4.3.2. An Improved Meta-Learning Method Based on Weight Guidance
The ultimate goal of the MAML algorithm is to make the initial parameters of the model balance the learning performance of each task. It does not seek the best learning performance of a single task, but it makes the performance of each task better and seeks the minimum joint training loss of all tasks. This is out of generalization considerations. However, at the level of knowledge transfer, considering that there are differences between tasks, treating the source tasks equally will inevitably lead to the problem of a low knowledge transfer rate.
To improve the efficiency of knowledge transfer, this section improves the learning process of MAML from the correlation between the source task and the target task and proposes a meta-learning improvement scheme based on weight guidance. The MMD value is used to measure the data distribution similarity between the target task and the source task dataset, and it is multiplied by the loss as a key reference for the final parameter update to increase the proportion of tasks with greater correlation in the meta-training process, thereby improving the task training efficiency of MAML. Unlike the transfer learning method in
Section 4.2, the MMD here is not embedded in the loss function of the model for iterative optimization, but as a definition of a similar correlation. It can be seen as a measure that assigns corresponding weights to different datasets to ensure that in the meta-learning framework, tasks that are more relevant to the test set can be considered more pertinently when determining the initial parameters of the final model. The new meta-training loss function is as follows:
where
is the weight of each task, and
α is a specific value; in this section,
to prevent the denominator from being 0. The smaller
is, the larger
is.
In this paper, the improved meta-learning framework is named MMD-MAML. As shown in
Figure 11, the meta-learning framework contains two loops. In the inner loop, each task is trained to obtain the training loss of each task. In the outer loop, the loss of all tasks is summarized, and then the gradient update is performed to obtain the optimal parameters of the model.
Through the gradient update of the inner loop and the outer loop, when the model is learning in different tasks, the initial parameters have prior knowledge of the historical task. When there is a new task, this prior knowledge can make the model quickly adapt to the new task.
According to Algorithm 2, this section measures the domain of each source task and the target task, obtains the correlation between each auxiliary task and the target task, and converts the correlation into the weight of the task training loss to control the training effect of each task on the target task. By changing the loss function of MAML, the direction of updating the initial model parameters in the parameter space is controlled, that is, the loss weight of the target task is greater than the loss weight of the source task, so that the parameters of the initial model are updated as far as possible towards the optimal parameter position of the target task.
Algorithm 2 LIB capacity estimation algorithm based on MMD-MAML. |
Input: Task distribution p(T); inner and outer loop steps α and β; multi-source domain migration framework Output: Capacity degradation rate Ratec |
While not done do: for do: Extract specific task Ti from task distribution p(T), and divide the dataset into support set and query set. A task is iterated n times, and the support set is input for model training to obtain the corresponding loss . The model parameters are updated by gradient descent: Use the query set in Ti to test the accuracy of the updated model and calculate the loss . end for Based on all , the initial model parameters are updated: end while Return Capacity degradation estimation model Modelmeta.
|
4.3.3. Meta-Learning Task Division
Similar to the single-source domain small-sample migration scheme, in the multi-source domain migration scheme, the learning task is still set based on the difference in charging conditions and the difference in battery type. The specific division methods are shown in
Table 4 and
Table 5, respectively. In the tables, # 1 represents the first dataset in a large class of batteries (NCA batteries), while * 1 and * 2 represent a small-sample scenario in a meta-testing dataset and the corresponding remaining samples, respectively.
Figure 12 shows the capacity degradation curve corresponding to each task dataset. It is worth emphasizing that the similarity of capacity degradation does not directly reflect the MMD value between knowledge domains, because MMD mainly focuses on the distribution of sample features without considering label information (i.e., capacity degradation). The features of different datasets are normalized to eliminate the influence of dimension. Then, the training weight of each task is calculated according to Equation (23) by calculating the MMD value of each dataset and the meta-test dataset. The smaller the MMD value is, the closer the feature distribution of the dataset is, and the larger the corresponding training weight is.
Figure 13 shows the weight ratio of different tasks in the case of fixed segment input and random segment input. It can be observed that there is a large correlation between the dataset of training task 2 and the dataset in the target task. Therefore, when performing the meta-training outer loop, training task 2 is given a greater weight to promote the parameter update more inclined to adapt to the dataset of task 2. This weight-guided meta-learning scheme helps to enhance the ability to learn the relevance of different tasks, thereby improving the performance of the meta-learning model.
In
Section 4.2, knowledge transfer between two different battery types under the single-source domain framework is carried out. As shown in
Table 5, according to the difference in battery type, four meta-training task datasets and one meta-testing target task dataset are divided. The battery capacity and cathode material between each task have a certain degree of difference, but the charging rate is uniformly selected at 0.5 C. Among them, the support set corresponds to the first
n% small-sample scenario, and the query set corresponds to the remaining samples.
The capacity degradation curve presented by the dataset based on the difference in battery type is shown in
Figure 14. It can be seen that the battery degradation curves corresponding to training tasks 2 and 3 show a certain similarity with the target dataset, which also increases the possibility of the model learning common features and patterns to a certain extent, thus improving the generalization performance on unknown tasks.
In addition, as can be seen from
Figure 15, compared with the situation based on different charging conditions, the weight difference between training tasks based on different battery types is not significant, and the weight of task 3 and task 1 is slightly higher than that of other tasks, which is slightly different from the traditional MAML method’s assumption that all task weights are consistent.
5. Experimental Results and Analysis
To eliminate the influence of dimension, a data normalization method is used. To prevent the model from overfitting, a regularization term and dropout layer are introduced. At the same time, RMSE and MAE are used to evaluate the generalization performance of the capacity estimation model.
In Equations (24) and (25), is the estimated value of capacity, is the true value of capacity, and N is the number of true values.
Due to the uncertainty caused by random segments, unexpected or unwanted fluctuations or changes occur in the data. Therefore, the wavelet threshold denoising method is used to remove the interference noise in the capacity data to extract the real signal in the data. As shown in Equations (26)–(28), the core idea of wavelet denoising is to decompose the signal into sub-bands of different frequencies by wavelet transform and to reduce noise by threshold processing of these sub-bands.
where
is the wavelet basis function, and in this paper, Daubechies wavelet is selected;
and
represent the scale and translation parameters, respectively;
is the wavelet coefficient, and
;
is the detail coefficient after soft threshold processing.
5.1. Capacity Estimation Results and Discussion Based on Partial Segment Input
From the Tongji University dataset, 60% was selected as the training set for model training, 10% as the validation set for model parameter tuning, and the remaining 30% was used as a test set to test the estimation performance of the model. Each partition of the training set, validation set, and test set is random, and two different input structures designed in
Section 3.1 are used for comparison. The battery capacity degradation estimation results based on partial segment input are shown in
Figure 16.
Compared with the CNN model, the hybrid neural network model (CNN-LSTM, CNN-GRU) can better fit the test data. The CNN-BiLSTM network has achieved better estimation results than the CNN-LSTM/GRU network because it can extract the time series characteristics of the battery from the forward and backward directions. At the same time, GateCNN combines the neural network structure of the convolutional layer and the gated recurrent layer, which can suppress the noise information in the features and improve the generalization ability of the network. Therefore, the basic model proposed in this paper obtains the optimal estimation results. At the same time, to make the model training results more stable, the instantaneous capacity appreciation caused by the experimental operation is eliminated.
Table 6 gives the RMSE and MAE values of the capacity estimation using different methods when the initial fixed segment is used as the input. It can be seen that the RMSE and MAE values of the GateCNN-BiLSTM model in the three batteries are the lowest. The corresponding RMSE values are 0.245%, 0.161%, and 0.391%, respectively. The MAE values are 0.201%, 0.121%, and 0.308%, respectively. This shows that the proposed model has good robustness.
When random segment charging data are used as input for Tongji University data, the variation in input features is large, which makes the data distribution more complicated, as shown in
Figure 17. Compared with the fixed segment input, the capacity estimation results obtained by random segment input fluctuate slightly, but the relative error is within −0.05%. According to
Table 7, the maximum RMSE and MAE in the three batteries are 0.623% and 0.594%, respectively.
The above results show that the output results of the GateCNN-BiLSTM model proposed in this paper show great stability on the datasets of these three different batteries, whether there is a fixed segment or random segment input. It is further proved that the GateCNN-BiLSTM model has strong robustness and generalization performance and can flexibly and accurately estimate the capacity degradation of battery cells.
To validate the proposed model on capacity degradation in other public datasets, the University of Maryland datasets were used, specifically the CS2_35, CS2_36, and CS2_37 datasets.
Figure 18 displays the results of the comparison of the proposed model to other contrastive models, under the condition of fixed initial short input segments. The proposed model achieved the best estimation results across all three battery datasets.
Table 8 presents the capacity degradation estimation errors of different data-driven models that rely on fixed segment input obtained from the University of Maryland datasets. The GateCNN-BiLSTM model exhibits the lowest RMSE and MAE for estimated capacity, which are 1.292% and 0.996%, respectively. The average RMSE and MAE are 1.551% and 1.196%. The proposed model outperforms all other models on all datasets, indicating its superior estimation accuracy and robustness.
Figure 19 illustrates the capacity estimation results of different models for University of Maryland datasets when subjected to random fragment input scenarios. By employing gating mechanisms to select feature information from random fragment input, the proposed model exhibits greater stability and closely approximates the actual capacity curve.
Additionally,
Table 9 indicates that the proposed model achieves lower capacity estimation errors when subjected to random fragment input scenarios. Furthermore, it can be observed that the model performs marginally better on the CS2_35 battery dataset, compared to other datasets, under both input structures. This superior performance may be attributed to the better mapping relationship between the health characteristics and capacity of this dataset.
Next, the data of the actual vehicle battery pack are used to verify the applicability of the battery capacity degradation evaluation model in practical applications. The results are shown in
Figure 20. The estimated value is very close to the target value. The maximum relative error based on the fixed short segment input is 0.944%, and the RMSE is 0.540%, respectively. The maximum relative error based on random short segment input is 0.998%, and RMSE is 0.761%, respectively. Therefore, the model proposed in this paper shows excellent accuracy and practical application in solving the problem of battery pack capacity degradation estimation.
5.2. Analysis and Discussion of Capacity Estimation Results Based on Small-Sample Learning
5.2.1. Single-Source Domain Migration Method Based on Domain Adaptation
In this section, the transfer learning model based on domain adaptation is used to estimate the capacity of different small-sample scenarios, and the influence of different sampling ratios is discussed.
Figure 21 and
Figure 22 show the battery capacity estimation results and errors for Tongji University data in the case of using the first
n% continuous samples, and
n is taken as 5, 10, 15, and 20. It can be seen that in the case of only extracting the first
n% of continuous samples, the model may rely too much on this part of the data, resulting in a weak generalization ability of the entire dataset. As the sample size increases, the generalization ability of the model on subsequent data may be affected due to the lack of follow-up information learning, resulting in a slower decline in error.
For the two migration schemes, with the increase in the sample size of the target dataset, there are also some differences in the model error based on the difference in charging conditions and the difference in battery types. For the migration scheme based on the difference in charging conditions, by performing migration learning on the same battery datasets of two different charging conditions, the model may more easily learn common patterns and features, because the datasets between the two charging conditions come from the same battery type. Therefore, as the sample size of the target data collection increases, the model may converge faster, and the accuracy will increase faster. For the migration schemes of different battery types, although the datasets under the same charging conditions have certain similarities, the model needs to adapt to the differences between different battery types, which may lead to more challenges in the learning and generalization process of the model. Therefore, as the sample size of the target data collection increases, the accuracy of the model may grow more slowly, because the model requires more data to better adapt to the differences between different batteries.
The MMD transfer learning algorithm proposed in this paper belongs to the feature-based transfer learning algorithm. In addition, there are also instance-based and parameter-based transfer learning algorithms. In this section, the MMD-based feature transfer learning algorithm, the instance-based transfer learning algorithm, and the parameter-based transfer learning algorithm are compared on the same GateCNN-BiLSTM model to verify the effectiveness of the MMD feature transfer learning algorithm. For the instance-based transfer learning method, the classic TrAdaBoost algorithm [
53] is selected. The main idea of this method is to select samples similar to the target domain from the source domain, re-weight them, and then add the weighted samples to the target domain training set to assist in the training of the target domain model. For the parameter-based transfer learning method, the Finetune algorithm is selected, which is also a classical algorithm [
54]. The core idea of this method is as follows: firstly, the neural network is trained by using the source domain data with sufficient samples, then some parameters of the network are frozen, and finally the network is retrained by using the target domain training set with limited samples. In this way, the migration of source domain model parameters to the target domain model is realized.
Figure 21 shows the estimation results of capacity degradation based on the difference in charging conditions when extracting the first 20% of the samples. To better evaluate the performance of each transfer learning algorithm,
Table 10 gives the RMSE and MAE values of each transfer learning algorithm for capacity degradation estimation. Among the three transfer learning algorithms, the MMD-based method has better accuracy, and the minimum RMSE and MAE are 1.558% and 1.318%, respectively.
In the case of capacity degradation estimation with battery type differences, as depicted in
Figure 23c,d, the knowledge transfer performance in the later stages of battery aging is relatively poor. This can be attributed to significant changes in the degradation mechanism of different batteries during the later stages of their lifespan. Consequently, the samples in the source domain struggle to effectively assist in training during the training process. According to the results presented in
Table 10, knowledge transfer based on battery type differences does not yield better estimation performance improvements compared to those based on differences in operating conditions. This suggests that the sample disparities caused by battery type outweigh the differences between charging conditions.
In the case of using the first 20% of the charging cycle as the target domain dataset, the neural network model lacks complete knowledge expression ability due to the absence of late aging information about the battery in the training data. This indicates that a data-driven model without the inclusion of a transfer learning algorithm cannot construct an accurate capacity degradation estimator solely based on a small amount of early cycle data from the battery charging condition.
It is important to highlight that the TrAdaBoost and Finetune methods did not yield satisfactory results. This can be attributed to significant distribution differences between the source domain data and the target domain data. TrAdaBoost attempts to address this issue by re-weighting certain source domain data and incorporating them into model training alongside the target domain data. On the other hand, the Finetune algorithm fine-tunes a model trained on source domain data using a small amount of target domain training data to better adapt to the target domain’s characteristics. Both of these transfer learning algorithms heavily rely on the similarity between the source and target domain data. However, in the capacity estimation study conducted here, there is a transfer between different datasets, resulting in substantial distribution differences between the source and target domain data. These significant distribution disparities greatly impact the performance of these two transfer learning algorithms, making it challenging to achieve the desired outcomes.
5.2.2. Multi-Source Domain Migration Method Based on Meta-Learning
To verify the effectiveness of the improved MMD-MAML framework based on weight guidance proposed in this paper, the Reptile [
55] and MAML methods are selected for comparison. The learning mechanism of Reptile is similar to that of MAML, as it focuses on improving the direction of gradient descent for model parameters. However, as mentioned earlier, the single-source domain transfer learning method tends to exhibit deviations in the later stages of the cycle due to differences in charging conditions and battery types.
From
Figure 24, it is evident that the capacity degradation estimation results achieved through multi-source domain transfer learning are significantly superior to those of single-source domain transfer. This improved performance effectively compensates for the errors in the later stages of the cycle. The reason behind this improvement is that MAML leverages its ability to comprehensively learn the knowledge expression of lithium-ion battery capacity degradation across multiple tasks. By dividing various charging conditions or battery types into source tasks and conducting meta-training, MAML can acquire the domain knowledge necessary for the target task. Additionally, the initial model parameters are tested using new small-scale data. Based on this, we introduce MMD to enhance MAML, considering the relevance of tasks, with the aim of further improving the accuracy of the model.
Table 11 presents the capacity estimation results for different charging conditions and battery types using the first 20% sample condition. In a comparison with the MAML and Reptile frameworks, it can be concluded that the capacity estimation results obtained by the MMD-MAML framework under different charging conditions closely align with the actual curve. Additionally, the RMSE and MAE values for MMD-MAML reach the minimum values, which are 0.254% and 0.157%, respectively. Compared to the single-source domain migration scheme, the estimation curve of the MMD-MAML framework better approximates the actual value in the later stages of battery capacity degradation. Despite the lack of post-cycle data in the input sample of the target task, the model can still learn the post-capacity-degradation pattern of the target battery through training on the source task, resulting in improved capacity estimation performance.
Regarding battery type differences, when considering the input features, the utilization of fixed segments as input yields the lowest RMSE and MAE for the estimated capacity, which are 0.758% and 0.678%, respectively. The variation in degradation mechanisms among different cathode materials is more significant compared to that of charging conditions. Consequently, the RMSE and MAE obtained using corresponding fixed short segment input increased by 198.42% and 221.85%, respectively. Similarly, the RMSE and MAE obtained using corresponding random short segment input increased by 155.70% and 242.18%, respectively. The meta-learning-based knowledge transfer method serves as an active training mechanism. Hence, even if the meta-testing task lacks sample data for a specific interval, the meta-learning method can actively learn the capacity degradation patterns across various batteries and aging intervals during the learning process of multiple source tasks. Although the capacity degradation estimator’s estimation error may increase in the later stages of the cycle, it is significantly reduced compared to the single-source domain method. Overall, the estimator provides a better fit for the battery’s capacity degradation.
6. Conclusions
This paper presents a novel learning-based approach for estimating the capacity degradation of lithium-ion batteries (LIBs). We introduce a hybrid neural network, GateCNN-BiLSTM, which combines gated convolutional and bi-directional long short-term memory, to establish a data-driven basic model. To address scenarios with limited previous charge and discharge cycle data, we develop a learning framework that combines single-source domain and multi-source domain transfer learning. This framework is applied to estimate the capacity degradation of LIBs and is validated using various aging datasets of different LIB types. The research findings are summarized as follows:
The GateCNN-BiLSTM basic model demonstrates the ability to capture advanced spatial features and forward–reverse time features of battery sequence data, enabling comprehensive feature extraction. By utilizing voltage and current data recorded during a short charging period, the model accurately estimates the current battery capacity state. Compared to traditional data-driven models, it achieves higher accuracy in battery capacity estimation across multiple battery datasets, even with random fragment inputs, the average root mean square error (RMSE) and mean absolute error (MAE) are 0.468% and 0.436% respectively. For vehicle battery pack capacity degradation estimation, the maximum relative error and RMSE for battery charging capacity estimation based on fixed short segment inputs are 0.944% and 0.540%, respectively. For charging capacity estimation based on random short segment inputs, the maximum relative error and RMSE are 0.998% and 0.761%, respectively. These results highlight the model’s wide applicability in capacity estimation for different battery types, without relying on a fixed input structure. Furthermore, considering the uncertainties in charging and discharging behavior in EVs, the model accurately estimates battery capacity degradation using short charging segment data.
In addressing the challenge of limited data in practical applications, we compare our transfer learning method with other approaches and find that the single-source domain feature transfer learning method based on maximum mean difference (MMD) effectively transfers battery health state information from a source domain with sufficient data to a target domain. This solves the problem of limited generalization in small-sample scenarios. By dividing the knowledge domain into charging condition difference and battery type difference, we analyze the variation in battery capacity estimation results based on different proportions of continuous charging samples. As the sample size increases, the capacity estimation error decreases, albeit at a slower rate. For the first 20% of the samples, the lowest RMSE and MAE are 1.558% and 1.318%, respectively. This indicates that our method significantly reduces the dependence of data-driven models on historical data quantity, improving the model’s performance even with limited data. Multi-task learning is implemented using the multi-source domain meta-learning MAML algorithm. By incorporating MMD to measure the correlation between the source task and the target task, the efficiency of meta-training is enhanced, compensating for any deviation in single-source domain knowledge transfer during the later stages of capacity degradation. This leads to more accurate and efficient capacity degradation estimation. In the future, with further optimization, we anticipate that the proposed battery capacity estimation method can be applied in cloud-based or vehicle-mounted battery management systems (BMSs).