Literature DB >> 35647454

Secondary Structural Ensemble Learning Cluster for Estimating the State of Health of Lithium-Ion Batteries.

Si-Zhe Chen1, Hongtao Zhang1, Long Zeng1, Yuanliang Fan2,3, Le Chang1, Yun Zhang1.   

Abstract

Accurate online state-of-health (SOH) estimation can improve the operational efficiency of lithium-ion batteries (LIBs) and ensure the safety of energy storage systems. However, the complex electrochemical properties of LIBs make accurate SOH estimation challenging. To overcome this challenge, we propose a secondary structural ensemble learning (SSEL) cluster. The proposed SSEL cluster includes multiple SSEL frameworks established separately within different SOH data intervals, allowing the identification of stable feature-SOH relationships. The adaptability and basic accuracy of each SSEL framework are guaranteed by various base learners and the corresponding stacking model and bagging model fusion. Each framework remains unique and specialized owing to the adoption of back propagation neural networks, which adjust learner weights based on the feature-SOH relationship at each interval. The effectiveness of the SSEL cluster was verified using the Oxford Battery Degradation Dataset 1. Comparisons showed that the proposed estimation method performs better than traditional machine learning methods.
© 2022 The Authors. Published by American Chemical Society.

Entities:  

Year:  2022        PMID: 35647454      PMCID: PMC9134389          DOI: 10.1021/acsomega.2c01589

Source DB:  PubMed          Journal:  ACS Omega        ISSN: 2470-1343


Introduction

Lithium-ion batteries (LIBs) have several advantages, including high energy density, a long service life, and a lack of memory effect.[1−6] Therefore, they are widely used in electric vehicles and portable devices and are also used for energy storage in power systems. However, non-emergency usage causes the continuous aging of LIBs, decreasing their actual capacity, affecting battery performance, and even causing serious accidents. Therefore, accurate state-of-health (SOH) estimation is critical for the safe use of LIBs. Current SOH prediction methods for LIBs can be classified into three types: direct measurement methods, model-based methods, and data-driven methods. In the direct measurement method,[7,8] SOH is evaluated based on aging-related characteristics, such as the capacity and internal resistance of the LIB. Although this method is centered on basic mechanisms and can be used for several types of batteries, high measurement accuracy can be achieved only under regulated laboratory conditions. To overcome the drawbacks of the direct measurement method, a model-based method was developed. This method uses empirical knowledge[9−14] to establish the relationship between measured battery signals and the SOH. However, the model parameters need to be adjusted as the battery ages, making computations increasingly complex. In the data-driven method, accurate SOH estimations can be made based on operational data, without analyzing internal chemical reactions. The data-driven method is widely used for SOH estimation. Hu et al.[15] showed that the K-nearest neighbor (KNN) algorithm can illustrate the relationship between battery features and the SOH, even though its multi-dimensional data processing performance is poor. Li et al.[16] developed a new random forest model that predicts LIB SOH using raw data without any preprocessing. Furthermore, because neural networks (NNs) have a strong nonlinear fitting ability, Shen et al.[17] employed a NN to evaluate SOH. Although both methods could accurately evaluate the SOH, they were prone to overfitting. Hence, support vector regression (SVR) was adopted to estimate the SOH, and good generalization to an unknown dataset was observed.[18] Nevertheless, the predicted performance of SVR remained sensitive to missing data.[19] To predict SOH with satisfactory accuracy, data-driven methods have been combined. A hybrid model consisting of extreme learning machine and random vector functional link networks was applied to predict SOH, ensuring efficient and accurate estimation.[20] In another study,[21] the weight-sharing feature of convolutional NNs (CNNs) and the sequential correlation of long short-term memory were jointly applied to improve SOH estimation accuracy. In the study by Dai et al.,[22] special features were used as inputs for a feed-forward NN, and the Markov model was then used to correct NN estimations and obtain the final SOH. This method combined the strong robustness of feed-forward NNs with the benefits Markov models offer in correcting fitting errors. Meng et al.[23] used an ensemble learner consisting of many SVR models to estimate SOH. They demonstrated that the accuracy and robustness of SOH estimation could be improved further through ensemble learning. However, current ensemble learners lack mechanisms that optimize internal weights based on the relationships between features and SOH (i.e., feature–SOH relationships). The highly nonlinear nature of datasets also compromises the accuracy of data-driven SOH estimation. In previous reports,[24,25] the empirical mode decomposition method was used to decompose the original data and reduce the impact of complicated existing relationships. This method effectively enhanced the accuracy of SOH estimation by reducing relationship complexity. However, in this method, data decomposition was required before each estimation. Thus, the method was not suitable for online application. Currently, two major challenges remain. The first is the establishment of a reasonable weight assignment mechanism for ensemble learning. The second is reducing the impact of highly nonlinearized data on SOH prediction accuracy without increasing the computational burden of online estimation. To overcome these two challenges, this article proposes a secondary structural ensemble learning (SSEL) cluster. The main contributions of this study are as follows: Multiple SSEL frameworks were separately trained with datasets involving different degrees of aging to reduce the impact of highly nonlinearized feature–SOH relationships. Across different SOH data intervals, the SSEL cluster could specifically optimize the weights of internal learners based on feature–SOH relationships. A generative adversarial network (GAN) was applied to expand the dataset, solving the problems caused by insufficient data. The remainder of this article describes the experimental data preprocessing process (Section ), the proposed SSEL cluster (Section ), and an experimental comparison between the SSEL cluster and other data-driven models, along with the factors affecting estimation accuracy (Section ). Finally, Section concludes the article.

Data Preprocessing

SOH Definition

SOH is represented by the current capacity and rated capacity,[26] as followswhere and are the current and rated capacity of the LIB, respectively.

Experimental Data

Oxford Battery Degradation Dataset 1,[27] which contains long-term battery aging data from eight Kokam (SLPB533459H4) 740 mAh lithium-ion pouch batteries, was used to evaluate the algorithms. These batteries were tested using a constant-current–constant-voltage (CC–CV) charging profile and a drive cycle discharging profile obtained from the Artemis urban cycle. The testing temperature was maintained at 40 °C. Characteristic measurements were taken every 100 cycles, with a 1 C charge and a 1 C discharge current. The aging curves are shown in Figure .
Figure 1

SOH aging curves of eight lithium-ion batteries.

SOH aging curves of eight lithium-ion batteries.

Feature Extraction

The powerful estimation ability of the data-driven method comes from the close relationship between model features and SOH; hence, the selection of appropriate features is imperative. The voltage, current, and temperature of a LIB during CC–CV charging are closely related to battery degradation. Moreover, these physical quantities are also easy to measure. Therefore, these parameters were chosen as the features for the SOH estimation of LIBs. The typical CC–CV charging curves of LIBs are shown in Figure . The voltage, current, and temperature curves were divided into 150 time-based subsections, and the average voltage, current, and temperature within each subsection were calculated as feature data.
Figure 2

Typical curve showing CC–CV charging in LIBs.

Typical curve showing CC–CV charging in LIBs.

GAN Data Expansion

The data-driven approach requires a large amount of experimental data for model training. A GAN[28] can be used to overcome the problem of insufficient experimental data. Through mutual gaming of the generator and the discriminator, new data that followed the patterns of the original dataset were generated. The generator G and discriminator D were trained through alternating iterations. The weights of the discriminator D were kept untrainable while training the generator G and vice versa. The GAN structure is shown in Figure .
Figure 3

GAN network structure.

GAN network structure. The complete GAN training function was as followswhere E(*) is the mathematical expectation, X is the real data, Z represents the random variables, denotes the distribution of the real data X, denotes the distribution of the random variable Z, G(*) is the output of the generator G, and D(*) is the output of the discriminator D. As the values of and increase, the discriminatory ability of the discriminator D also increases. As the value of decreases, the data generated by the generator G become more realistic. Through alternate iterations, the distribution pattern of G(Z) approaches the distribution pattern of the real data X. The generator G has six fully connected layers that transform the three-dimensional random variable Z into 151-dimensional generated data. The first 150 dimensions are the features, and the 151st dimension is the SOH value. The generated data and real data were inputted into the discriminator separately. After running through four fully connected layers, the discriminator could verify the truth or falsity of each data point.

SSEL Cluster

Proposed SSEL Cluster for SOH Estimation

In this article, an SSEL cluster was proposed to improve the accuracy of SOH estimation. A host of SSEL frameworks with different SOH data intervals were established to identify a stable relationship between features and SOH. In each SSEL framework, the different data-driven models were considered base learners for the first-level stacking model[29] and bagging model fusion.[30] The outputs of the base learners, stacking model fusion, and bagging model fusion were used as inputs for the second-level back propagation neural network (BPNN) fusion model. By training the BPNN, a weight assignment mechanism was generated, improving the model’s function fitting ability.

First-level Stacking Model Fusion

First-level stacking model fusion was performed using different models to understand the feature–SOH relationships from different perspectives, which enabled the models to complement each other. The stacking model fusion ensured basic estimation accuracy in each SSEL framework. Its principle is shown in Figure .
Figure 4

Principle of stacking model fusion.

Principle of stacking model fusion. The implementation of the stacking model fusion is shown in Figure . The procedure was as follows.
Figure 5

Implementation of the stacking model fusion.

Implementation of the stacking model fusion. Step 1. Based on the sub-training sets, each base learner was assessed by fourfold cross-validation. Three of the sub-training sets, S1, S2, S3, and S4, were used to train the base learner. Then, the remaining sub-training set and the Stest set were used to validate and test the trained base learners. The results of the four validations were recorded as the column vectors a1, a2, a3, and a4, and the corresponding testing results were recorded as the column vectors b1, b2, b3, and b4. Step 2. A new column vector was generated by vertically merging the column vectors a1, a2, a3, and a4. This vector was defined as the new training set feature A1. The new testing set feature B1 was generated by averaging the elements in the corresponding positions of the column vectors b1, b2, b3, and b4. Step 3. Steps 1 and 2 were looped until all base learners were trained, validated, and tested to obtain new training set features A1, A2, ..., A, and new testing set features B1, B2, ..., B. Step 4. The training set features of the meta-learner were formed by merging the column vectors A1, A2, ..., A horizontally into a matrix. The real column vector Areal corresponding to A1, A2, ..., A was jointly used as the training set label for the meta-learner. A1, A2, ..., A and Areal were fed into the meta-learner for training. Step 5. The testing set features of the meta-learner were formed by merging the column vectors B1, B2, ..., B horizontally into a matrix. The real column vector Breal corresponding to B1, B2, ..., B was jointly used as the testing set label for the meta-learner. B1, B2, ..., B and Breal were fed into the meta-learner for testing.

First-level Bagging Model Fusion

Stacking model fusion has created large biases in some SOH intervals due to the use of the fusion. Given the obvious differences between the bagging and stacking models in terms of fusion principles, the bagging model fusion was added within each SSEL framework to improve estimation accuracy further. The principle of the bagging model fusion is shown in Figure .
Figure 6

Principle of the bagging model fusion.

Principle of the bagging model fusion. Simple random sampling with replacement was performed in the training set to construct P sub-training sets. The P sub-training sets and the testing set were used to train and test each base learner separately. Subsequently, the average of all predictions from these learners was calculated to obtain the final estimation.

Second-level BPNN Fusion Model

As the LIB ages, the relationships between features and SOH change. Owing to fixed internal weights, ordinary models cannot adapt to such continuous variation. Therefore, multiple SSEL frameworks were separately established within different SOH data intervals. The SSEL frameworks in the intervals comprise the proposed SSEL cluster. The relationships learned by the SSEL framework for each SOH interval were as followswhere , represent our model’s labels, features, and the learned relationship between in the x-th SOH interval, respectively. H represents the feature dimension number of each physical quantity, and X represents the number of SOH intervals., , and are the voltage, current, and temperature features collected in the x-th SOH interval, respectively. In each SSEL framework, the BPNN fusion model was trained to obtain the weights of the base learners. As shown in Figure , the stacking model fusion, bagging model fusion, and base learners were incorporated while training the BPNN fusion model. The role of the base learners is to ensure that the different types of feature–SOH relationships are well-fitted by the SSEL framework. The stacking and bagging model fusion algorithms guarantee the basic estimation accuracy of the SSEL framework. Furthermore, the BPNN fusion model assigns weights for each learner. SOH estimation in the SSEL framework is performed as follows.where Q is the number of learners; and are the outputs of the i-th learner and the SSEL framework in the x-th SOH interval, respectively; and is the weight of the i-th learner in the x-th SOH interval.
Figure 7

SSEL framework of one SOH interval.

SSEL framework of one SOH interval.

Procedure for the SOH Estimation Method

The implementation procedure for the SSEL cluster is shown in Figure . The steps are as follows.
Figure 8

Implementation procedure for the SSEL cluster.

Implementation procedure for the SSEL cluster. Step 1. The voltage, current, temperature, and SOH values from the CC–CV charging profile of the LIBs were used as the original dataset. Then, the data were normalized and divided into the training, validation, and testing sets. Step 2. A GAN was used to expand the training set. The validation and testing sets were both divided into multiple intervals based on the SOH. Step 3. The expanded training set was used for training various base learners, stacking model fusion, and bagging model fusion. The base learners included XGBoost, the light gradient boosting machine (LGBM), SVR, extra tree regressor, decision tree regressor, linear regressor, KNN, and CNN. The linear regressor was chosen as the base learner for bagging model fusion. The base learners for stacking model fusion were the extra tree regressor, decision tree regressor, and linear regressor. The meta-learner for stacking model fusion was the KNN. Step 4. Each SSEL framework was trained using the validation set of the corresponding interval. In each SSEL framework, the outputs of the base learners, stacking model fusion, and bagging model fusion were used as inputs for the single-layer BPNN. Moreover, the weights of all learners were distributed based on the BPNN fusion model. SOH estimation for the SSEL cluster was performed based on eq . Step 5. The SSEL cluster was tested using the testing set of the corresponding interval. The tested model could then be used for LIB SOH estimation.

Results and Discussion

To prove the superiority of the SSEL cluster, the proposed model was compared with other models using data from the Oxford Battery Degradation Dataset 1. In addition, the effects of the dividing intervals method, GAN data expansion, and fusion models on LIB SOH estimation were evaluated. The root mean square error (RMSE) was used as the evaluation index for SOH estimation.where SOHtrue  and SOHest  are the true and predicted values, respectively, and N is the number of training samples.

Performance of the SSEL Cluster

The Oxford Battery Degradation Dataset 1 included data from eight batteries. The data were collected in the SOH range of 70–100%, which was defined as the complete life cycle of the battery. In case I, data from batteries no. 1–7 were divided into the training (80%) and validation sets (20%), and data from battery no. 8 were used as the testing set. The training set was expanded by using a GAN for training various base learners, stacking model fusion, and bagging model fusion. The validation and testing sets were divided into three intervals corresponding to the SOH ranges of 70–80, 80–90, and 90–100%. Each SSEL framework was trained using the validation set of the corresponding interval. Based on training results, the weights of learners in each SSEL framework were determined (Table ). The SSEL cluster was used to estimate the SOH of battery no. 8; the estimates and errors are shown in Figure .
Table 1

Weights of the SSEL Cluster for Case I and Case II

caseSOH interval (%)bagging (%)stacking (%)extra tree regressor (%)CNN (%)decision tree regressor (%)XGBoost (%)LGBM (%)KNN (%)SVR (%)linear regressor (%)
I90–1001.8725.89–7.29–16.9115.8215.762.5630.06–6.0938.48
 80–90–6.5921.69–5.6622.21–23.4610.2529.624.8818.9027.50
 70–801.4634.90–0.2613.04–10.0732.384.74–12.77–4.2440.80
II90–10037.62–4.5038.964.778.95–5.2917.7613.59–6.08–5.96
 80–9035.317.1626.70–9.4422.3112.89–4.19–13.62–7.2930.57
 70–806.4525.825.94–20.16–7.2627.2521.10–13.2732.4317.50
Figure 9

SOH estimation and errors for battery no. 8.

SOH estimation and errors for battery no. 8. In case II, data from battery no. 4 were used as the testing set, and data from the remaining batteries were used as the training and validation sets. The above preprocessing, training, and testing processes were re-executed to obtain the weights of learners in each SSEL framework (Table ). The SSEL cluster was used to estimate the SOH of battery no. 4; the estimates and errors are shown in Figure .
Figure 10

SOH estimation and errors for battery no. 4.

SOH estimation and errors for battery no. 4. Table presents the internal weights of the SSEL cluster for batteries no. 8 and no. 4. The weight of each learner was adjusted across different SOH intervals, enabling each SSEL framework to fit the corresponding feature–SOH relationships accurately. Notably, either the stacking or bagging model was always maintained as a heavyweight in all SOH intervals for batteries no. 8 and no. 4, ensuring the basic accuracy of SOH estimation. The proposed model was compared with other models using the same training and testing sets as those of cases 1 and 2. XGBoost, LGBM, SVR, extra tree regressor, decision tree regressor, linear regressor, KNN, and CNN were used as comparative models. The results for cases 1 and 2 are shown in Table . The RMSE of all SSEL frameworks was within 0.6%, indicating that they outperformed the other models. SOH estimation results and error curves were plotted for the four comparative models with the highest estimation accuracy in cases 1 and 2 (Figures and 12, respectively). As shown in Figures –12, other models could not guarantee estimation accuracy across the entire SOH interval, especially in specific SOH ranges.
Table 2

Estimation Results

  SSEL cluster
comparative models
caseitem90–100%interval80–90% interval70–80% intervalXGBoostLGBMSVRextra tree regressordecision tree regressorlinear regressorKNNCNN
IRMSE (%)0.35560.42210.57181.24660.67874.67740.83890.92900.70790.71121.3210
 accuracy (%)99.644499.577999.428298.753499.321395.322699.161199.071099.292199.288898.6790
IIRMSE (%)0.35680.39940.10490.97621.01423.02791.77101.81850.68950.90311.4010
 accuracy (%)99.643299.600699.895199.023898.985896.972198.229098.181599.310599.096998.5990
Figure 11

SOH estimation and errors for battery no. 8. (a–d) LGBM, KNN, linear regressor, and extra tree regressor fitting curves for battery no. 8, respectively. (e–h) LGBM, KNN, linear regressor, and extra tree regressor error curves for battery no. 8, respectively.

Figure 12

SOH estimation and errors for battery no. 4. (a–d) LGBM, KNN, linear regressor, and XGBoost fitting curves for battery no. 4, respectively. (e–h) LGBM, KNN, linear regressor, and XGBoost error curves for battery no. 4, respectively.

SOH estimation and errors for battery no. 8. (a–d) LGBM, KNN, linear regressor, and extra tree regressor fitting curves for battery no. 8, respectively. (e–h) LGBM, KNN, linear regressor, and extra tree regressor error curves for battery no. 8, respectively. SOH estimation and errors for battery no. 4. (a–d) LGBM, KNN, linear regressor, and XGBoost fitting curves for battery no. 4, respectively. (e–h) LGBM, KNN, linear regressor, and XGBoost error curves for battery no. 4, respectively.

Effect of the Dividing Interval Method

To explore the influence of the dividing intervals method, a single SSEL framework was trained and tested based on the validation and testing sets without any division. The results are presented in Table . By assigning weights to each learner, the fitting ability of single SSEL frameworks could be improved. However, the SOH estimation accuracy was substantially lower than that of comparative models. During the entire life cycle of the LIB, each learner within the single SSEL framework corresponded to only one fixed weight. Therefore, the changing relationship between the features and SOH was not accurately represented. Hence, the SOH estimation was less accurate than the SSEL cluster-based SOH estimation.
Table 3

Comparison between the Performance of the Dividing Interval SSEL Cluster and the Single SSEL Framework

 SSEL cluster RMSE (%)
single SSEL framework RMSE (%)
dataset90–100%80–90%70–80%70–100%
battery no. 80.35560.42210.57180.6097
battery no. 40.35680.39940.10490.5888

Effect of GAN Data Expansion

The effect of GAN data expansion on the accuracy of LIB SOH estimation was evaluated. Training for various base learners, stacking model fusion, and bagging model fusion were performed using the unexpanded original training set. Then, the SSEL cluster was trained and tested with the dividing interval method. The results are displayed in Table . It can be seen that GAN data expansion had a positive effect on the accuracy of SOH estimation. This was because the procedure prevented overfitting.
Table 4

Effect of GAN Data Expansion on the SSEL Cluster

 dividing intervals RMSE (%)
 90–100%
80–90%
70–80%
datasetGANwithout GANGANwithout GANGANwithout GAN
battery no. 80.35560.61390.42210.59260.57180.8842
battery no. 40.35680.49290.39940.57570.10490.5433

Effect of Model Fusion

The effect of model fusion on LIB SOH estimation was evaluated. Table shows the RMSE obtained after using stacking model fusion, bagging model fusion, and the SSEL cluster. In general, the SSEL cluster provided better accuracy for SOH prediction than the stacking model or bagging model fusion.
Table 5

Comparison between the Performance of Different Fusion Models for Battery no. 8 and no. 4

  first-level fusion model RMSE (%)
 
batterySOH intervals (%)stackingbaggingSSEL cluster RMSE (%)
no. 890–1000.45861.12100.3556
 80–900.67080.55560.4221
 70–800.76810.59560.5718
no. 490–1000.42570.45210.3568
 80–900.79060.29270.3994
 70–800.26370.26400.1049
Tables and 5 show that in case I, stacking model fusion outperformed all comparative models during the entire life cycle, except during the 70–80% SOH interval. Similarly, bagging model fusion outperformed all comparative models, except during the 90–100% SOH interval. In case II, the RMSE obtained after applying stacking model fusion was lower than that obtained with other comparative models, and it was inferior to the linear regressor only in the 80–90% SOH interval. Bagging model fusion provided the best estimation accuracy compared with all the other models. This showed that the stacking model and bagging model fusion play an important role in ensuring the basic estimation accuracy of the SSEL cluster.

Conclusions

This article presents an SSEL cluster for LIB SOH estimation based on voltage, current, and temperature measurements. The dataset, which was expanded using GAN, was divided into intervals based on true SOH values. Then, an independent SSEL framework was built into each interval to fit the feature–SOH relationships accurately. Various base learners enabled the SSEL framework to accommodate the complex relationships between features and SOH; stacking model fusion and bagging model fusion ensured basic estimation accuracy; the BPNN fusion model assigned weights to each learner to allow optimal cooperation among learners. Owing to these mechanisms, the SSEL cluster provided improved SOH estimation accuracy. Through comparisons with other models, the superiority of the SSEL cluster was verified. Moreover, the role of the dividing interval method, GAN data expansion, and fusion in improving the accuracy of SOH estimation was explored experimentally.
  1 in total

1.  Evaluation of Various Offline and Online ECM Parameter Identification Methods of Lithium-Ion Batteries in Underwater Vehicles.

Authors:  Peiyu Chen; Chengyi Lu; Zhaoyong Mao; Bo Li; Chiyu Wang; Wenlong Tian; Mengjie Li; Yunwei Xu
Journal:  ACS Omega       Date:  2022-08-19
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.