Next Article in Journal
Optimal Time Frequency Fusion Symmetric Dot Pattern Bearing Fault Feature Enhancement and Diagnosis
Previous Article in Journal
Extraction and Validation of Biomechanical Gait Parameters with Contactless FMCW Radar
Previous Article in Special Issue
Optimizing Battery Charging Using Neural Networks in the Presence of Unknown States and Parameters
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Lithium-Ion Battery Health Predictions by Hybrid-Grained Graph Modeling

1
College of Electronics and Information Engineering, Sichuan University, Chengdu 610065, China
2
School of Cyber Science and Engineering, Sichuan University, Chengdu 610207, China
3
College of Biomedical Engineering, Sichuan University, Chengdu 610065, China
4
Systems Hub, The Hong Kong University of Science and Technology (Guangzhou), Guangzhou 511453, China
*
Authors to whom correspondence should be addressed.
Submission received: 12 May 2024 / Revised: 20 June 2024 / Accepted: 25 June 2024 / Published: 27 June 2024

Abstract

:
Predicting the health status of lithium-ion batteries is crucial for ensuring safety. The prediction process typically requires inputting multiple time series, which exhibit temporal dependencies. Existing methods for health status prediction fail to uncover both coarse-grained and fine-grained temporal dependencies between these series. Coarse-grained analysis often overlooks minor fluctuations in the data, while fine-grained analysis can be overly complex and prone to overfitting, negatively impacting the accuracy of battery health predictions. To address these issues, this study developed a Hybrid-grained Evolving Aware Graph (HEAG) model for enhanced prediction of lithium-ion battery health. In this approach, the Fine-grained Dependency Graph (FDG) helps us model the dependencies between different sequences at individual time points, and the Coarse-grained Dependency Graph (CDG) is used for capturing the patterns and magnitudes of changes across time series. The effectiveness of the proposed method was evaluated using two datasets. Experimental results demonstrate that our approach outperforms all baseline methods, and the efficacy of each component within the HEAG model is validated through the ablation study.

1. Introduction

Lithium-ion batteries, as a type of high energy density and lightweight energy storage device, are widely utilized in energy fields such as smartphones, laptops, drones, and electric vehicles [1,2]. However, the performance of lithium-ion batteries can deteriorate over repeated charge–discharge cycles, a phenomenon referred to as battery aging, resulting from physical and chemical changes caused by routine usage and operation [3]. Battery aging has the potential to lead to significant safety incidents and economic losses. State-of-health (SOH) refers to the ratio of a battery’s maximum capacity to its rated capacity, and it can also serve as a criterion for assessing battery aging [4,5]. SOH is primarily influenced by factors such as the number of charge–discharge cycles, charge–discharge rate, operating temperature, and internal resistance [6].
Current methodologies for predicting lithium-ion battery SOH are broadly classified into filter-based [7], machine learning [8], and deep learning techniques [9]. Filter techniques have limited capacity for handling large datasets, while machine learning requires intensive manual feature engineering and domain expertise, leading to potential labor intensiveness and suboptimal generalization. Deep learning traditionally struggles to capture dynamic temporal dependencies between input time series. As illustrated in Figure 1, dynamic dependencies typically manifest as either coarse-grained or fine-grained. Coarse-grained dependencies are akin to observing the overall trend over a long period, such as the daily temperature variation over a year, providing a broad view but missing finer details. Fine-grained dependencies focus on short-term variations, like hourly temperature changes within a day, capturing detailed fluctuations but potentially overlooking the broader trend. Dynamic graph networks are frequently employed in existing work to model evolving dependencies. However, predicting SOH is usually a long-term process influenced by intermittent factors [10]. Coarse-grained models struggle to capture sudden events and transient impacts in input time series, while fine-grained models find it challenging to capture the long-term dynamic temporal dependencies of SOH [11]. Consequently, both types of models may introduce certain errors in predicting SOH.
To address the limitations of existing models, we introduce the HEAG for predicting the lifespan of lithium-ion batteries. This model captures fine-grained evolving dependencies at each timestamp and integrates these insights into subsequent modeling stages. HEAG leverages a heterogeneous attention mechanism to aggregate fine-grained temporal information from multiple nodes on the graph. Additionally, our approach includes an FDG that relies on the Graph Attention (GAT) Mechanism. The FDG allows us to discern the interdependencies among various sequences at individual time nodes, capturing the patterns and magnitudes of variations within each sequence. Lastly, we deploy a Coarse-grained Dependency Graph generator to model temporal information across the entire timeline. This facilitates comprehensive modeling of the time series and enhances the understanding of mutual dependencies among different sequences.
In conclusion, the contributions of our work can be summarized as follows:
  • In response to the limitations of traditional methods for predicting the health status of lithium batteries, we have introduced the HEAG model. This model utilizes hybrid-granularity time scales to significantly enhance predictive accuracy.
  • To address the complex interplay of coarse-grained and fine-grained temporal dependencies within multiple input sequences more effectively, we have developed the FDG and the CDG. These tools help analyze the mutual dependencies across individual time points and entire time windows, offering new perspectives on heterogeneous correlation modeling among sequences.
  • We have conducted extensive experiments using two publicly available datasets to demonstrate that the HEAG model can more accurately discern the correlations among various input sequences. This ability leads to improved predictions of the SOH of lithium batteries.

2. Related Works

2.1. Traditional Methods

Traditional methods for predicting the SOH of batteries are primarily classified into two categories: (1) filter-based techniques, and (2) machine learning approaches.
Filter-based models employ various techniques, including the Kalman filter [12,13], particle filter [14,15], and others. For instance, particle filters estimate SOH using stochastic models [16], while novel unbiased Kalman filters also assess battery SOH [17]. Additionally, extended Kalman filters with Lebesgue sampling [18] and adaptive methods considering parameter uncertainty using the Wiener process [19] have been proposed. Although these filter methods have significant advantages in self-correlation abilities, they have limitations when handling large datasets, especially in predicting the SOH of multiple battery cells [20].
In contrast, machine learning methods [21,22] are more effective at handling complex data and have been widely adopted for predicting battery SOH [23]. For example, ensemble learning [24], Gaussian process regression [25], and support vector regression [26] can all predict the SOH of lithium-ion batteries. Experimental results show that machine learning methods outperform filter-based approaches in terms of accuracy. However, machine learning also faces certain limitations. It is less capable of handling massive datasets compared to deep learning and relies on manually extracted features, complicating preprocessing and limiting the models’ generalizability in real-world applications [27].

2.2. Deep Learning-Based Methods

To effectively utilize the historical information from each battery unit, deep learning algorithms have been increasingly adopted [1,28]. These include CNN [29], RNN [30], LSTM [31], TCN [32], and Transformers [33].
CNN and LSTM combinations can identify battery aging features [34], and RNN models estimate the SOH of high energy density batteries [35]. You et al. designed an RNN-based degradation model for real-world driving data [36].
The BLS-LSTM hybrid model combines traditional statistical methods with neural networks, promoting battery health monitoring and energy management [37]. Optimized sliding windows and steady-state filtering reduce redundancy [38]. Multi-channel LSTM enhances flexibility in lifespan prediction, showcasing deep learning’s advantages in handling complex patterns and temporal dependencies [39]. Combining LSTM with Gaussian processes [40] or time series forecasting [41] leverages probabilistic and temporal modeling strengths. TCN addresses long-term dependencies, providing more accurate battery health predictions [42,43]. Integrating empirical mode decomposition with TCN reduces interference and enhances SOH prediction robustness [44]. Despite higher predictive accuracy than traditional methods, LSTM and TCN may overlook dynamic relationships between input sequences and SOH, leading to biases.
Transformers capture extensive dependencies and improve computational efficiency, advancing sequence modeling. Scholars like Chen [45] use denoising autoencoders to preprocess data for Transformer networks, while Luo [46] identified electrochemical impedance spectroscopy as a novel feature for Transformer-based SOH estimation. The Exponential Smoothing Transformer with Seasonal and Growth Embedding (SGEformer) has been proposed for predicting SOH [47]. These methods reflect the trend of integrating and advancing machine learning and deep learning for battery health assessment. However, Transformers, though adept at learning fine-grained inter-sequence relationships, face challenges in terms of high computational complexity and struggle to model coarse-grained relationships effectively.

2.3. Graph-Based Methods

While deep learning has demonstrated proficiency in handling complex data, it often struggles to unveil the inter-feature correlations essential for enhancing predictive model accuracy and robustness. Graph Convolutional Networks (GCNs) have gained prominence for their ability to analyze data relationships through undirected graphs, thus predicting SOH [48]. Derived models like MGCN [49] tackle complex network analysis in SOH prediction; GCN-DA [50] improves accuracy under multifactorial influences; CL-GraphSAGE [51] boosts scalability and efficiency for extensive data analysis; and CGGN-DCO [52] enhances efficiency in capturing complex temporal and spatial patterns. However, challenges such as MGCN’s high computational demand, GCN-DA’s cumbersome adaptation process, CL-GraphSAGE’s potential information loss, and CGGN-DCO’s complex training process may impede effective learning of health indicators tracking battery degradation. Our paper introduces the HEAG model, integrating coarse and fine-grained analyses for SOH prediction, achieving satisfactory results in computational complexity and predictive accuracy.

3. Methods

3.1. Problem Formulation

This study provides a tabulated summary and notation of key symbols in Table 1. We model the health state of lithium-ion batteries as a graph M = ( Q , E ) , with Q embodying N factors influencing battery life, interconnected by E . Each factor q within Q has a dynamic input x = { x ( 1 ) , x ( 2 ) , , x ( t ) } , representing values over time. The variable t represents the cycle, which is the number of charge–discharge cycles of the battery. Each cycle impacts the SOH of the battery. Therefore, the SOH prediction problem is defined as a time series prediction problem because the SOH changes over time and with the number of cycles, with each cycle depending on the previous ones. Continuous cycles form a time series. After feature engineering, the refined input sequence is X R L × J , where L is the sequence length, and J is the number of dimensions at each time slot.
In our study, we delve into single-step prediction, which predicts the immediate future state using a historical data window. For input data X with a sliding window size s, we generate sub-sequences X ( t s + 1 : t ) of length L s and use the subsequent value x ( t + 1 ) as the target output y. Our goal is to craft a parameterized function F , which maps these sequences to future states, thereby enhancing prediction accuracy.

3.2. Model Architecture

The HEAG model’s structure, depicted in Figure 2, comprises four integral components: Graph and Time Series Processing, a Fine-grained Dependency Graph (FDG), a Coarse-grained Dependency Graph (CDG), and the Prediction module. The model employs time series processing to distill comprehensive features, setting the stage for advanced feature extraction. Through the intermediary steps involving the FDG and the CDG, the model culminates in the Prediction module, where it learns to project the feature vector v ( t + 1 ) onto the forecasted value y ^ ( t + 1 ) .

3.2.1. Feature Engineering

We derived two continuous features from each univariate sequence: (1) stability, represented by a normalized series, and (2) trend, depicted by a rate of change series. The computations for the normalized series and the rate of change series are detailed in Equation (1).
x ( i ) = x ( i ) min ( x ( t s + 1 : t ) ) max ( x ( t s + 1 : t ) ) min ( x ( t s + 1 : t ) ) , x ( i ) = ( x ( i ) x ( i 1 ) ) / x ( i 1 ) .

3.2.2. Fine-Grained Evolving Aware Graph

In this section, we discuss efficient extraction methods for precise, fine-grained temporal dependencies using graph structures.
Given the flow sequence of the target influencing factors X R L × J and the adjacent sequence Z R K × L × J , K + 1 independent FDG units are initialized for each sequence, with no parameter sharing between the FDG units. The architecture of the FDG unit is shown in Figure 3. Specifically, at timestamp t, X ( t ) and Z ( t ) are inputted into the first cell of their respective encoders. For instance, mapping ( x ( t ) , x ( t ) , x ( t ) ) to a high-dimensional space generates a representation vector ( t ) with size J. The state vector e ( 1 ) R 1 × D is randomly initialized at the first timestamp and continuously updated throughout the sequence. The temporally aware representation v ( t 1 ) at the last timestamp is also considered as one of the inputs at timestamp t. By following the recurrent units based on the graph in Equation (2), the state vector is updated to e ( t ) at each timestamp and is prepared to enter the next cell. The temporal information of this input information flow at this timestamp is represented by the vector o ( t ) R 1 × D .
i ( t ) = σ ( W i T [ v ( t 1 ) , o ^ ( t ) ] + b i ) , f ( t ) = σ ( W f T [ v ( t 1 ) , o ^ ( t ) ] + b f ) , e ^ ( t ) = tanh ( W e T [ v ( t 1 ) , o ^ ( t ) ] + b e ) , g ( t ) = σ ( W g T [ v ( t 1 ) , o ^ ( t ) ] + b g ) , e ( t ) = f ( t ) e ( t 1 ) + i ( t ) e ^ ( t ) , o ( t ) = g ( t ) tanh ( e ( t ) ) .
Here, i ( t ) is the input gate, f ( t ) is the forget gate, and g ( t ) is the output gate. The learnable parameters at each gate, W i , W f , W e , W g R J × 2 J , are accompanied by biases b i , b f , b e , b g . The symbol ⊙ indicates the element-wise product. After extracting temporal information at both neighbor and target encoder sides, the temporal representations of neighbor influencing factors p ( t ) R K × D and o ( t ) are inputted into the GAT module, which will be discussed in the next paragraph.
The Graph Attention (GAT) Mechanism was proposed to address the shortcomings of the Graph Convolutional Network (GCN), which struggles with inductive tasks on dynamic graphs and with capturing temporal dependencies due to its static topology. To capture fine-grained temporal dependencies and sustain influence from relevant influencing factors, the time attention score α i ( t ) between the target influencing factors and each relevant influencing factor is calculated and updated at each timestamp. Given the time representations of the target influencing factors and relevant influencing factors, the mechanism for information aggregation is depicted in Equation (3):
h i ( t ) = W q T g a ( t ) × W k T p i ( t ) , α i ( t ) = h i ( t ) i = 1 K h i ( t ) , v ( t ) = R e L U ( W f u s e 1 T [ o ( t ) , i = 1 K α i ( t ) W v T p i ( t ) ] ) .
In Equation (3), W k is the learnable weight of the key mapping function, W q is the weight of the query function, and W v is the weight of the value function. After a linear transform by W k , the representation matrix of neighbor influencing factors is queried by the query vector, which is transformed from the representation of the target influencing factors. The attention scores facilitate the weighted aggregation of information from these neighbors. Following this, a fusion layer updates the temporal representation v ( t ) of the target influencing factors. Post-aggregation, the temporal information at timestamp t of the target influencing factors is also shared with neighboring factors, as detailed in Equation (4), where W f u s e 2 denotes the weight of a new fusion layer.
p i ( t ) = R E L U ( W f u s e 2 T [ o ( t ) , p i ( t ) ] )
The message passing operation is completed on each neighbor influencing factor to update their representations. Then, the updated temporal representation vectors of influencing factors at the present timestamp are inputted to the FDGNet cell for the next timestamp until the whole sequence is modeled.

3.2.3. Coarse-Grained Dependency Graph

Acknowledging the inevitable loss of global information during the recurrent modeling process, the resulting fine-grained representation v proves inadequate. Previous studies have highlighted that fixed adjacency matrices might not accurately reflect the true connections between influencing factors. To address these issues, we have developed an implicit Coarse-grained Dependency Graph (CDG) generator that captures evolving global dependency information. Additionally, due to the continuous nature of the time series, simplistic pooling or naive addition of representations may either omit critical information captured in prior steps or introduce redundancy into the feature space. As illustrated in Figure 4, for representations at all time steps of both neighbor and target influencing factors p and v , we employ an addition operation with a computational attenuation coefficient γ , effectively aggregating information into space E R ( K + 1 ) × D . With v as an example, the process is formulated as follows:
γ ( t ) = 1 s t , E = concat ( t s γ ( t ) v ( t ) , t s γ ( t ) p 1 ( t ) , , t s γ ( t ) p N ( t ) ) .
Then, we apply a similarity matrix θ , where the similarity between each influencing factor θ i j is measured by the inner product. Through a non-parameterized softmax function, a sparse adjacency matrix A t is computed as follows:
A t = σ ( ( log ( θ i j / ( 1 θ i j ) ) .
Here, σ represents the activation function. Through the CDG generator, an auxiliary implicit graph is formed, capturing the global temporal representation. During each training iteration, A t is updated to maintain statistical alignment with the evolving dependency matrix, ensuring continual optimization post training. At the final time step t, the representations of v ( t ) and p ( t ) , along with A t , are fed into the graph convolutional network, updating the target influencing factors’ representation to v ( t ) . This process facilitates achieving mixed granularity.

3.3. Result Prediction and Loss Function

The module utilizes a Multi-Layer Perceptron (MLP) for result prediction and loss analysis to generate predictions at time y ^ at t + 1 . Specifically, the module consists of two linear neural layers, with a non-linear activation function and a dropout operation deployed between them. Through this MLP block, the output vector v ( t ) from the target influencing factor encoder is transformed into the prediction value at time y ^ ( t + 1 ) at t + 1 . Considering the robustness of HEAG in regression tasks, the smooth L1 loss is chosen as the loss function, defined as follows:
L = 0.5 × ( y y ^ ) 2 , i f   | y y ^ |   < 0.5 | y y ^ | 0.5 , o t h e r w i s e .

4. Experiments

4.1. Battery Datasets

To enhance the credibility of the experimental results, we selected two publicly available datasets used in baseline methods for evaluation, ensuring that the samples were strictly chosen according to comparison criteria. These datasets were released by two different research institutions at different times. The dataset information statistics are shown in Table 2.

4.1.1. NASA Dataset

The HEAG model’s efficacy was validated using a battery dataset from NASA’s Ames Prognostics Center of Excellence [53], as cited in recent studies [50,54,55]. This dataset is divided into three subsets, with each subset encompassing specific batteries: Subset 1 features data from batteries numbered 5, 6, 7, and 18; Subset 2 from batteries 29, 30, 31, and 32; and Subset 3 from batteries 25, 26, and 27. These batteries were tested under three operational conditions: charging, discharging, and impedance, with detailed recordings of current, voltage, and temperature during operation. The charging protocol involved constant current (CC) charging at 1.5 A until the voltage hit 4.2 V, followed by constant voltage (CV) charging until the current fell to below 20 mA. The discharge settings varied slightly across experiments. The NASA dataset includes various features such as cycle number, ambient temperature, date/time, capacity, and the measurements of battery terminal voltage, battery output current, and battery temperature for each cycle, all of which are used as inputs for the model. The primary output from this model is the SOH. The repeated charging and discharging cycles cause accelerated degradation and capacity loss in lithium-ion batteries, as depicted in Figure 5, which illustrates the capacity degradation trajectories across the subsets.

4.1.2. CALCE Dataset

The CALCE dataset is frequently utilized for research on SOH estimation [56,57]. It encompasses six subsets, known as CS2, which are differentiated by various cycle aging processes, such as CC–CV charging and CC discharge. In the charging stage, batteries charge under a constant current of 0.5 C until reaching a voltage of 4.2 V, and then they continue charging at this voltage until the current falls below 0.05 A. For discharging, the batteries discharge at various constant currents until the voltage falls to 2.7 V. This study focuses on four batteries within the dataset (CS2-35, CS2-36, CS2-37, CS2-38), all of which discharge at a constant rate of 1 C. These were examined using an Arbin battery tester. Within the CALCE dataset, the model inputs include date/time, cycle, capacity, resistance, CCCT (Constant Current Charging Time), and CVCT (Constant Voltage Charging Time), with the model providing predictions over a span of 72 cycles and providing output for 16 cycles.

4.2. Evaluation Metrics

To fairly assess the model performance, RMSE (Root Mean Square Error), MAE (Mean Absolute Error), and MedAE (Median Absolute Error) are utilized to measure the discrepancy between the actual values and the predicted values. Let y n , l denote the ground truth and y ^ n , l represent the predicted value. The definitions of these three metrics are as follows:
M A E = 1 N L n = 1 N l = 1 L y n , l y ^ n , l ,
R M S E = 1 N L n = 1 N l = 1 L y n , l y ^ n , l 2 ,
MedAE = median ( | y n , l y ^ n , l | ) .

4.3. Baseline Methods

To rigorously assess the efficacy of our novel model across diverse applications, we meticulously curated an assortment of both traditional and state-of-the-art benchmark methods for a comprehensive comparative study. These methodologies are broadly classified into sequence and graph-based models.
For sequence analysis, our arsenal included ARIMA [58] for forecasting time series data, linear regression (LR) [59] for tackling classification and regression challenges, GRU [60] for the intricate modeling and forecasting of sequential data, and the hybrid CNN+LSTM [34] model that fuses the strengths of CNNs and LSTMs for enhanced sequence data analysis. Furthermore, we explored the potential of TCN [43] and Transformer [61] models for advanced sequence-to-sequence learning, with the Transformer model employing a sophisticated self-attention mechanism for superior sequence data handling. The State-of-Health Prediction of Lithium-Ion Batteries using SGEformer [47] model was recently proposed, which integrates seasonal and growth embedding layers into the Transformer architecture, enhancing its ability to handle time-series data specifically tailored for battery life prediction.
Transitioning to graph-based models, we harnessed MGCN [49] to capture features at different scales by introducing multiple receptive fields and integrating these into a more robust feature representation, effectively enhancing feature representation and relationship mining among signals. Additionally, we referenced the GCN-DA [62] model, which incorporates a dual attention mechanism into a graph convolutional network to enhance the extraction and utilization of relevant features from time-series data. To further refine the quality of graph-influencing factor embeddings, we utilized CL-GraphSAGE [51], which leverages the strengths of both CNN and LSTM to extract temporal features, while the GraphSAGE component addresses spatial interdependencies. Additionally, the CGGN-DCO [52] model enhances the ability to capture temporal dependencies between features by integrating gated graph neural networks with dynamic convolutional operations.

4.4. Implementation Settings

During the implementation phase, the Adam optimizer was employed with a constant learning rate of 0.0001, and a batch size of 64 was established. The model underwent training over 20 epochs, with the hidden vector dimensionality steadfastly set at 64. To curb the risk of overfitting, a dropout rate of 0.1 was instituted. For the selection of the most effective parameter fusion, a grid search approach was harnessed, meticulously calibrating parameters within the validation dataset. This calibration included K values in the set [ 2 , 4 , 6 , 8 ] , as well as s values of [ 11 , 16 ] .
All experimental procedures were executed on a 64-bit Linux server, outfitted with an NVIDIA GeForce GTX 3090 GPU. We developed HEAG and the comparison baseline models utilizing Python 3.8 and PyTorch 1.7.0. Each experimental iteration was performed five times to ensure the statistical robustness of our findings.
For models lacking available official code in subsequent sections, we took the initiative to craft implementations independently; for those with accessible official code, we utilized it directly. To guarantee equitable comparisons, we meticulously tuned the parameters for all models, targeting their peak performance levels.

4.5. Comparison Results and Analysis

As shown in Table 3 and Table 4, in a detailed comparative analysis of SOH prediction models across the CALCE and NASA datasets, the HEAG distinguishes itself with exceptional performance. This model achieves the lowest RMSE and MAE, indicative of its advanced predictive abilities for battery health. Traditional methods like ARIMA lack the adaptability to capture intricate degradation patterns due to their linear nature and fall short in handling the nonlinearities present in battery usage data.
LR and GRU show improvements over traditional statistical methods but still face challenges with complex temporal sequences. These machine learning models often require expert domain knowledge for feature engineering, which is labor-intensive and which limits scalability.
Deep learning models, including CNN combined with LSTM, TCN, and Transformers, enhance the ability to understand sequential dependencies. They leverage the temporal information contained within the input data more effectively than traditional machine learning models. Despite this, their prediction accuracy, while better, is not as consistent or high-reaching as that of HEAG.
HEAG stands out not only against traditional deep learning approaches but also when compared to newer models such as SGEformer, MGCN, GCN-DA, CL-GraphSAGE, and CGGN-DCO. HEAG’s nuanced approach to temporal dependencies, utilizing both coarse- and fine-grained insights, enables a more sophisticated capture of SOH patterns in lithium-ion batteries.
HEAG’s robustness stems from its comprehensive approach to temporal analysis, blending the attention to detail seen in fine-grained analysis with the overarching patterns recognized by coarse-grained analysis. This dual approach allows the HEAG model to understand the immediate fluctuations and the long-term trends within the battery’s discharge cycles. This level of detail enables HEAG to capture both sudden, transient changes and long-term degradation patterns, enriching the model’s comprehension of battery health and its predictive precision.
In conclusion, while the deep learning models show significant potential, the hybrid approach taken by HEAG, with its ability to synthesize multi-scale temporal patterns, sets a new standard for accuracy and reliability in SOH prediction for lithium-ion batteries.

4.6. Ablation Study

For our comprehensive evaluation of the HEAG model, we placed special emphasis on the vital roles played by fine-grained and coarse-grained information in influencing the model’s performance metrics. To dissect these influences with greater precision, we conducted two key ablation experiments: one removing fine-grained information (labeled as HEAG w/o fine) and another eliminating coarse-grained information (labeled as HEAG w/o coarse). The purpose of these experiments was to delineate the individual and combined effects of these types of information on the model’s predictive accuracy.
Insights drawn from Figure 6a,b, corresponding to the NASA and CALCE datasets, respectively, emphasize the significance of both types of information. Figure 6a clearly displays a performance reduction in key metrics such as RMSE, MAE, and MedAE by 17.5%, 23.5%, and 28.7%, respectively, on the NASA dataset following the removal of fine-grained and coarse-grained information. Similarly, Figure 6b, which focuses on the CALCE dataset, reveals decreases of 19.7%, 24.9%, and 20.1% in these metrics after the same omissions.
Our extensive analysis across these datasets not only confirms the negative impact of omitting any type of information on model performance but also highlights the model’s refined response to different dataset characteristics. This in-depth examination not only reaffirms the resilience and adaptability of the HEAG model but also provides deep insights into how fine-grained and coarse-grained information differentially and collectively influence model performance across various scenarios. Therefore, our study not only showcases the HEAG model’s strengths in complex predictive tasks but also establishes a detailed framework for future research to delve into the complex effects of model components on performance outcomes.

4.7. Case Study: Cross-Battery Evaluation

We utilized the CALCE dataset for cross-battery evaluation, involving four batteries. One battery was designated as the target test set, while the other three batteries were trained together. This process was repeated four times, with each battery serving as the target test set once. As shown in Figure 7a,b, the experimental results produced four graphs. By observing these graphs, we derived several observations.
Shape Analysis: The plots Figure 7a–d display the estimated SOH against the ground truth across the discharge cycles for four batteries. A consistent pattern is observed, with a plateau in the early cycles followed by a steeper decline, suggesting a characteristic battery degradation curve. However, some plots show divergence in this pattern, indicating potential improvements in capturing the lifecycle nuances of individual batteries.
Bias and Deviation: All graphs indicate some degree of bias, where the estimated SOH either overestimates or underestimates the ground truth. The deviation is more pronounced as the batteries near the end of their lifecycle, highlighting a challenge in accurately predicting the steepness and timing of SOH decline.
Anomaly Detection Ability: The predictive model’s ability to capture anomalous fluctuations is inconsistent. Initial fluctuations in SOH estimation, particularly in plot Figure 7d, reveal a sensitivity to anomalies, yet the model does not consistently detect all abrupt changes, suggesting a need for refined modeling techniques to improve prediction accuracy and reliability.
Overall, while the model shows promise in generalizing across different batteries, the case study reveals limitations in shape adherence, bias accuracy, and anomaly detection, guiding future model enhancements.

5. Conclusions

In this paper, we present a novel evolving aware graph model designed to predict the health status of lithium-ion batteries. This model, known as HEAG, integrates both Fine-grained Dependency Graph (FDG) and Coarse-grained Dependency Graph (CDG) components to discern temporal dependencies at various granularities. Specifically, the FDG captures dependencies between sequences at individual time points, while the CDG perceives dependencies across entire time series. Extensive experiments were conducted using publicly available datasets containing discharge cycle data such as current, voltage, temperature, and other parameters. Prior to experimentation, feature engineering was employed to optimize efficiency and accelerate convergence. The results demonstrate that our method accurately identifies dependencies between different time series and predicts the SOH with high precision. Compared to the baseline methods and four new methods discussed in this paper, our approach shows superior performance in terms of RMSE, MAE, and MedAE. Additionally, the necessity of each component was validated through ablation studies, and cross-evaluation case studies of the batteries yielded favorable outcomes. Moving forward, the effectiveness of the proposed method will be assessed using various types of lithium-ion battery datasets to confirm its broader applicability.

Author Contributions

Methodology, C.X.; resources, Z.Z.; software, C.X.; supervision, J.W. (Jun Wang) and J.W. (Jiyao Wang); visualization, H.L.; writing—original draft, C.X.; writing—review & editing, J.W. (Jiyao Wang). All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The NASA and CALCE datasets used in this study can be accessed at https://www.nasa.gov/content/prognostics-center-of-excellence-data-set-repository and https://calce.umd.edu/battery-data, respectively, with access dates of 16 January 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SOHState-of-health
FDGFine-grained Dependency Graph
CDGCoarse-grained Dependency Graph
HEAGHybrid-grained Evolving Aware Graph
CNNConvolutional Neural Network
RNNRecurrent Neural Network
GRUGated Recurrent Network
LSTMLong Short-term Memory
GATGraph Attention
TCNTemporal Convolutional Network
CCConstant Current
CVConstant Voltage
MLPMulti-Layer Perceptron

References

  1. Xu, F.; Yang, F.; Fei, Z.; Huang, Z.; Tsui, K.L. Life prediction of lithium-ion batteries based on stacked denoising autoencoders. Reliab. Eng. Syst. Saf. 2021, 208, 107396. [Google Scholar] [CrossRef]
  2. Attia, P.M.; Grover, A.; Jin, N.; Severson, K.A.; Markov, T.M.; Liao, Y.H.; Chen, M.H.; Cheong, B.; Perkins, N.; Yang, Z.; et al. Closed-loop optimization of fast-charging protocols for batteries with machine learning. Nature 2020, 578, 397–402. [Google Scholar] [CrossRef] [PubMed]
  3. He, W.; Williard, N.; Chen, C.; Pecht, M. State of charge estimation for Li-ion batteries using neural network modeling and unscented Kalman filter-based error cancellation. Int. J. Electr. Power Energy Syst. 2014, 62, 783–791. [Google Scholar] [CrossRef]
  4. Yao, L.; Wen, J.; Xu, S.; Zheng, J.; Hou, J.; Fang, Z.; Xiao, Y. State of health estimation based on the long short-term memory network using incremental capacity and transfer learning. Sensors 2022, 22, 7835. [Google Scholar] [CrossRef] [PubMed]
  5. Christensen, G.; Younes, H.; Hong, H.; Widener, C.; Hrabe, R.H.; Wu, J.J. Nanofluids as media for high capacity anodes of lithium-ion battery—A review. J. Nanofluids 2019, 8, 657–670. [Google Scholar] [CrossRef]
  6. Zeng, J.; Liu, S. Research on aging mechanism and state of health prediction in lithium batteries. J. Energy Storage 2023, 72, 108274. [Google Scholar] [CrossRef]
  7. Plett, G.L. Extended Kalman filtering for battery management systems of LiPB-based HEV battery packs: Part 1. Background. J. Power Sources 2004, 134, 252–261. [Google Scholar] [CrossRef]
  8. Severson, K.A.; Attia, P.M.; Jin, N.; Perkins, N.; Jiang, B.; Yang, Z.; Chen, M.H.; Aykol, M.; Herring, P.K.; Fraggedakis, D.; et al. Data-driven prediction of battery cycle life before capacity degradation. Nat. Energy 2019, 4, 383–391. [Google Scholar] [CrossRef]
  9. Ma, G.; Zhang, Y.; Cheng, C.; Zhou, B.; Hu, P.; Yuan, Y. Remaining useful life prediction of lithium-ion batteries based on false nearest neighbors and a hybrid neural network. Appl. Energy 2019, 253, 113626. [Google Scholar] [CrossRef]
  10. Gao, X.; Dong, X.; Xing, Z.; Jamila, S.; Hong, H.; Jiang, H.; Zhang, J.; Ju, Z. Ether-based electrolytes enable the application of nitrogen and sulfur co-doped 3D graphene frameworks as anodes in high-performance sodium-ion batteries. Nanoscale 2023, 15, 1568–1582. [Google Scholar] [CrossRef]
  11. Xu, Y.; Wu, S.; Wang, B.; Yang, M.; Wu, Z.; Yao, Y.; Wei, Z. Two-stage fine-grained image classification model based on multi-granularity feature fusion. Pattern Recognit. 2024, 146, 110042. [Google Scholar] [CrossRef]
  12. Andre, D.; Appel, C.; Soczka-Guth, T.; Sauer, D.U. Advanced mathematical methods of SOC and SOH estimation for lithium-ion batteries. J. Power Sources 2013, 224, 20–27. [Google Scholar] [CrossRef]
  13. Xu, W.; Wang, S.; Jiang, C.; Fernandez, C.; Yu, C.; Fan, Y.; Cao, W. A novel adaptive dual extended Kalman filtering algorithm for the Li-ion battery state of charge and state of health co-estimation. Int. J. Energy Res. 2021, 45, 14592–14602. [Google Scholar] [CrossRef]
  14. Liu, D.; Yin, X.; Song, Y.; Liu, W.; Peng, Y. An on-line state of health estimation of lithium-ion battery using unscented particle filter. IEEE Access 2018, 6, 40990–41001. [Google Scholar] [CrossRef]
  15. Schwunk, S.; Armbruster, N.; Straub, S.; Kehl, J.; Vetter, M. Particle filter for state of charge and state of health estimation for lithium–iron phosphate batteries. J. Power Sources 2013, 239, 705–710. [Google Scholar] [CrossRef]
  16. Dong, G.; Chen, Z.; Wei, J.; Ling, Q. Battery health prognosis using Brownian motion modeling and particle filtering. IEEE Trans. Ind. Electron. 2018, 65, 8646–8655. [Google Scholar] [CrossRef]
  17. Chen, N.; Hu, X.; Gui, W.; Zou, J. Estimation of li-ion battery state of charging and state of healthy based on unsented Kalman filtering. In Proceedings of the The 26th Chinese Control and Decision Conference (2014 CCDC), Changsha, China, 31 May–2 June 2014; pp. 4725–4729. [Google Scholar]
  18. Yan, W.; Zhang, B.; Zhao, G.; Tang, S.; Niu, G.; Wang, X. A battery management system with a Lebesgue-sampling-based extended Kalman filter. IEEE Trans. Ind. Electron. 2018, 66, 3227–3236. [Google Scholar] [CrossRef]
  19. Zhang, J.; Jiang, Y.; Li, X.; Huo, M.; Luo, H.; Yin, S. An adaptive remaining useful life prediction approach for single battery with unlabeled small sample data and parameter uncertainty. Reliab. Eng. Syst. Saf. 2022, 222, 108357. [Google Scholar] [CrossRef]
  20. Thelen, A.; Li, M.; Hu, C.; Bekyarova, E.; Kalinin, S.; Sanghadasa, M. Augmented model-based framework for battery remaining useful life prediction. Appl. Energy 2022, 324, 119624. [Google Scholar] [CrossRef]
  21. Greenbank, S.; Howey, D.A. Piecewise-linear modelling with automated feature selection for Li-ion battery end-of-life prognosis. Mech. Syst. Signal Process. 2023, 184, 109612. [Google Scholar] [CrossRef]
  22. Luo, J.; Fang, S.C.; Deng, Z.; Tian, Y. Robust kernel-free support vector regression based on optimal margin distribution. Knowl.-Based Syst. 2022, 253, 109477. [Google Scholar] [CrossRef]
  23. He, J.; Tian, Y.; Wu, L. A hybrid data-driven method for rapid prediction of lithium-ion battery capacity. Reliab. Eng. Syst. Saf. 2022, 226, 108674. [Google Scholar] [CrossRef]
  24. Gou, B.; Xu, Y.; Feng, X. State-of-health estimation and remaining-useful-life prediction for lithium-ion battery using a hybrid data-driven method. IEEE Trans. Veh. Technol. 2020, 69, 10854–10867. [Google Scholar] [CrossRef]
  25. Richardson, R.R.; Osborne, M.A.; Howey, D.A. Gaussian process regression for forecasting battery state of health. J. Power Sources 2017, 357, 209–219. [Google Scholar] [CrossRef]
  26. Li, X.; Yuan, C.; Wang, Z. State of health estimation for Li-ion battery via partial incremental capacity analysis based on support vector regression. Energy 2020, 203, 117852. [Google Scholar] [CrossRef]
  27. Akbar, K.; Zou, Y.; Awais, Q.; Baig, M.J.A.; Jamil, M. A machine learning-based robust state of health (soh) prediction model for electric vehicle batteries. Electronics 2022, 11, 1216. [Google Scholar] [CrossRef]
  28. Chang, C.; Wang, Q.; Jiang, J.; Wu, T. Lithium-ion battery state of health estimation using the incremental capacity and wavelet neural networks with genetic algorithm. J. Energy Storage 2021, 38, 102570. [Google Scholar] [CrossRef]
  29. Al-Dulaimi, A.; Zabihi, S.; Asif, A.; Mohammadi, A. A multimodal and hybrid deep neural network model for remaining useful life estimation. Comput. Ind. 2019, 108, 186–196. [Google Scholar] [CrossRef]
  30. Lu, J.; Xiong, R.; Tian, J.; Wang, C.; Hsu, C.W.; Tsou, N.T.; Sun, F.; Li, J. Battery degradation prediction against uncertain future conditions with recurrent neural network enabled deep learning. Energy Storage Mater. 2022, 50, 139–151. [Google Scholar] [CrossRef]
  31. Yayan, U.; Arslan, A.T.; Yucel, H. A novel method for SoH prediction of batteries based on stacked LSTM with quick charge data. Appl. Artif. Intell. 2021, 35, 421–439. [Google Scholar] [CrossRef]
  32. Zhou, D.; Wang, B. Battery health prognosis using improved temporal convolutional network modeling. J. Energy Storage 2022, 51, 104480. [Google Scholar] [CrossRef]
  33. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language models are unsupervised multitask learners. OpenAI blog 2019, 1, 9. [Google Scholar]
  34. Li, P.; Zhang, Z.; Grosu, R.; Deng, Z.; Hou, J.; Rong, Y.; Wu, R. An end-to-end neural network framework for state-of-health estimation and remaining useful life prediction of electric vehicle lithium batteries. Renew. Sustain. Energy Rev. 2022, 156, 111843. [Google Scholar] [CrossRef]
  35. Eddahech, A.; Briat, O.; Bertrand, N.; Delétage, J.Y.; Vinassa, J.M. Behavior and state-of-health monitoring of Li-ion batteries using impedance spectroscopy and recurrent neural networks. Int. J. Electr. Power Energy Syst. 2012, 42, 487–494. [Google Scholar] [CrossRef]
  36. You, G.W.; Park, S.; Oh, D. Diagnosis of electric vehicle batteries using recurrent neural networks. IEEE Trans. Ind. Electron. 2017, 64, 4885–4893. [Google Scholar] [CrossRef]
  37. Zhao, S.; Zhang, C.; Wang, Y. Lithium-ion battery capacity and remaining useful life prediction using board learning system and long short-term memory neural network. J. Energy Storage 2022, 52, 104901. [Google Scholar] [CrossRef]
  38. Wang, S.; Takyi-Aninakwa, P.; Jin, S.; Yu, C.; Fernandez, C.; Stroe, D.I. An improved feedforward-long short-term memory modeling method for the whole-life-cycle state of charge prediction of lithium-ion batteries considering current-voltage-temperature variation. Energy 2022, 254, 124224. [Google Scholar] [CrossRef]
  39. Park, K.; Choi, Y.; Choi, W.J.; Ryu, H.Y.; Kim, H. LSTM-based battery remaining useful life prediction with multi-channel charging profiles. IEEE Access 2020, 8, 20786–20798. [Google Scholar] [CrossRef]
  40. Zhao, J.; Zhu, Y.; Zhang, B.; Liu, M.; Wang, J.; Liu, C.; Zhang, Y. Method of Predicting SOH and RUL of Lithium-Ion Battery Based on the Combination of LSTM and GPR. Sustainability 2022, 14, 11865. [Google Scholar] [CrossRef]
  41. Audin, P.; Jorge, I.; Mesbahi, T.; Samet, A.; De Beuvron, F.D.B.; Boné, R. Auto-encoder LSTM for Li-ion SOH prediction: A comparative study on various benchmark datasets. In Proceedings of the 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), Pasadena, CA, USA, 13–16 December 2021; pp. 1529–1536. [Google Scholar]
  42. Zhen, Y.; Fang, J.; Zhao, X.; Ge, J.; Xiao, Y. Temporal convolution network based on attention mechanism for well production prediction. J. Pet. Sci. Eng. 2022, 218, 111043. [Google Scholar] [CrossRef]
  43. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  44. Zhou, D.; Li, Z.; Zhu, J.; Zhang, H.; Hou, L. State of health monitoring and remaining useful life prediction of lithium-ion batteries based on temporal convolutional network. IEEE Access 2020, 8, 53307–53320. [Google Scholar] [CrossRef]
  45. Chen, D.; Hong, W.; Zhou, X. Transformer network for remaining useful life prediction of lithium-ion batteries. IEEE Access 2022, 10, 19621–19628. [Google Scholar] [CrossRef]
  46. Luo, K.; Zheng, H.; Shi, Z. A simple feature extraction method for estimating the whole life cycle state of health of lithium-ion batteries using transformer-based neural network. J. Power Sources 2023, 576, 233139. [Google Scholar] [CrossRef]
  47. Fauzi, M.R.; Yudistira, N.; Mahmudy, W.F. State-of-Health Prediction of Lithium-Ion Batteries using Exponential Smoothing Transformer with Seasonal and Growth Embedding. IEEE Access 2024, 12, 14659–14670. [Google Scholar] [CrossRef]
  48. Wei, Y.; Wu, D. Remaining useful life prediction of bearings with attention-awared graph convolutional network. Adv. Eng. Inform. 2023, 58, 102143. [Google Scholar] [CrossRef]
  49. Li, T.; Zhao, Z.; Sun, C.; Yan, R.; Chen, X. Multireceptive field graph convolutional networks for machine fault diagnosis. IEEE Trans. Ind. Electron. 2020, 68, 12739–12749. [Google Scholar] [CrossRef]
  50. Wei, Y.; Wu, D. Prediction of state of health and remaining useful life of lithium-ion battery using graph convolutional network with dual attention mechanisms. Reliab. Eng. Syst. Saf. 2023, 230, 108947. [Google Scholar] [CrossRef]
  51. Yao, X.Y.; Chen, G.; Pecht, M.; Chen, B. A novel graph-based framework for state of health prediction of lithium-ion battery. J. Energy Storage 2023, 58, 106437. [Google Scholar] [CrossRef]
  52. Wei, Y.; Wu, D. State of health and remaining useful life prediction of lithium-ion batteries with conditional graph convolutional network. Expert Syst. Appl. 2024, 238, 122041. [Google Scholar] [CrossRef]
  53. Saha, B.; Goebel, K. Battery Data Set; NASA AMES Prognostics Data Repository; NASA Ames Research Center: Moffett Field, CA, USA, 2007. [Google Scholar]
  54. Zhou, J.; Wang, S.; Cao, W.; Xie, Y.; Fernandez, C. State of Health Prediction of Lithium-ion Batteries Based on SSA Optimized Hybrid Neural Network Model. Electrochim. Acta 2024, 487, 144146. [Google Scholar] [CrossRef]
  55. Zou, C.; Chen, X.; Zhang, Y. State of Health prediction of lithium-ion batteries based on temporal degeneration feature extraction with Deep Cycle Attention Network. J. Energy Storage 2023, 65, 107367. [Google Scholar] [CrossRef]
  56. He, W.; Williard, N.; Osterman, M.; Pecht, M. Prognostics of lithium-ion batteries based on Dempster–Shafer theory and the Bayesian Monte Carlo method. J. Power Sources 2011, 196, 10314–10321. [Google Scholar] [CrossRef]
  57. Zhou, Y.; Wang, S.; Xie, Y.; Shen, X.; Fernandez, C. Remaining useful life prediction and state of health diagnosis for lithium-ion batteries based on improved grey wolf optimization algorithm-deep extreme learning machine algorithm. Energy 2023, 285, 128761. [Google Scholar] [CrossRef]
  58. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice; OTexts: Melbourne, Australia, 2018. [Google Scholar]
  59. Kang, S.H.; Noh, T.W.; Lee, B.K. Machine Learning-based SOH Estimation Algorithm Using a Linear Regression Analysis. Trans. Korean Inst. Power Electron. 2021, 26, 241–248. [Google Scholar]
  60. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  61. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2017; Volume 30. [Google Scholar]
  62. Liu, X.; Zhao, Z.; Zhang, Y.; Liu, C.; Yang, F. Social network rumor detection method combining dual-attention mechanism with graph convolutional network. IEEE Trans. Comput. Soc. Syst. 2022, 10, 2350–2361. [Google Scholar] [CrossRef]
Figure 1. The possible dynamic interactions of variables in SOH prediction under two observation scales. The black box represents the sliding horizontal window, and the different colored spheres represent the representation vectors of different series learned by the model.
Figure 1. The possible dynamic interactions of variables in SOH prediction under two observation scales. The black box represents the sliding horizontal window, and the different colored spheres represent the representation vectors of different series learned by the model.
Sensors 24 04185 g001
Figure 2. Structure and input/output dynamics of the hybrid network. Note: The time in the time series (such as Constant Current Charge Time, Capacity, Constant Voltage Charge Time, etc.) is represented by the cycle number.
Figure 2. Structure and input/output dynamics of the hybrid network. Note: The time in the time series (such as Constant Current Charge Time, Capacity, Constant Voltage Charge Time, etc.) is represented by the cycle number.
Sensors 24 04185 g002
Figure 3. Cell structure of the FDG.
Figure 3. Cell structure of the FDG.
Sensors 24 04185 g003
Figure 4. The structure of the CDG generator.
Figure 4. The structure of the CDG generator.
Sensors 24 04185 g004
Figure 5. The capacity degradation trajectory of three battery subsets.
Figure 5. The capacity degradation trajectory of three battery subsets.
Sensors 24 04185 g005
Figure 6. Ablation study of NASA and CALCE. The numerical unit is 10 2 .
Figure 6. Ablation study of NASA and CALCE. The numerical unit is 10 2 .
Sensors 24 04185 g006
Figure 7. The SOH estimation results for batteries CS2-35, CS2-36, CS2-37, and CS2-38 within the CALCE dataset.
Figure 7. The SOH estimation results for batteries CS2-35, CS2-36, CS2-37, and CS2-38 within the CALCE dataset.
Sensors 24 04185 g007
Table 1. A summary of symbols and descriptions.
Table 1. A summary of symbols and descriptions.
SymbolDescription
M The SOH of lithium-ion batteries.
Q Factors related to the state-of-health of lithium-ion batteries.
E The interplay of influences among various factors.
NThe number of influencing factors on battery life.
LThe length of the information flow.
JThe number of subseries after feature engineering.
XThe finalized input series of each computational influencing factor.
DThe dimension of the hidden vector.
E The space after information aggregation.
ZThe input series of graph-based neighbor influencing factors.
KThe number of input sequences.
sThe size of the sliding historical window.
A t The global temporal sparse adjacency matrix.
W The learnable parameter of the neural layer.
b The bias of the neural layer.
xThe original series of the adjacent influencing factors.
x The normalized series of the adjacent influencing factors.
x The trend series of the adjacent influencing factors.
XThe integrated input series of the adjacent influencing factors.
yThe ground truth of each training series.
y ^ The output of the HEAG.
e The state vector in the HEAG.
o The output temporal representation vector.
p The representation vectors of adjacent influencing factors.
v The final temporal representation vector for prediction.
γ The attenuation coefficient.
α The attention weight learned by the attention mechanism.
L The loss function used in this work.
Table 2. Dataset information statistics.
Table 2. Dataset information statistics.
DatasetBattery PackSequence NumberAverage Length
CALCECS2-35256735
CS2-36265639
CS2-37277968
CS2-38288236
NASA5-6-7-184857
25-26-27-284739
25-4420821
45-46-47-484769
49-50-51-524807
53-54-55-564792
Table 3. Experimental data based on the CALCE dataset. The bold font highlights the superior performance of the HEAG model compared to other methods.
Table 3. Experimental data based on the CALCE dataset. The bold font highlights the superior performance of the HEAG model compared to other methods.
MethodMetrics35363738Average
ARIMA [58]RMSE ( 10 2 )51.06053.17749.73842.69149.166
MAE ( 10 2 )48.40550.60848.97139.93146.979
MedAE ( 10 2 )49.33452.36150.05941.79848.388
LR [59]RMSE ( 10 2 )49.20046.73646.49341.89046.080
MAE ( 10 2 )45.48346.90546.98438.00544.344
MedAE ( 10 2 )42.10347.08243.23339.77743.049
GRU [60]RMSE ( 10 2 )32.53634.61331.61125.66531.106
MAE ( 10 2 )29.29631.21528.71823.94328.293
MedAE ( 10 2 )30.44034.50432.13625.72730.702
CNN+LSTM [34]RMSE ( 10 2 )27.63530.62028.12723.03327.354
MAE ( 10 2 )26.23130.24027.43619.93925.961
MedAE ( 10 2 )27.30331.42029.45421.12927.326
TCN [43]RMSE ( 10 2 )12.29812.90911.9379.27211.604
MAE ( 10 2 )10.67312.29010.3718.06310.349
MedAE ( 10 2 )10.97013.18310.4098.92310.871
Transformer [61]RMSE ( 10 2 )16.03319.06517.91413.21916.558
MAE ( 10 2 )14.89916.38615.13212.90114.830
MedAE ( 10 2 )15.01719.26016.43313.12315.958
SGEformer [47]RMSE ( 10 2 )1.9922.7302.0831.7472.138
MAE ( 10 2 )1.2422.1531.5830.9161.473
MedAE ( 10 2 )1.8662.3971.9991.1581.855
MGCN [49]RMSE ( 10 2 )4.1576.6894.6893.4434.745
MAE ( 10 2 )2.8665.3433.3212.2103.435
MedAE ( 10 2 )3.0056.0964.3432.9484.098
GCN-DA [50]RMSE ( 10 2 )1.8494.8851.9421.8082.621
MAE ( 10 2 )1.0213.0961.6960.9151.682
MedAE ( 10 2 )1.7213.7842.0141.5402.265
CL-GraphSAGE [51]RMSE ( 10 2 )0.9871.1271.0150.8931.005
MAE ( 10 2 )0.7551.0820.8630.5700.817
MedAE ( 10 2 )0.7200.9520.7190.5980.747
CGGN-DCO [52]RMSE ( 10 2 )0.7830.9130.8100.6230.782
MAE ( 10 2 )0.5440.7550.5300.3390.542
MedAE ( 10 2 )0.6590.8200.6340.4070.630
HEAGRMSE ( 10 2 )0.5500.6130.5440.4110.530
MAE ( 10 2 )0.3410.5170.4530.3160.407
MedAE ( 10 2 )0.3690.5520.4910.3450.439
Table 4. Experimental data based on the NASA dataset. The bold font highlights the superior performance of the HEAG model compared to other methods.
Table 4. Experimental data based on the NASA dataset. The bold font highlights the superior performance of the HEAG model compared to other methods.
MethodMetrics Subset 1 Subset 2 Subset 3 AVG
No. 5 No. 6 No. 7 No. 18 No. 29 No. 30 No. 31 No. 32 No. 25 No. 26 No. 27
ARIMARMSE ( 10 2 )14.20738.15437.08721.04419.41010.2319.06711.63516.45420.11938.59921.455
MAE ( 10 2 )13.84437.46936.89919.09819.4518.0608.71613.00016.36517.49738.55220.814
MedAE ( 10 2 )14.84438.36937.85018.21819.2758.5288.38213.30616.20315.47038.61620.824
LRRMSE ( 10 2 )13.17036.33636.17919.63519.0228.1318.62311.88215.61119.90638.18920.608
MAE ( 10 2 )12.40436.87836.14718.00517.5938.0246.74611.06315.81316.66537.19519.685
MedAE ( 10 2 )12.97337.46538.23416.68819.2606.8848.00412.46014.15113.43236.44619.636
GRURMSE ( 10 2 )8.46131.79132.67915.87615.5384.9793.4638.34111.52916.42334.19416.661
MAE ( 10 2 )7.36532.96232.31213.30715.2983.6693.1037.92310.54412.11632.87315.588
MedAE ( 10 2 )9.89433.88934.18114.12415.5024.4764.0828.52610.08010.41033.34816.229
CNN + LSTMRMSE ( 10 2 )7.24930.62030.55013.28012.6202.3712.1815.4789.21613.70031.49014.432
MAE ( 10 2 )6.04730.24030.98011.85012.4501.9391.5565.3248.97510.02031.44013.711
MedAE ( 10 2 )7.42031.42031.06011.17012.7701.6341.2665.6638.8997.50531.23013.640
TCNRMSE ( 10 2 )2.1742.4682.4452.4402.3212.7222.4063.0743.6393.7373.0642.772
MAE ( 10 2 )1.8862.2912.4342.5212.3752.5842.3753.1483.4663.1233.0502.660
MedAE ( 10 2 )1.9802.1972.4492.2622.3062.6562.4233.1933.4153.0393.1272.641
TransformerRMSE ( 10 2 )4.52025.03025.3444.8567.8574.0492.0691.9312.61823.0508.0359.942
MAE ( 10 2 )3.61816.61016.3253.8774.6033.4911.8261.5902.24418.2407.3487.252
MedAE ( 10 2 )2.75113.26013.4493.2242.0643.1761.7231.2722.20211.5806.3905.554
SGEformerRMSE ( 10 2 )0.9871.5471.4831.4211.2581.9461.4152.3662.6292.6802.0211.796
MAE ( 10 2 )1.2251.5321.5591.2791.2111.7981.6472.3922.4852.4751.8441.768
MedAE ( 10 2 )1.2671.0971.1961.3601.5122.0531.2582.4382.6251.9541.9401.700
MGCNRMSE ( 10 2 )0.6591.6891.8771.2680.4911.1590.4721.0491.6752.2921.6451.298
MAE ( 10 2 )0.5381.3431.4601.0180.4161.1180.4161.0341.5521.5441.6271.097
MedAE ( 10 2 )0.4781.0961.1670.9340.3841.1390.4181.0121.5171.0801.6390.988
GCN-DARMSE ( 10 2 )0.8491.1291.0121.5181.3511.8661.1881.8492.4142.5331.9091.602
MAE ( 10 2 )1.0041.0381.0101.2110.9421.6251.1042.0912.4011.8671.9401.476
MedAE ( 10 2 )0.7740.9751.1371.1601.2781.4571.0982.0052.3721.7971.7531.437
CL-GraphSAGERMSE ( 10 2 )0.8301.1161.2671.2991.0921.5961.3581.8262.3472.6721.7701.561
MAE ( 10 2 )0.7231.0791.1811.3381.1571.4571.1622.0382.1311.8871.7261.443
MedAE ( 10 2 )0.7040.9711.2711.1511.1031.5241.1161.9932.1961.7541.9451.430
CGGN-DCORMSE ( 10 2 )0.2280.5130.5010.7110.4800.9800.5671.2741.6861.9411.1570.913
MAE ( 10 2 )0.1640.3550.4600.5530.4060.8640.4891.2431.5661.3461.1290.780
MedAE ( 10 2 )0.1200.2200.4750.4770.4330.8810.4661.2591.5421.1461.1660.744
HEAGRMSE ( 10 2 )0.1220.3570.3820.5870.3320.8480.4231.1171.5751.8311.0240.782
MAE ( 10 2 )0.1090.2350.3110.4310.2570.7550.3561.1331.4181.2440.9970.659
MedAE ( 10 2 )0.1180.1010.3220.3250.3180.7620.3091.1251.4301.0431.0310.626
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xing, C.; Liu, H.; Zhang, Z.; Wang, J.; Wang, J. Enhancing Lithium-Ion Battery Health Predictions by Hybrid-Grained Graph Modeling. Sensors 2024, 24, 4185. https://doi.org/10.3390/s24134185

AMA Style

Xing C, Liu H, Zhang Z, Wang J, Wang J. Enhancing Lithium-Ion Battery Health Predictions by Hybrid-Grained Graph Modeling. Sensors. 2024; 24(13):4185. https://doi.org/10.3390/s24134185

Chicago/Turabian Style

Xing, Chuang, Hangyu Liu, Zekun Zhang, Jun Wang, and Jiyao Wang. 2024. "Enhancing Lithium-Ion Battery Health Predictions by Hybrid-Grained Graph Modeling" Sensors 24, no. 13: 4185. https://doi.org/10.3390/s24134185

APA Style

Xing, C., Liu, H., Zhang, Z., Wang, J., & Wang, J. (2024). Enhancing Lithium-Ion Battery Health Predictions by Hybrid-Grained Graph Modeling. Sensors, 24(13), 4185. https://doi.org/10.3390/s24134185

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop