Abstract

This paper discusses the application of condition monitoring to a battery system used in a hybrid electric vehicle (HEV). Battery condition management systems (BCMSs) are employed to ensure the safe, efficient, and reliable operation of a battery, ultimately to guarantee the availability of electric power. This is critical for the case of the HEV to ensure greater overall energy efficiency and the availability of reliable electrical supply. This paper considers the use of state and parameter estimation techniques for the condition monitoring of batteries. A comparative study is presented in which the Kalman and the extended Kalman filters (KF/EKF), the particle filter (PF), the quadrature Kalman filter (QKF), and the smooth variable structure filter (SVSF) are used for battery condition monitoring. These comparisons are made based on estimation error, robustness, sensitivity to noise, and computational time.

1. Introduction

Condition monitoring is an essential process for fault detection and diagnosis. It involves monitoring system states or parameters over an operational period, where abnormal values or significant changes would indicate a fault. Quite often direct measurements of the states are not available due to limitations in design or cost. In these cases, state and parameter estimation techniques can be used for information extraction. Condition monitoring of systems allows proper maintenance to be scheduled, which helps reduce unscheduled downtime of manufacturing equipment, as well as the cost to repair damaged systems [1, 2]. An important area for condition monitoring is energy management for hybrid electric (HEVs) and battery electric vehicles (BEVs).

In general, HEVs have two power sources: a gasoline engine and an electric motor. In full hybrid vehicles, the engine and the motor can operate separately or simultaneously. The motor is used mainly during acceleration, startup, reverse mode, and in regenerative braking. A traction battery pack is used to provide power to the motor. It is recharged by a generator or during regenerative braking. The performance of an HEV is largely dependent on a balance between the gasoline engine and the electric motor, optimized with respect to fuel consumption based on vehicle conditions [3]. Many different types of control methods have been applied to balance the power and energy requirements of HEVs, including fuzzy logic [4–6], genetic algorithms [7], dynamic programming [8, 9], Pareto optimization [10], and intelligent mechanism designs [11]. These control strategies rely heavily on the availability of battery power to balance the operation of the gasoline engine versus the electric motor. The available battery power may be obtained from the state-of-charge (SOC) information [3, 12]. Further to the SOC, the battery state-of-health (SOH) is required in order to help determine whether a battery would fail subject to a certain load [13]. The SOC cannot be measured directly with electric signals, and as such it often, needs to be estimated through current and voltage relationships [3]. Poor estimation or control of the SOC may lead to improper charging conditions and can degrade the efficiency and reliability of the batteries [13]. Hence, proper condition monitoring of batteries plays a pivotal role in the optimization of HEV performance, as well as extending the lifetime and increasing the reliability of the batteries [14].

State and parameter estimation techniques are an integral part of condition monitoring and are used when direct measurements of the states are not available. One of the most commonly studied methods for estimation is the Kalman filter (KF) for linear systems and its extended form (EKF) for nonlinear systems [15–19]. Operating conditions such as battery SOC, power fade, capacity fade, resistance, and instantaneous available power have been estimated well using the EKF [17]. Other methods such as the sigma-point Kalman filtering (SPKF) and support vector machine (SVM) have also been used for condition monitoring and fault diagnosis [20–22].

In critical applications, such as automotive that require added safety and reliability, the choice of the estimation method is very important and should be selected based on the linearity of the system, performance, robustness or sensitivity to noise and computational difficulty and time. In relation to battery condition monitoring, this paper presents a quantitative and qualitative comparison of the following methods: the Kalman and extended Kalman filters (KF/EKF), the particle filter (PF), the quadrature Kalman filter (QKF), and the smooth variable structure filter (SVSF).

2. State and Parameter Estimation Techniques

State and parameter estimation is essential for sensing and information processing in model-based condition monitoring. Estimation theory involves information extraction by tracking changes in physical parameters or operational states of the system. This paper studies four popular strategies: the Kalman and the extended Kalman filters (KF/EKF), the particle filter (PF), the quadrature Kalman filter (QKF), and the smooth variable structure filter (SVSF).

2.1. Kalman and Extended Kalman Filters

Even after 50 years, the Kalman filter (KF) remains the most studied and one of the most popular tools used in state estimation [23–26]. It may be applied to linear dynamic systems in the presence of Gaussian white noise, and provides an elegant and statistically optimal solution by minimizing the mean-squared error. The impact that the KF has had on estimation and control problems is considered by some scientists and engineers to be one of the greatest achievement in engineering and signal processing [26]. It is a method that utilizes measurements linearly related to the states, and error covariance matrices, to generate a gain referred to as the Kalman gain. This gain is applied to the a priori state estimate, thus creating an a posteriori estimate. The estimation process continues in a predictor-corrector fashion while maintaining a statistically minimal state error covariance matrix for linear systems.

The following two equations describe the system dynamic model and the measurement model used in general for state estimation. Refer to Appendix A for a description of the nomenclature π‘₯π‘˜+1=π΄π‘˜π‘₯π‘˜+π΅π‘˜π‘’π‘˜+π‘€π‘˜,π‘§π‘˜+1=πΆπ‘˜+1π‘₯π‘˜+1+π‘£π‘˜+1.(1)

The next five equations form the KF algorithm and are used in an iterative fashion. Equation (2) extrapolates the a priori state estimate, and (3) is the corresponding a priori error covariance. The Kalman gain may be calculated by (4), and is used to update the state estimate and error covariance, described by (5) and (6), respectively. Μ‚π‘₯π‘˜+1|π‘˜=π΄π‘˜Μ‚π‘₯π‘˜|π‘˜+π΅π‘˜π‘’π‘˜,𝑃(2)π‘˜+1|π‘˜=π΄π‘˜π‘ƒπ‘˜|π‘˜π΄π‘‡π‘˜+π‘„π‘˜πΎ,(3)π‘˜=π‘ƒπ‘˜+1|π‘˜πΆπ‘‡π‘˜ξ€ΊπΆπ‘˜π‘ƒπ‘˜+1|π‘˜πΆπ‘‡π‘˜+π‘…π‘˜ξ€»βˆ’1,(4)Μ‚π‘₯π‘˜+1|π‘˜+1=Μ‚π‘₯π‘˜+1|π‘˜+πΎπ‘˜ξ€Ίπ‘§π‘˜βˆ’πΆπ‘˜π‘₯π‘˜+1|π‘˜ξ€»π‘ƒ,(5)π‘˜+1|π‘˜+1=ξ€ΊπΌβˆ’πΎπ‘˜πΆπ‘˜ξ€»π‘ƒπ‘˜+1|π‘˜.(6)

The effects due to model uncertainties can have a large impact on the stability and performance of the KF [26, 27]. For nonlinear systems, the EKF may be used. It is conceptually similar to the KF process. The nonlinear system and measurement matrices are linearized according to their corresponding Jacobian, which is a first-order partial derivative. This linearization introduces uncertainties in the estimation process; such that overlooked nonlinearities in the system may cause the EKF to become unstable [26].

2.2. Particle Filter

The particle filter (PF) has many forms: Monte Carlo filters, interacting particle approximations [28], bootstrap filters [29], condensation algorithm [30], and survival of the fittest [31], to name a few. Compared to the KF, it is newer, being introduced in 1993. Since then, the PF has become a very popular method for solving nonlinear estimation problems, ranging from predicting chemical processes to target tracking. The PF takes the Bayesian approach to dynamic state estimation, in which one attempts to accurately represent the probability distribution function (PDF) of the values of interest [32]. The PDF contains all of the pertinent statistical information and may be considered as holding the solution to the estimation problem [32]. Essentially, the distribution holds a probability of values for the state being observed. The stronger or tighter the prediction PDF, the more accurate the state estimate.

The PF obtains its name from the use of weighted particles or β€œpoint masses” that are distributed throughout the PDF to form an approximation. These particles are used in an iterative process to obtain new particles and associated importance weights, with the goal of creating a more accurate approximation of the PDF. In general, as the number of implemented particles becomes large, the PDF becomes more accurate [32]. An important step in the PF is that of resampling, which eliminates particles with low weights and multiplies those with high weights [32]. This helps to avoid the degeneracy problem with the PF, which refers to only one particle having a significant importance weight after a large number of recursions. Furthermore, it also increases the accuracy of the PDF approximation by replicating particles with high weights. The sequential importance resampling (SIR) algorithm is a very popular form of the PF and may be summarized by (7) to (10). The first equation draws samples or particles from the proposal distribution. π‘₯π‘˜(𝑛)ξ‚€π‘₯βˆΌπœ‹π‘˜βˆ£π‘₯(𝑛)π‘˜βˆ’1,π‘¦π‘˜ξ‚.(7) The next equation updates the importance weights up to a normalizing constant ξπœ”π‘˜(𝑛)=πœ”(𝑛)π‘˜βˆ’1π‘ξ‚€π‘¦π‘˜βˆ£π‘₯π‘˜(𝑛)𝑝π‘₯π‘˜(𝑛)∣π‘₯(𝑛)π‘˜βˆ’1ξ‚πœ‹ξ‚€π‘₯π‘˜βˆ£π‘₯(𝑛)π‘˜βˆ’1,π‘¦π‘˜ξ‚.(8) Next, the normalized weights are calculated for each particle πœ”π‘˜(𝑛)=ξπœ”π‘˜(𝑛)βˆ‘π‘›π‘–=1ξπœ”π‘˜(𝑖).(9)

Finally, a constant known as the effective number of particles is calculated as shown in (10). Resampling is performed if the effective number of particles is lower than some design threshold 𝑁eff=1βˆ‘π‘›π‘–=1ξ‚€πœ”π‘˜(𝑖)2.(10)

2.3. Quadrature Kalman Filter

Similar to the PF, the quadrature Kalman filter (QKF) is a type of Bayesian filter that is able to model dynamic processes which are nonlinear and subject to non-Gaussian noise. In 2007, it was proposed that a set of Gauss-Hermite quadrature points could be used to parameterize the PDF [33, 34]. When compared with the EKF, it was found that the QKF method provides a more accurate least-squares solution [33]. Figure 1 shows the PDF of a nonlinearly transformed Gaussian random variable. The mean of the EKF appears to be biased, and the covariance is obviously far from the true covariance. However, the QKF mean and covariance match the true values quite well, with the 5 point QKF working the best. The main drawback to this method is the fact that the number of terms used in the Gaussian sum grows exponentially, which means that more memory will be used over a longer period of time [33]. The process of this filter is similar to the PF, in the sense that it may be solved recursively using two stages (time and measurement update). Please refer to Appendix B for the full QKF algorithm.

2.4. Smooth Variable Structure Filter

In 2002, the variable structure filter (VSF) was introduced as a new predictor-corrector method used for state and parameter estimation [27, 35]. It is a type of sliding mode estimator, where gain switching is used to ensure that the estimates converge to true state values. An internal model of the system, either linear or nonlinear, is used to predict an a priori state estimate. A corrective term is then applied to calculate the a posteriori state estimate, and the estimation process is repeated iteratively. The SVSF was later derived from the VSF and uses a simpler and less complex gain calculation [36]. In its present form, the SVSF is stable and robust to modeling uncertainties and noise, given an upperbound on uncertainties [36]. The basic concept of the SVSF is shown in Figure 2. Assume that the solid line in Figure 2 is a trajectory of some state (amplitude versus time). An initial value is selected for the state estimate. The estimated state is pushed towards the true value until it reaches a subspace around the actual state trajectory, referred to as the existence subspace. Once the value enters the existence subspace, the estimated state is forced to remain within it and into switching along the system state trajectory [36].

The SVSF method is model based and applies to smooth nonlinear dynamic equations. The estimation process may be summarized by (11) to (16) and is repeated iteratively. An a priori state estimate is calculated using an estimated model of the system. This value is then used to calculate an a priori estimate of the measurement defined by (12). A corrective term, referred to as the SVSF gain, is calculated as a function of the error in the predicted output, as well as a gain matrix and the smoothing boundary layer width. The corrective term calculated in (13) is then used in (14) to find the a posteriori state estimate Μ‚π‘₯π‘˜+1|π‘˜=𝐹̂π‘₯π‘˜|π‘˜,π‘’π‘˜ξ€Έ,(11)Μ‚π‘§π‘˜+1|π‘˜=𝐢̂π‘₯π‘˜+1|π‘˜πΎ,(12)π‘˜+1=ξπΆβˆ’1|||ξ‚€|||π‘’π‘§π‘˜+1|π‘˜|||𝐴𝐡𝑆|||𝑒+π›Ύπ‘§π‘˜|π‘˜|||𝐴𝐡𝑆|||π΄π΅π‘†ξ‚€π‘’βˆ˜satπ‘§π‘˜+1|π‘˜ξ‚,,Ξ¨(13)Μ‚π‘₯π‘˜+1|π‘˜+1=Μ‚π‘₯π‘˜+1|π‘˜+πΎπ‘˜+1𝑒,(14)π‘§π‘˜|π‘˜=π‘§π‘˜βˆ’Μ‚π‘§π‘˜|π‘˜π‘’,(15)π‘§π‘˜+1|π‘˜=π‘§π‘˜+1βˆ’Μ‚π‘§π‘˜+1|π‘˜.(16)

Two critical variables in this process are the a priori and a posteriori output error estimates, defined by (15) and (16), respectively [36]. Note that (15) is the output error estimate from the previous time step and is used only in the gain calculation.

3. Condition Monitoring of a Battery System in a Hybrid Electric Vehicle

A variety of batteries have been studied in literature, most notably lead-acid and lithium-ion batteries [13, 15, 19, 20, 37]. Lead-acid batteries are the oldest type of rechargeable batteries, and are most commonly found in motor vehicles. Lithium-ion batteries are also a form of rechargeable battery, which contain lithium in its positive electrode (cathode). These batteries are usually found in portable consumer electronics (i.e., laptops or notebooks) due to particularly high energy-to-weight ratios, slow self-discharge, and a lack of memory effect (i.e., where a battery loses its maximum energy capacity over time) [16]. In recent years, lithium-ion batteries have slowly entered the hybrid electric vehicle market, due to the fact that they offer better energy density compared to standard batteries [38].

The operation of batteries may be studied by using the advanced vehicle simulator (ADVISOR), which was written in MATLAB and Simulink by the US Department of Energy and the National Renewable Energy Laboratory [39–41]. ADVISOR is used for the analysis of performance and fuel economy of three vehicle types: conventional, electric, and hybrid vehicles [39]. In 2001, the resistance-capacitance (RC) battery model was first implemented in ADVISOR [42]. The electrical model consists of three resistors (𝑅𝑒, 𝑅𝑐, and 𝑅𝑑) and two capacitors (𝐢𝑏 and 𝐢𝑐). The first capacitor (𝐢𝑏) represents the capability of the battery to chemically store a charge, and the second capacitor (𝐢𝑐) represents the surface effects of a cell [41]. The resistances and capacitances vary with changing SOC and temperature (𝑇) [41]. ADVISOR offers two different datasets for the RC battery model: lithium-ion and nickel-metal hydride chemistries. For the purposes of this study, the lithium-ion chemistry was used in conjunction with the RC battery model. Figure 3 illustrates the equivalent circuit diagram of the RC model.

A standard model of a parallel hybrid electric vehicle referred to within ADVISOR as the Annex VII PHEV was used for this study. This model has been developed by the International Energy Agency (IEA), which is an international research community for the development and commercialization of hybrid and electric vehicles [43]. The model is based on data obtained from published sources and national (U.S.) research test data [39]. The battery system of the HEV represents the battery pack which stores energy on board the HEV. The system accepts a power request and returns the available power from the battery, as well as the SOC, voltage and current [39].

The nonlinear equations that describe the system may be derived from the RC battery model of Figure 3, as shown in Appendix C. For the purposes of condition monitoring of the battery, two voltages 𝑉𝐢𝑏 and 𝑉𝐢𝑐 as well as two capacitance parameters 𝐢𝑏 and 𝐢𝑐 need to be estimated. Should a fault exist in either of the battery capacitors, one would be able to determine this given the corresponding change in the parametric value. Further to the equations found in Appendix C, a discrete-time state space model of the Capacitor voltages may be defined as follows: ξ‚Έπ‘‰πΆπ‘π‘˜+1π‘‰πΆπ‘π‘˜+1ξ‚Ή=βŽ‘βŽ’βŽ’βŽ’βŽ£βˆ’π‘‡π‘ πΆπ‘(𝑅𝑒+𝑅𝑐)𝑇+1𝑠𝐢𝑏(𝑅𝑒+𝑅𝑐)𝑇𝑠𝐢𝑐(𝑅𝑒+𝑅𝑐)βˆ’π‘‡π‘ πΆπ‘(𝑅𝑒+𝑅𝑐)⎀βŽ₯βŽ₯βŽ₯⎦+1π‘˜ξ‚Έπ‘‰πΆπ‘π‘˜π‘‰πΆπ‘π‘˜ξ‚Ή+βŽ‘βŽ’βŽ’βŽ’βŽ£π‘‡π‘ π‘…π‘πΆπ‘(𝑅𝑒+𝑅𝑐)𝑇𝑠𝑅𝑒𝐢𝑐(𝑅𝑒+𝑅𝑐)⎀βŽ₯βŽ₯βŽ₯βŽ¦π‘˜πΌπ‘†π‘˜.(17)

For parameter estimation, (17) is used to formulate a state vector that would include parameters such that 𝑉π‘₯=𝐢𝑏𝑉𝐢𝑐𝐢𝑏𝐢𝑐𝑇.(18)

By rearranging (17), and further to Appendix C, the general model used for state and parameter estimation is obtained as π‘₯π‘˜+1ξ€·π‘₯=π‘“π‘˜,π‘’π‘˜ξ€Έ+π‘€π‘˜,π‘§π‘˜+1ξ€·π‘₯=β„Žπ‘˜+1,π‘’π‘˜+1ξ€Έ+π‘£π‘˜+1.(19)

To implement the EKF method, the nonlinear system had to be linearized using a first-order Taylor series approximation. The linearized form of the system equation is provided in Appendix D.

4. Estimation Results

This paper presents a comparative performance for the application of the extended Kalman filter, the particle filter, the quadrature Kalman filter, and the smooth variable structure filter for condition monitoring of RC batteries. The study is conducted by simulation using the ADVISOR battery model. The model parameters are documented in [39].

Figures 4 and 5 illustrate the input current (𝐼𝑠) and the output voltages 𝑉𝐢𝑏 and 𝑉𝐢𝑐 used in the simulation. The parameter values 𝐢𝑏 and 𝐢𝑐 are made to vary in order to simulate fault conditions as shown in Figures 6 and 7. Parameters 𝐢𝑏 and 𝐢𝑐 as well as the states 𝑉𝐢𝑏 and 𝑉𝐢𝑐 are estimated using the four methods described in Section 2.

A comparison and discussion of the results is provided using two simulated cases: one with noise and one with noise as well as modeling errors. Note that the sampling time used in the simulation was 0.01 seconds.

4.1. Extended Kalman Filter Results

For linear dynamic systems in the presence of Gaussian white noise, the KF provides an elegant and statistically optimal solution by minimizing the mean-squared estimation error. The EKF is used for nonlinear problems. The following covariance matrices (error, process, and measurement, resp.) were used for the EKF and were obtained by trial-and-error: ⎑⎒⎒⎒⎣⎀βŽ₯βŽ₯βŽ₯⎦,⎑⎒⎒⎒⎣⎀βŽ₯βŽ₯βŽ₯⎦.𝑃=1303030301303030301303030301𝑄=𝑅=1000010000100001(20)

The EKF worked very well in the absence of modeling errors, as shown in Figure 8. Its performance degraded considerably when modeling errors were added, as illustrated in Figure 9.

As shown in Figure 10, the EKF was very sensitive to the selection of the initial conditions; however, it often recovered and performed very well. This is important to note because without selecting good starting points, the filter performed the worst. However, when good initial conditions were selected, the EKF yields the best results (based on RMSE and the assumption that there were no uncertainties in the model). When uncertainties (besides white measurement and process noise) were present in the filter model, in this example, the EKF became unstable (as shown in Figure 9) and failed to yield reasonable results. The simulation time for the EKF was one of the fastest, and compared to the other methods, it was computationally easy.

4.2. Particle Filter Results

For the PF method, a large number of particles (500) were used. Its application results are shown in Figures 11 and 12. The PF was able to estimate fairly well in the presence of noise. However, when modeling uncertainties were introduced, the PF had difficulty tracking the true voltages. As shown in Figure 12(b), after about 20 seconds, the PF was able to recover and provide a good estimate.

The PF provided satisfactory results. However, the technique had the highest RMSE (for the first case). That being said, when compared to the EKF, no Jacobian matrix had to be calculated to linearize the system matrix, as weighted particles were used instead. This may be an attractive feature if the system is too difficult to linearize. A large number of particles (which were required to accurately estimate the PDF), and the resampling feature of the PF, contributed to a slower computational time. When uncertainties were added to the filter model (as shown in Figure 12), the PF was able to recover slowly from modeling errors and after some time, accurately represent the first two states. There were significant errors in the third and fourth states; however this, as previously mentioned, may be attributed to measurements being only available for the first two states.

4.3. Quadrature Kalman Filter Results

For the QKF, the initial values of the covariance matrices were set to the identity matrix. The following results shown in Figures 13 and 14 were generated using the QKF based on the simulation setup. The QKF performed very well for both cases. However, like the previous methods, when uncertainties were introduced, it was unable to accurately track the two parameters associated with the battery (𝐢𝑏 and 𝐢𝑐) as shown in Figure 14.

The QKF provided good results for the system. Like the PF, there was no need to calculate the Jacobian, since weighted quadrature points were used instead. One of the drawbacks of the QKF is the computational demand when a large number of states and quadrature points are used. This leads to a slower calculation time. For example, since four states were required, an array of 81 quadrature points was required (34). If a more accurate model was required, five quadrature points per state could be used. However, this would further increase the computational time. For this system, it was found that increasing the number of quadrature points had a negligible effect on the accuracy. Note that when uncertainties were added to the filter model, the QKF worked extremely well. However, the slow computation time is its main hindrance.

4.4. Smooth Variable Structure Filter Results

The constant diagonal matrix (𝛾) was set to 0.4, and the smoothing boundary layer thicknesses (πœ“) were, respectively, set to 1Γ—10βˆ’3 and 1Γ—10βˆ’2 for the two states and the two parameters, respectively. The results are shown in Figures 15 and 16, and were generated using the SVSF based on the simulation setup. The SVSF method is very robust to noise and modeling uncertainties, as demonstrated by the accurate estimation in both cases.

The SVSF yielded very good results for condition monitoring of the battery system. Similar to the previous two methods, it did not require calculation of the Jacobian matrix, and also required a time delay (signal extraction for the last two states). The SVSF simulation time was just as fast as the EKF, and the results were similar for the case with noise. When errors were added to the filter model, the SVSF worked the best (in terms of RMSE). Although there were still large estimation errors for the third and fourth states, they were finite and stable. It is important to note that the SVSF was robust and not sensitive to changes in the filter parameters and the initial conditions.

4.5. Summary of Results

The results of the simulations were compared based on estimation error, robustness, sensitivity to noise, and computational difficulty and time. As shown in Table 1, the EKF performed best interms of estimation error, in the case of only noise (referred to as Case  (1)). In the presence of modeling uncertainties (Case  (2)), the EKF’s performance severely degraded. The SVSF yielded the most accurate estimation for the case involving modeling uncertainties. Both the PF and QKF performed moderately well, but at a cost of computational time, as shown in the last row of Table 1.

The error in Table 1 refers tothe root mean-squared error (RMSE), calculated as follows: 1RMSE=𝑁𝑁𝑖=1ξ€·π‘₯π‘–βˆ’Μ‚π‘₯𝑖2.(21)

Table 2 gives a quantitative comparison of the four filters. The robustness was based on observations made on varying the noise levels and considering its impact on model errors. The sensitivities were determined by tuning the filters various parameters. Computation (simulation) time was determined for each method and may be compared on a relative basis.

5. Conclusions

This paper discussed the application of condition monitoring to an RC battery system, typically found within a hybrid electric vehicle. State and parameter estimation techniques are important as they are responsible for providing accurate estimates of the states when reliable measurements are unavailable, and hence ensure successful condition monitoring. A comparative study was presented in which the Kalman and the extended Kalman filters (KF/EKF), the particle filter (PF), the quadrature Kalman filter (QKF), and the smooth variable structure filter (SVSF) were applied for condition monitoring. These estimation methods were compared based on estimation error, robustness, sensitivity to noise, and computation time. Table 3 summarizes the results of the comparison. For the case without modeling errors, the EKF worked the best in terms of RMSE and computational speed. When modeling errors were present, which is common in physical applications, the SVSF was shown to work significantly better than the other methods in terms of stability and RMSE.

Appendices

A. List of Nomenclature

𝐴:System matrix𝐡: Input matrix𝐢:Output matrix𝑒:State estimation error𝑓:Nonlinear system or process equationβ„Ž:Nonlinear measurement or output equationπ‘˜:Time step index𝐾:Gain value (KF, QKF, or SVSF)π‘š:Number of measurements𝑛:Number of states𝑁eff:Effective number of particles𝑃: Error covariance matrix𝑄:System noise covariance matrix𝑅:Measurement noise covariance matrixSat:Saturation function𝑑:Simulation time𝑇𝑠:Sampling time𝑒: Input𝑣:Measurement noise𝑀:System noise𝑀𝑙:Quadrature point weightπ‘₯:System states𝑦:Observed measurement or output𝑧:Measurement output𝛾:Constant diagonal gain matrix with elements having values between 0 and 1πœ”:Particle weightΞ¨:Smoothing boundary layerπœ‹:Probability distribution𝜏:Sampling timeΜ‚β€Œ:Denotes an estimated value~:Denotes an error valueβ‹…:On top of a parameter denotes a time derivative.

Furthermore, note that subscript π‘˜+1βˆ£π‘˜ refers to an a priori time step and that the subscript π‘˜+1βˆ£π‘˜+1 refers an a posteriori time step. A superscript of 𝑇 denotes a matrix transpose.

B. Quadrature Kalman Filter Algorithm

The following is a summary of the QKF algorithm, directly as presented in [33, 34]. This process was used to obtain the presented results. There are two updates at each step: the time update and the measurement update.

Time Update Step
(1)Assuming at time π‘˜ the posterior density function 𝑝(π‘₯π‘˜βˆ’1βˆ£π‘§π‘˜βˆ’1)=𝑁(Μ‚π‘₯π‘˜βˆ’1|π‘˜βˆ’1,π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1) is known, then we may factorize as follows: π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1=ξ”π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1ξ‚€ξ”π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1𝑇.(B.1)(2)Evaluate the quadrature points {𝑋𝑙,π‘˜βˆ’1|π‘˜βˆ’1}π‘šπ‘™=1 as 𝑋𝑙,π‘˜βˆ’1|π‘˜βˆ’1=ξ”π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1πœ‰π‘™+Μ‚π‘₯π‘˜βˆ’1|π‘˜βˆ’1.(B.2)(3)Evaluate the propagated quadrature points {π‘‹βˆ—π‘™,π‘˜|π‘˜βˆ’1}π‘šπ‘™=1 as:π‘‹βˆ—π‘™,π‘˜|π‘˜βˆ’1𝑋=𝑓𝑙,π‘˜βˆ’1|π‘˜βˆ’1,π‘’π‘˜βˆ’1ξ€Έ,π‘˜βˆ’1.(B.3)(4)Estimate the predicted state: Μ‚π‘₯π‘˜|π‘˜βˆ’1=ξ“π‘šπ‘™=1π‘€π‘™π‘‹βˆ—π‘™,π‘˜|π‘˜βˆ’1.(B.4)(5) Estimate the predicted error covariance: π‘ƒπ‘˜|π‘˜βˆ’1=ξ“π‘šπ‘™=1π‘€π‘™π‘‹βˆ—π‘™,π‘˜|π‘˜βˆ’1π‘‹βˆ—π‘‡π‘™,π‘˜|π‘˜βˆ’1βˆ’Μ‚π‘₯π‘˜|π‘˜βˆ’1Μ‚π‘₯π‘‡π‘˜|π‘˜βˆ’1+π‘„π‘˜.(B.5)

Measurement Update Step
(1)Factorize: π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1=ξ”π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1ξ‚€ξ”π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1𝑇.(B.6)(2)Evaluate the quadrature points {𝑋𝑙,π‘˜βˆ’1|π‘˜βˆ’1}π‘šπ‘™=1 as: 𝑋𝑙,π‘˜βˆ’1|π‘˜βˆ’1=ξ”π‘ƒπ‘˜βˆ’1|π‘˜βˆ’1πœ‰π‘™+Μ‚π‘₯π‘˜βˆ’1|π‘˜βˆ’1.(B.7)(3)Evaluate the propagated quadrature points {𝑍𝑙,π‘˜|π‘˜βˆ’1}π‘šπ‘™=1 as: 𝑍𝑙,π‘˜|π‘˜βˆ’1𝑋=β„Žπ‘™,π‘˜|π‘˜βˆ’1,π‘’π‘˜ξ€Έ,π‘˜.(B.8)(4)Estimate the predicted measurement: Μ‚π‘§π‘˜|π‘˜βˆ’1=ξ“π‘šπ‘™=1𝑀𝑙𝑍𝑙,π‘˜|π‘˜βˆ’1.(B.9)(5)Estimate the innovation covariance matrix 𝑃𝑧𝑧,π‘˜|π‘˜βˆ’1=π‘…π‘˜+ξ“π‘šπ‘™=1𝑀𝑙𝑍𝑙,π‘˜|π‘˜βˆ’1𝑍𝑇𝑙,π‘˜|π‘˜βˆ’1βˆ’Μ‚π‘§π‘˜|π‘˜βˆ’1Μ‚π‘§π‘‡π‘˜|π‘˜βˆ’1.(B.10)(6)Estimate the cross covariance matrix 𝑃π‘₯𝑧,π‘˜|π‘˜βˆ’1=ξ“π‘šπ‘™=1𝑀𝑙𝑋𝑙,π‘˜|π‘˜βˆ’1𝑍𝑇𝑙,π‘˜|π‘˜βˆ’1βˆ’Μ‚π‘₯π‘˜|π‘˜βˆ’1Μ‚π‘§π‘‡π‘˜|π‘˜βˆ’1.(B.11)(7)Estimate the Kalman gain: π‘Šπ‘˜=𝑃π‘₯𝑧,π‘˜|π‘˜βˆ’1π‘ƒβˆ’1𝑧𝑧,π‘˜|π‘˜βˆ’1.(B.12)(8)Estimate the updated step: Μ‚π‘₯π‘˜|π‘˜=Μ‚π‘₯π‘˜|π‘˜βˆ’1+π‘Šπ‘˜ξ€·π‘§π‘˜βˆ’Μ‚π‘§π‘˜|π‘˜βˆ’1ξ€Έ.(B.13)(9)Estimate the corresponding error covariance: π‘ƒπ‘˜|π‘˜=π‘ƒπ‘˜|π‘˜βˆ’1βˆ’π‘Šπ‘˜π‘ƒπ‘§π‘§,π‘˜|π‘˜βˆ’1π‘Šπ‘‡π‘˜.(B.14)

C. Derivation of the RC Battery Model Equations

The output voltage may be calculated by summing the voltages of each element in the circuit. Summation of the outer loop and inner loop voltages yield two equations for the output voltage, respectively: 𝑉𝑂=𝐼𝑆𝑅𝑑+𝐼𝑏𝑅𝑒+𝑉𝐢𝑏,𝑉(C.1)𝑂=𝐼𝑆𝑅𝑑+𝐼𝑐𝑅𝑐+𝑉𝐢𝑐.(C.2)

Recall Kirchhoff’s current law, which states that the total current entering a node must equal the total current leaving, such that the supply current is defined by 𝐼𝑆=𝐼𝑏+𝐼𝑐.(C.3) Rearranging for 𝐼𝑐 in (C.3) and substituting into (C.2) yields 𝑉𝑂=𝐼𝑆𝑅𝑑+ξ€·πΌπ‘†βˆ’πΌπ‘ξ€Έπ‘…π‘+𝑉𝐢𝑐.(C.4) Equating (C.4) with the first output voltage (C.1) and solving for 𝐼𝑏 yields 𝐼𝑏=𝐼𝑆𝑅𝑐𝑅𝑒+𝑅𝑐+𝑉𝐢𝑐𝑅𝑒+π‘…π‘βˆ’π‘‰πΆπ‘π‘…π‘’+𝑅𝑐.(C.5) Note that ̇𝑉𝐢𝑏=𝐼𝑏/𝐢𝑏 such that (C.5) becomes ̇𝑉𝐢𝑏=𝐼𝑆𝑅𝑐𝐢𝑏𝑅𝑒+𝑅𝑐+𝑉𝐢𝑐𝐢𝑏𝑅𝑒+π‘…π‘ξ€Έβˆ’π‘‰πΆπ‘πΆπ‘ξ€·π‘…π‘’+𝑅𝑐.(C.6) Similarly, the above approach may be used to solve for the second capacitor voltage rate of change: ̇𝑉𝐢𝑐=𝐼𝑆𝑅𝑒𝐢𝑐𝑅𝑒+𝑅𝑐+𝑉𝐢𝑏𝐢𝑐𝑅𝑒+π‘…π‘ξ€Έβˆ’π‘‰πΆπ‘πΆπ‘ξ€·π‘…π‘’+𝑅𝑐.(C.7) The output voltage may be solved by manipulating (C.1) through (C.3). Setting 𝐼𝑐=πΌπ‘†βˆ’πΌπ‘ in (C.2), and rearranging for 𝐼𝑏, results in the following: 𝐼𝑏=𝑅𝑑+𝑅𝑐𝑅𝑐𝐼𝑆+π‘‰πΆπ‘π‘…π‘βˆ’π‘‰π‘‚π‘…π‘.(C.8) Rearranging (C.1) for 𝐼𝑏,and setting the result equal to (C.8) yields π‘‰π‘‚π‘…π‘’βˆ’π‘‰πΆπ‘π‘…π‘’βˆ’π‘…π‘‘π‘…π‘’πΌπ‘†=𝑅𝑑+𝑅𝑐𝑅𝑐𝐼𝑆+π‘‰πΆπ‘π‘…π‘βˆ’π‘‰π‘‚π‘…π‘.(C.9) Manipulating (C.9) and simplifying for the output voltage results in 𝑉𝑂=𝑅𝑐𝑅𝑒+𝑅𝑐𝑉𝐢𝑏+𝑅𝑒𝑅𝑒+𝑅𝑐𝑉𝐢𝑐+𝑅𝑑+𝑅𝑒𝑅𝑐𝑅𝑒+𝑅𝑐𝐼𝑆.(C.10)

D. Linearization of the Battery System

The battery model described in Section 3 requires linearization for the EKF to be implemented. The linearization was based on the Jacobian matrix and the four states (where for neatness, π‘ŠπΆπ‘=1/𝐢𝑏 and π‘ŠπΆπ‘=1/𝐢𝑐) πœ™π‘˜=βŽ‘βŽ’βŽ’βŽ’βŽ’βŽ’βŽ’βŽ’βŽ’βŽ£πœ•π‘“1πœ•π‘‰πΆπ‘πœ•π‘“1πœ•π‘‰πΆπ‘πœ•π‘“1πœ•π‘ŠπΆπ‘πœ•π‘“1πœ•π‘ŠπΆπ‘πœ•π‘“2πœ•π‘‰πΆπ‘πœ•π‘“2πœ•π‘‰πΆπ‘πœ•π‘“2πœ•π‘ŠπΆπ‘πœ•π‘“2πœ•π‘ŠπΆπ‘πœ•π‘“3πœ•π‘‰πΆπ‘πœ•π‘“3πœ•π‘‰πΆπ‘πœ•π‘“3πœ•π‘ŠπΆπ‘πœ•π‘“3πœ•π‘ŠπΆπ‘πœ•π‘“4πœ•π‘‰πΆπ‘πœ•π‘“4πœ•π‘‰πΆπ‘πœ•π‘“4πœ•π‘ŠπΆπ‘πœ•π‘“4πœ•π‘ŠπΆπ‘βŽ€βŽ₯βŽ₯βŽ₯βŽ₯βŽ₯βŽ₯βŽ₯βŽ₯βŽ¦π‘˜,(D.1) where the functions in (D.1) are described by 𝑓1𝑇=βˆ’π‘†π‘‰πΆπ‘π‘ŠπΆπ‘π‘…π‘’+𝑅𝑐+𝑉𝐢𝑏+π‘‡π‘†π‘‰πΆπ‘π‘ŠπΆπ‘π‘…π‘’+𝑅𝑐+π‘‡π‘†π‘…π‘π‘ŠπΆπ‘π‘…π‘’+𝑅𝑐𝐼𝑆,𝑓2𝑇=βˆ’π‘†π‘‰πΆπ‘π‘ŠπΆπ‘π‘…π‘’+π‘…π‘βˆ’π‘‡π‘†π‘‰πΆπ‘π‘ŠπΆπ‘π‘…π‘’+𝑅𝑐+𝑉𝐢𝑐+π‘‡π‘†π‘…π‘’π‘ŠπΆπ‘π‘…π‘’+𝑅𝑐𝐼𝑆,𝑓3=π‘ŠπΆπ‘,𝑓4=π‘ŠπΆπ‘.(D.2)

The linearized form (πœ™) of the system matrix, using the state vector described by (18), is as follows: πœ™π‘˜=βŽ‘βŽ’βŽ’βŽ’βŽ£πœ™11πœ™12πœ™130πœ™21πœ™220πœ™24⎀βŽ₯βŽ₯βŽ₯⎦00100001π‘˜,(D.3) where the elements are described by: πœ™11=βˆ’π‘‡π‘†π‘Šπ‘˜πΆπ‘π‘…π‘’π‘˜+π‘…π‘π‘˜πœ™+1,12=π‘‡π‘†π‘Šπ‘˜πΆπ‘π‘…π‘’π‘˜+π‘…π‘π‘˜,πœ™13=π‘‡π‘†π‘…π‘’π‘˜+π‘…π‘π‘˜ξ€·βˆ’π‘‰π‘˜πΆπ‘+π‘‰π‘˜πΆπ‘+π‘…π‘π‘˜πΌπ‘†π‘˜ξ€Έ,πœ™21=π‘‡π‘†π‘Šπ‘˜πΆπ‘π‘…π‘’π‘˜+π‘…π‘π‘˜,πœ™22=βˆ’π‘‡π‘†π‘Šπ‘˜πΆπ‘π‘…π‘’π‘˜+π‘…π‘π‘˜πœ™+1,24=π‘‡π‘†π‘…π‘’π‘˜+π‘…π‘π‘˜ξ€·π‘‰π‘˜πΆπ‘βˆ’π‘‰π‘˜πΆπ‘+π‘…π‘’π‘˜πΌπ‘†π‘˜ξ€Έ.(D.4)