Abstract

A method is proposed to approximate the main features or patterns including interventions that may occur in a time series. Collision data from the Ontario Ministry of Transportation illustrate the approach using monthly collision counts from police reports over a 10-year period from 1990 to 1999. The domain of the time series is partitioned into nonoverlapping subdomains. The major condition on the approximation requires that the series and the approximation have the same average value over each subdomain. To obtain a smooth approximation, based on the second difference of the series, a few iterations are necessary since an iteration over one subdomain is affected by the previous iteration over the adjacent subdomains.

1. Introduction

Graduated licensing system (GLS) is a method of gradual exposure of young novice drivers into the driving environment, allowing them to obtain initial experience with driving under supervision, followed by more independent driving under higher-risk circumstances [1]. This model was widely incorporated into driver licensing programs across the US and Canada as well as other countries over the 1990s. Most of these programs have incorporated similar restrictions into their initial phases [2]. These include driving with supervision, restricted driving at night, limited teenage passengers, and zero blood alcohol level while driving. This method has had limited long-term evaluation in North America, but long-term followup in New Zealand suggested a reduced but persistent long-term reduction in young driver collisions as a result of its implementation [3]. The collision data for Ontario drivers, around the time of the introduction of the GLS ([2], p. 126), illustrate the variety of approximations of time series that are possible.

There are many practical techniques for smoothing a time series [4]. The smoothed value at a point is a weighted average involving the elements in the series that are within a local window about the point. One way to generate the weights in a moving average involves a local polynomial approximation of order 3 (or 5) where the window includes 5 (or 7 et cetera) points. The weights are then defined by regression. Another approach defines the weights in terms of an appropriate kernel, and this method applies more generally to bivariate data [5]. The advantages of local estimates compared with global estimates are discussed in [6].

The first step in the computational process involves the partition of the domain of the time series into subdomains. The subdomains are then labelled as odd-numbered or even-numbered. Iterations are then performed over the odd numbered subdomains followed by the iterations over the even-numbered subdomains. This process is numerically efficient since the iterations over one set of subdomains update the boundary conditions for the iterations over the remaining subdomains [7]. To determine a smooth and accurate approximation, this set of iterations is repeated a few times.

This paper is organized as follows. Section 2 describes the form of the approximations, the partition of the time series, and the minimization over the subdomains of the partition. In Section 3, the equations for the approximation are derived, and some computational details are given in Section 4. Under certain assumptions, approximations of time series with variable spacing are possible. The time series and their approximations are presented in Section 5, and an example outlines the approach for step level changes, missing data, and outliers in Section 6. In Section 7, the approximation over a subdomain is determined by a fourth-order polynomial and a straight line. Finally, guidelines for the application of the proposed approach to a time series are outlined in β€œConcluding remarks.”

2. The General Model

The equation for the general model for the approximation of the time series {π‘π‘‘βˆ£π‘‘=1,…,𝑁} is𝑍𝑑=π‘›βˆ’1ξ“π‘˜=0ξ€·π‘‚π‘˜π‘‘+π‘„π‘˜π‘‘ξ€Έ+𝑅𝑛𝑑,(2.1) where 𝑄0𝑑=𝑇𝑑+𝑀𝑑 and 𝑅𝑛𝑑=π΄π‘›π‘‘π‘Šπ‘›π‘‘. In these equations, π‘‚π‘˜π‘‘ is the term for outliers (if present); π‘„π‘˜π‘‘ is the π‘˜th approximation; 𝑇𝑑 is a trend and includes the level changes; 𝑀𝑑 is a nonperiodic oscillatory function; 𝑅𝑛𝑑 is the remainder; 𝐴𝑛𝑑 is a measure of the variation of the remainder. A restriction on the approximations requires that the root-mean-square (RMS) value of the remainder is a decreasing sequence with increasing 𝑛. The form of (2.1) is similar to the asymptotic expansion of a function that contains a small positive parameter ([8], p. 1–4).

The partition of the domain of the time series {𝑍𝑑}, henceforth denoted by 𝑍𝑑, is chosen in order to accurately approximate possible patterns in the time series. Let 𝑃={πΈπ‘˜βˆ£π‘˜=1,…,𝑀} be a partition of [1,𝑁] where the nonoverlapping subdomains are πΈπ‘˜={π‘‘π‘˜βˆ’π‘›π‘˜+1,…,π‘‘π‘˜}, π‘›π‘˜β‰₯1, where π‘›π‘˜ is the number of elements in the π‘˜th subdomain πΈπ‘˜. The overlapping subdomains, over which the iterations are computed, are defined as πΈπ‘œπ‘˜={π‘‘π‘˜βˆ’π‘›π‘˜,πΈπ‘˜,π‘‘π‘˜+1}. For π‘˜=1, πΈπ‘œ1={0,𝐸1,𝑛1+1} and for π‘˜=𝑀, πΈπ‘œπ‘€={π‘βˆ’π‘›π‘,𝐸𝑁,𝑁+1} so that the overlapping subdomains are defined over the interval [0,𝑁+1].

An approximation 𝑄𝑑 of a time series 𝑍𝑑, where 𝑍𝑑=𝑄𝑑+𝑅𝑑, is determined by a few iterations 𝐼𝑛𝑑 starting with 𝐼0𝑑=𝑍𝑑. Once the desired accuracy is obtained, the last iteration is defined as 𝑄𝑑. All iterations and the approximation 𝑄𝑑 along with the remainder 𝑅𝑑 satisfy the following properties.(1)An iteration over πΈπ‘˜ has the same average value π‘Žπ‘˜ as the time series ξ“π‘‘βˆˆπΈπ‘˜π‘π‘‘=ξ“π‘‘βˆˆπΈπ‘˜πΌπ‘›π‘‘=π‘Žπ‘˜,(2.2) and, hence, the average value of the remainder 𝑅𝑑 is zero over this subdomain. If 𝑍𝑑 is a measure of β€œenergy” in the process, then 𝑄𝑑 conserves energy over each subdomain in the partition. In the particular case π‘›π‘˜=1, πΈπ‘˜={π‘‘π‘˜} and 𝑑=π‘‘π‘˜ is a fixed point for the approximation so that π‘„π‘‘βˆΆ=𝑍𝑑 at 𝑑=π‘‘π‘˜.(2)The measure of smoothness of the iterations at time 𝑑 for the 𝑛th iteration is defined by 𝛿𝑛𝑑=𝐼𝑛𝑑+1+πΌπ‘›π‘‘βˆ’1βˆ’2𝐼𝑛𝑑 which is the second difference of 𝐼𝑛𝑑 at time 𝑑. The norm on πΈπ‘œπ‘˜ is defined as the RMS value Ξ”π‘›π‘˜=βŽ›βŽœβŽœβŽβˆ‘π‘‘βˆˆπΈπ‘˜ξ€·π›Ώπ‘›π‘‘ξ€Έ2π‘›π‘˜βŽžβŽŸβŽŸβŽ 1/2.(2.3) Provided that the number of elements in πΈπ‘˜ is greater than 1, then the condition that Ξ”π‘›π‘˜ has a minimum value is imposed. 𝐼𝑛𝑑 is required for 𝑑=π‘‘π‘˜βˆ’π‘›π‘˜ and 𝑑=π‘‘π‘˜+1 in πΈπ‘œπ‘˜ to determine 𝛿𝑛𝑑 at the endpoints of πΈπ‘˜. These two values of 𝐼𝑛𝑑 are the boundary conditions for the minimization on πΈπ‘˜.(3)For most of the examples presented in this paper, 𝑑=1 and 𝑑=𝑁 are fixed points so that 𝑄1=𝑍1 (𝑛1=1) and 𝑄𝑀=𝑍𝑁 (𝑛𝑀=1). These values provide the boundary conditions for the minimization over the adjacent subdomain. For the general case where 𝑛1>1 in 𝐸1 (𝑛𝑀>1 in 𝐸𝑀), one of the boundary conditions is missing in πΈπ‘œ1 (πΈπ‘œπ‘€) so that an external boundary condition is required as described in the last paragraph of Section 3.(4)In some cases there are two or more approximations over one or more subdomains and a criterion is required to choose the best approximation. From (2.1), π‘…π‘˜π‘‘=π‘„π‘˜π‘‘+π‘…π‘˜+1t(𝑅0π‘‘βˆΆ=𝑍𝑑) where π‘„π‘˜π‘‘ is determined from π‘…π‘˜π‘‘ over a partition π‘ƒπ‘˜. For the example in Section 5, the simplest case occurs when π‘ƒπ‘˜ is a refinement of π‘ƒπ‘˜βˆ’1; that is, π‘ƒπ‘˜=βˆͺπ‘ƒβ„“π‘˜ where π‘ƒβ„“π‘˜ covers 𝐸ℓ in π‘ƒπ‘˜βˆ’1. Then an approximation over 𝐸ℓ is π‘„π‘˜π‘‘=0 and the other is defined by π‘ƒβ„“π‘˜. Let the RMS value of the remainder π‘…π‘‘π‘˜+1 over π‘ƒβ„“π‘˜ be denoted by π‘†π‘˜+1, and π‘†π‘˜ is the RMS value over 𝐸ℓ in π‘…π‘˜π‘‘. The approximation defined by π‘ƒβ„“π‘˜ is a significant improvement if the ratio π‘†π‘˜+1/π‘†π‘˜β‰€πœ– for a chosen value of πœ–. As shown in Section 7, an upper bound for πœ– takes on values between 0.75 and 0.9. For the example in Section 6 involving an outlier, there are two approximations for 𝑄0𝑑.(5)The magnitude 𝐴𝑛𝑑 of the remainder 𝑅𝑛𝑑 is defined to be the RMS value of the remainder over each subdomain in a partition, and this definition implies that the RMS value of the series π‘Šπ‘›π‘‘ in (2.1) is equal to 1. For the example presented in Section 5, the subdomains are uniform with 12 elements.

3. Mathematical Details

Given the iterates πΌπ‘‘π‘›βˆ’1 and π›Ώπ‘‘π‘›βˆ’1 on πΈπ‘œπ‘˜, the iterates 𝐼𝑛𝑑 for π‘‘βˆˆπΈπ‘˜ are computed such that the sum of squares of 𝛿𝑛𝑑 has a minimum value. For the moment, the first and last interval 𝐸1 and 𝐸𝑀 are excluded. The following variables are required to set up the equations for the minimization over the subdomain:π‘‹π‘›π‘˜=ξ‚ƒπ›Ώπ‘›π‘‘π‘˜βˆ’π‘›π‘˜+1,…,π›Ώπ‘›π‘‘π‘˜ξ‚„ξ…ž,π‘Œπ‘›π‘˜=ξ‚ƒπΌπ‘›π‘‘π‘˜βˆ’π‘›π‘˜+1,…,πΌπ‘›π‘‘π‘˜ξ‚„ξ…ž,π΅π‘›π‘˜=ξ‚ƒπΌπ‘›π‘‘π‘˜βˆ’π‘›π‘˜,0,…,0,πΌπ‘›π‘‘π‘˜+1ξ‚„ξ…ž(3.1) are π‘›π‘˜Γ—1 matrices and a prime on a matrix indicates the transposed matrix. From the definition of 𝛿𝑛𝑑 in Section 2, these matrices are related by π‘‹π‘›π‘˜βˆ’π΄π‘Œπ‘›π‘˜=π΅π‘›π‘˜, where 𝐴 is a π‘›π‘˜Γ—π‘›π‘˜ tridiagonal symmetric matrix with elements{(βˆ’2,1);(1,βˆ’2,1);…;(1,βˆ’2,1);(1,βˆ’2)},(3.2) where βˆ’2 is on the main diagonal. The equations for the iterations are obtained by replacing π΅π‘›π‘˜ with π΅π‘˜π‘›βˆ’1. Since the sum of 𝐼𝑛𝑑 for π‘‘βˆˆπΈπ‘˜ is a constant for all 𝑛, then π»ξ…žπ‘Œπ‘›π‘˜=ξ“π‘‘βˆˆπΈπ‘˜π‘π‘‘=π‘Žπ‘˜π‘›π‘˜,(3.3) where 𝐻=[1,1,…,1]ξ…ž, and π‘Žπ‘˜ is the average value of 𝑍𝑑 over πΈπ‘˜. The condition on the sum in terms of π‘‹π‘›π‘˜ is πΈξ…ž(π‘‹π‘›π‘˜βˆ’π΅π‘˜π‘›βˆ’1)=π‘Žπ‘˜π‘›π‘˜, where 𝐸 is the solution of 𝐴𝐸=𝐻. Thus, πΈβ€²π‘‹π‘›π‘˜=π‘›π‘˜πœŒπ‘˜π‘›βˆ’1 whereπœŒπ‘˜π‘›βˆ’1=π‘Žπ‘˜βˆ’ξ‚€πΌπ‘‘π‘›βˆ’1π‘˜βˆ’1+πΌπ‘‘π‘›βˆ’1π‘˜+12,(3.4) and 𝐸[1]=𝐸[π‘›π‘˜]=βˆ’π‘›π‘˜/2 (Section 7). The solution for π‘‹π‘›π‘˜ such that (π‘‹π‘›π‘˜)ξ…žπ‘‹π‘›π‘˜ has a minimum is π‘‹π‘›π‘˜=πœŒπ‘˜π‘›βˆ’1𝐺2(π‘›π‘˜)𝐸, where 𝐺2(π‘›π‘˜)=π‘›π‘˜/(𝐸′𝐸). Finally, Ξ”π‘›π‘˜=|πœŒπ‘˜π‘›βˆ’1|𝐺(π‘›π‘˜), and the solution π‘Œπ‘›π‘˜ isπ‘Œπ‘›π‘˜=πΌπ‘‘π‘›βˆ’1π‘˜βˆ’π‘›π‘˜π‘‰1+πΌπ‘‘π‘›βˆ’1π‘˜+1𝑉2+πœŒπ‘˜π‘›βˆ’1𝑉3,𝑉(3.5)1[𝑖]𝑖=1βˆ’ξ€·π‘›π‘˜ξ€Έ+1,𝑉2[𝑖]=π‘–ξ€·π‘›π‘˜ξ€Έ+1,𝐴𝑉3=𝐺2ξ€·π‘›π‘˜ξ€Έ.(3.6) The iteration 𝐼𝑛𝑑=π‘Œπ‘›π‘˜[𝑖], where 𝑖=1,…,π‘›π‘˜ and 𝑑=π‘‘π‘˜βˆ’π‘›π‘˜+1,…,π‘‘π‘˜ in πΈπ‘˜, respectively. For a given π‘›π‘˜, 𝐸 and then 𝑉3 are uniquely determined, and 𝑉3 is computed in advance for all of the subdomains that occur in a time series.

3.1. External Boundary Conditions

If 𝑑=1 is not a fixed point, then the subdomain πΈπ‘œ1 requires a boundary condition. Here are three possible external boundary conditions to impose at 𝑑=0 that can be used to reflect the possible behavior of the time series near the endpoint. (1)𝐼0π‘›βˆ’1=𝐼1π‘›βˆ’1. The slope of the tangent is zero at 𝑑=1. (2)𝐼0π‘›βˆ’1 is defined so that 𝜌1π‘›βˆ’1=0 in (3.4). This condition implies that the approximation over 𝐸1 is a segment of a straight line. (3)An iterative process is used to obtain 𝑍0 such that the RMS value of the remainder over the adjacent subdomain(s) has a minimum value.

Similarly, if 𝑑=𝑁 is not a fixed point, then the external boundary conditions are obtained by replacing 𝐼0π‘›βˆ’1, 𝐼1π‘›βˆ’1, 𝜌1π‘›βˆ’1, 𝑍0 with πΌπ‘›βˆ’1𝑁+1, πΌπ‘π‘›βˆ’1, πœŒπ‘€π‘›βˆ’1, 𝑍𝑁+1, respectively.

4. Computational Aspects

For a time series 𝑍𝑑 and a partition 𝑃, there is a related series defined by Z𝑑=π‘Žπ‘˜ for π‘‘βˆˆπΈπ‘˜, where π‘Žπ‘˜ is the average value of 𝑍𝑑 for π‘‘βˆˆπΈπ‘˜. This property holds for all of the approximations in this paper,

The approximations for a time series 𝑍𝑑 and the averaged time series Z𝑑 are the same to the desired accuracy provided that the same partition and the same external boundary conditions (if any) are applied.

Consequently, any time series with variable spacing can be approximated provided that the estimates of the average values of the time series over the subdomains are adequate.

The approximation for the averaged series is employed especially for larger subdomains (π‘›π‘˜β‰ˆ12 or more). The efficiency of the computations is increased if the boundary conditions in the first set of iterations (even and odd) are the average of the four values of the series that straddle the subdomains πΈπ‘˜βˆ’1 and πΈπ‘˜. The averaged series was used in all computations, although the approximation obtained from 𝑍𝑑 may be more efficient in special cases.

It is convenient to introduce another notation to represent a partition: 𝑃 = {𝑛1,𝑛2,…;…;…,𝑛𝑀}, where the number of elements in the subdomains in the first block is {𝑛1,𝑛2,…} and in the last block by {…,𝑛𝑀}. These blocks are a convenient way to separate the seasons or a set of months. Also, the approximation π‘„π‘˜π‘‘ obtained by iterating the time series 𝑛 times, using the partition π‘ƒπ‘˜, is denoted by π’«π‘›π‘˜{π‘…π‘˜π‘‘} (𝑅0π‘‘βˆΆ=𝑍𝑑). The number of iterations 𝑛 is determined from the difference𝐷𝑛𝑑=π’«π‘›π‘˜ξ€½π‘…π‘˜π‘‘ξ€Ύβˆ’π’«π‘˜2π‘›ξ€½π‘…π‘˜π‘‘ξ€Ύ(4.1) by imposing the condition that max(|𝐷𝑛𝑑|)<𝐿. 𝐿=1 in Figures 1 and 2, 𝐿=0.04 in Figures 3 and 4, and 𝐿=0.01 in Figure 5. All calculations in this paper were performed using Maple software [9].

5. Applications

Two time series, provided by the Ontario Ministry of Transportation ([2], p. 126), illustrate the approximations. The graph of the time series for the monthly accidents for young novice drivers is given in Figure 1 where the main feature here is the intervention that occurs at 52 months owing to the introduction of the GLS on April 1, 1994. The corresponding graph for all drivers is shown in Figure 3 where the sharp drop in the graph from the maximum in December/January to April, except for the last 2 years, is a strong feature of the series.

In Figure 1, the uniform partition 𝑃0 = {1,3;4;…;4;3,1}, except for the first and last blocks, provides a smooth approximation 𝑄0𝑑=𝒫40{𝑍𝑑} and captures the intervention well. In this case, the first and last elements are fixed (𝑄01=𝑍1 and 𝑄0120=𝑍120). Since the sum of the elements in the remainder over πΈπ‘˜ is zero, then 𝒫0{𝑅1𝑑}={0}. Other uniform partitions are possible where there are 3 or 6 elements in each subdomain. The approximation in the former case is slightly less smooth than 𝑄0𝑑 in Figure 1, and the RMS value of the remainder is 26. For the case of 6 elements, the RMS value of the remainder is 30. A more accurate approximation is obtained if the subdomains have two elements; however, this approximation has an angular appearance since it more closely approximates the time series.

In Figure 2, the approximation 𝑄0𝑑 of Figure 1 is expressed as a sum of a trend and an oscillatory series. The partition for the trend 𝑇𝑑=𝒫6𝑇{𝑄0𝑑} is 𝑃𝑇 = {12,12,12,12;6,6,6,6;12,12,12,12}. The external boundary condition implies that the tangent is horizontal at the endpoints of the series. The trend in this example is defined as a seasonal approximation of the time series where the subdomains contain 6 elements over the domain of the intervention. The remainder is the oscillatory series 𝑀𝑑=𝑄0π‘‘βˆ’π‘‡π‘‘.

In Figure 3, the points for January or December (plus one November) and April are fixed points for the approximation 𝑄0𝑑=𝒫30{𝑍𝑑}. The partition is 𝑃0={1,2,1,7,1;3,1,7,1;3,1,7,1;3,1,8;1,2,1,7,1;3,1,6,2;1,2,1,8;1,2,1,6,1;2,2,1,8;1,2,1,7,1}. The second approximation captures the increase in the number of accidents that occur in the summer months by approximating the remainder 𝑅1𝑑 in 𝑍𝑑=𝑄0𝑑+𝑅1𝑑 to obtain 𝑅1𝑑=𝑄1𝑑+𝑅2𝑑 where 𝑄1𝑑=𝒫31{𝑅1𝑑}. The partition 𝑃1 is a refinement of 𝑃0 where 7 is replaced with 3,4;6 with 3,3; 8 with 3,5. Consequently, the approximations 𝑄0𝑑 and 𝑄0𝑑+𝑄1𝑑 have the same average value over the subdomains of 𝑃0.

The subdomains that are the same in the two partitions 𝑃0 and 𝑃1 are indicated by the intervals over which the approximation is zero in Figure 3. For the remaining intervals, the ratio of the RMS value of the remainder 𝑅2𝑑 in Figure 4 to the RMS value of 𝑅1𝑑 is equal to 0.49 so that this second approximation is significant. Furthermore, each of the ten segments of 𝑄1𝑑, excluding the segments in which the approximation is identically equal to zero, has a value between 0.40 and 0.52. The approximation 𝑄0𝑑+𝑄1𝑑 of 𝑍𝑑 and the remainder 𝑅2𝑑 appear in Figure 4.

6. Level Changes, Missing Data, and Outliers

For a step level change between 𝑑=𝜏 and 𝑑=𝜏+1, an approximation may not provide an adequate approximation for the time series in the subdomains on both sides of the step. For 𝑑>𝜏, an external boundary condition is applied at 𝑑=𝜏 such that the remainder of the approximation has a minimum RMS value. The same approach is applied to the series for 𝑑<𝜏+1. These ideas are illustrated in Figure 5 where the partition for the approximation 𝑄𝑑=𝒫6{𝑇𝑑} over [1,120] is 𝑃 = {1,11;12;…;12;11,1}. The approximation exhibits a phenomenon that is similar to a Fourier series near a discontinuity in that the approximation overshoots on the right and undershoots to the left of the jump. The maximum and minimum of 𝑄𝑑 are 1.176 and βˆ’0.173. Moving away from the jump in either direction, the oscillations of π‘„π‘‘βˆ’π‘‡π‘‘ occur with rapidly decreasing amplitude. The details in item 4 of Section 2 are applied over the subdomains adjacent to 𝑑=60 to choose between the two approximations of the trend. Interventions and level shifts, from an autoregressive moving average point of view, are presented in [10] and [11].

A simple example indicates the approach for a series that has a missing value or a possible outlier at 𝑑=6. The series is {𝑍𝑑}= {0.6,βˆ’0.3,βˆ’0.5,βˆ’0.2,0.4,𝑍6,1,0.8,0.9,1.1,1.0,1.2} where 𝑍6 is not defined in the case of a missing value. For both cases, the approximation 𝑄0𝑑 is determined for the partition 𝑃= {1;4;1;5;1}, where the value of the series is 𝑋 at 𝑑=6, such that the RMS value of the remainder is a minimum. A good initial estimate for 𝑋 is the average value of the time series in a window about 𝑑=6 [5]. Then an iterative process is started to obtain π‘‹β‰ˆ1.25, as shown in Figure 6, and the RMS value of the remainder is 0.104. The smoothest approximation over 𝑃 occurs for π‘‹β‰ˆ0.12, where βˆ‘211𝛿2𝑑 has a minimum value. For the case of a possible outlier, 𝑂0𝑑=0 for 𝑑≠6 in (2.1) and 𝑂06=𝑍6βˆ’1.25 provided that the ratio of the RMS value of the approximation with 𝑋=1.25 to the RMS value of the approximation under the assumption that 𝑍6 is not an outlier satisfies the condition in item 4 of Section 2.

7. Properties of Approximations

Approximations of Random Samples
The point of this exercise is to determine the πœ– in item 4 of Section 2 such that the only reasonable approximation for a series of random samples is 𝑄0𝑑 equal to the mean of the series. 12,000 random samples from the normal distribution with a mean of 0 and a standard deviation of 1 were generated using Maple to form 100 time series with 120 elements in each series. For each series, five approximations were determined where the subdomains of the uniform partition contained 3, 4, 6, 12, and 24 elements. The external boundary condition for the approximation is the condition of zero slope of the tangent. For each series, the ratio of the RMS value of the remainder for the approximation to the RMS value of the series were calculated, and the results are given in Table 1. The approximations corresponding to 24 and 12 elements are smooth and appear to reflect an underlying pattern in the series; whereas, for the cases 3 and 4, the approximations are contorted. An upper bound for πœ– is less than the minimum values in the range.

Quartic Polynomial
The terms in the equation (3.5) for the approximation over πΈπ‘œπ‘˜ have a simple interpretation. For the first two terms in (3.5), (𝑖,𝑉1[𝑖]) and (𝑖,𝑉2[𝑖]) are points on straight lines. For the last term, Maple solves 𝐴𝐸=𝐻 for 𝐸 and (3.6) for 𝑉3 exactly; consequently, an accurate computation shows that (𝑖,𝐸[𝑖]) and (𝑖,𝑉3[𝑖]) are points on the graph of a quadratic and a quartic polynomial, respectively.
To describe the properties of 𝐸 and 𝑉3, it is necessary to change variables from 𝑑 to 𝑠 where 𝑠=π‘‘βˆ’[π‘‘π‘˜βˆ’(π‘›π‘˜βˆ’1)/2] and 𝑠=0 is the central point of the interval πΈπ‘œπ‘˜. In the 𝑠 variable, the boundary conditions are applied at the points 𝑠=Β±π‘ π‘œ, where π‘ π‘œ=(π‘›π‘˜+1)/2. The equation of the quadratic polynomial 𝑣(𝑠,π‘›π‘˜) for 𝐸 is defined by 𝑑2𝑣/𝑑𝑠2=1 and 𝑣=0 at 𝑠=Β±π‘ π‘œ so that 𝑣=(𝑠2βˆ’π‘ 2π‘œ)/2. For any integer 𝑖, 𝐸[𝑖] is equal to 𝑣 at the corresponding value for 𝑠, and 𝐸′𝐸=((2π‘ π‘œ)5βˆ’2π‘ π‘œ)/120. The equation for the quartic polynomial 𝑒(𝑠,π‘›π‘˜) for 𝑉3, provided π‘›π‘˜β‰₯3, is determined by 𝑑4𝑒/𝑑𝑠4=𝐺2(π‘›π‘˜), where the roots of the equation for 𝑒 are 𝑠=Β±π‘ π‘œ and 𝑠2=5𝑠2π‘œ+1. Thus, 𝑒=𝐺2(π‘›π‘˜)(𝑠2βˆ’π‘ 2π‘œ)(𝑠2βˆ’5𝑠2π‘œβˆ’1)/24. 𝐺(π‘›π‘˜) is a measure of the smoothness of the approximation over πΈπ‘œπ‘˜: G(2) = 1.0, G(3) = 0.59, G(4) = 0.39, G(6) = 0.21, G(12) = 0.062, and G(24) = 0.017.

Concluding Remarks
The major input for the approximation of a time series involves the partition of the domain. Initially a uniform partition is chosen and, if seasonal behavior is present in the series, a subset of the partitions cover the domain for the seasons. In general, as the length of the subintervals decreases, the approximation is less smooth and the accuracy of the approximation increases. The best approximation occurs at the point at which the approximation is acceptably smooth. The subintervals can be enlarged to determine a much smoother approximation that can be labelled as a trend while still respecting the seasonal aspects of the series; however, if an intervention is present, then some adjustment of the partition may be necessary in the region of the intervention. For time series with a well-defined local maximum or minimum, the approximation can be assigned the same value as the series by taking the partition to be a single point of the domain. For series with jumps and other complexities, examples are provided to suggest how to proceed in these cases.
An approach in the literature, as indicated in the introduction, defines the approximation at a point as a weighted average of the values of the values of the time series in a window about the point. This approach may smooth out interesting features in the time series and, if applied over a smaller intervals, the approximation will not be smooth. Since the proposed model is not based on regression, a comparison of the two approaches has not been considered.