Abstract

The classical linear filter can successfully filter the components from a time series for which the frequency content does not change with time, and those nonstationary time series with time-varying frequency (TVF) components that do not overlap. However, for many types of nonstationary time series, the TVF components often overlap in time. In such a situation, the classical linear filtering method fails to extract components from the original process. In this paper, we introduce and theoretically develop the G-filter based on a time-deformation technique. Simulation examples and a real bat echolocation example illustrate that the G-filter can successfully filter a G-stationary process whose TVF components overlap with time.

1. Introduction and Background

In this paper we develop filters for G-stationary processes, which are nonstationary processes with time-varying frequency (TVF) behavior. We begin with a very brief discussion of linear filters since the filters we develop here are generalizations of the linear filter. The traditional linear filter is defined as βˆ«π‘Œ(𝑑)=π‘‹βˆ—π‘Ž(𝑑)=+βˆžβˆ’βˆžπ‘‹(π‘‘βˆ’π‘’)π‘Ž(𝑒)𝑑𝑒,whereπ‘‘βˆˆ(βˆ’βˆž,∞). Letting 𝑃𝑋(𝑓)and π‘ƒπ‘Œ(𝑓) denote the power spectra of the stationary input and output processes 𝑋(𝑑)and π‘Œ(𝑑), respectively, the fundamental filtering theorem states that π‘ƒπ‘Œ(𝑓)=|𝐴(𝑓)|2𝑃𝑋(𝑓), where ∫𝐴(𝑓)=βˆžβˆ’βˆžπ‘Ž(𝑑)π‘’βˆ’2πœ‹π‘–π‘“π‘‘π‘‘π‘‘is the frequency response function [1]. The importance of this result is that it provides information concerning how the filter, as represented by the squared frequency response function, |𝐴(𝑓)|2, impacts the frequency behavior of the output signal. This can be useful in designing filters with specific properties. Linear filters are commonly used to β€œfilter out” certain frequencies from a time series realization. In this discussion we will consider low-pass and high-pass filters which are designed to remove the high-frequency behavior and low frequencies, respectively. We will use the popular Butterworth filter (see [2]). The squared frequency response function of an 𝑁th order low-pass Butterworth filter is given by||||𝐴(𝑓)2=1ξ€·1+𝑓/𝑓𝑐2𝑁,(1.1) where 𝑓𝑐 is the cutoff frequency, and 𝑁 is the order of the filter. The frequency response function for a high-pass Butterworth filter has a similar form. Traditional linear filters such as the Butterworth filter can successfully extract components from stationary processes where the frequency behavior of the data does not change with time. However, for many nonstationary time series with time varying frequency (TVF) behavior, the time-invariant assumption of traditional filters often results in the failure to extract TVF components from such processes. Example 1.1 examines the application of the Butterworth filter on data containing time-varying frequencies.

Example 1.1 (linear filter applied to TVF data). Figure 1(a) shows a realization of length 𝑛=400 from the model 𝑋(𝑑)=cos36πœ‹ln(𝑑+175)+πœ“1ξ€»ξ€Ί+.5cos85πœ‹ln(𝑑+175)+πœ“2ξ€»,(1.2) where πœ“1=1 and πœ“2=1.59. The TVF behavior is clear, both in the data in Figure 1(a) and the two components that are shown in Figures 1(b) and 1(c). However, it should be noted that the frequency content at the beginning of the β€œlow frequency” component (Figure 1(b)) is about the same as that toward the end of the β€œhigh frequency” component seen in Figure 1(c). For this reason, a Butterworth-type filter will not be able to completely separate the components. In Figures 1(d) and 1(e) we show the results of 3rd order low-pass Butterworth filters applied to the data in Figure 1(a) with 𝑓𝑐=.15 and .06, respectively. In Figure 1(d) the cutoff frequency, 𝑓𝑐=.15,does a good job of extracting the low-frequency behavior for time 𝑑≀100, but toward the end of the signal, both components pass through the filter. This occurs since the frequency behavior (which decreases with time) in the β€œhigh frequency” component has decreased to the point that it passes through the filter with 𝑓𝑐=.15. Figure 1(e) shows the effects of lowering the cut off frequency to 𝑓𝑐=.06. For 𝑑β‰₯250 the filter does a reasonable job of extracting only the low-frequency component. However, the effect of lowering the cut-off frequency is that neither the high- nor low-frequency componentpasses through the filter toward the beginning of the data. In fact, because of the overlap in frequency content between the first part of the signal in Figure 1(b) and the latter part of the signal in Figure 1(c), no cut-off frequency will be able to separate these two components using a filter such as the Butterworth.
The data set in Example 1.1 is a realization from an M-stationary process which is a special case of the G-stationary processes that have been introduced to generalize the concept of stationarity [3]. As illustrated above, current time-invariant linear filtering would have to be adjusted in order to filter the data from a G-stationary processes. Baraniuk and Jones [4] used the theory of unitary equivalence systems for designing generic classes of signal analysis and processing systems based on alternate coordinate systems. Applying the principle of unitary equivalence systems, the filtering procedure is to (1) preprocess the data, (2) apply a traditional time-invariant filtering method, and (3) convert the results to the original time scale. To solve the filtering problem for G-stationary processes, in this paper we define the G-filter and illustrate its application on signals that originated from G-stationary processes.
The paper is organized as follows. In Section 2 we discuss G-stationary models with focus on strategies for fitting 𝐆(𝝀)-stationary models of Jiang, et al. [3]. In Section 3 we present results concerning G-filters designed for filtering data from G-stationary models, and in Section 4 we introduce a straight-forward implementation of the G-filter for extracting components such as those shown in Figures 1(b) and 1(c).

2. Time Deformation Method and the G-Stationary Process

A stationary process is a stochastic process in which the distribution does not change with time. A large volume of statistical theory and methods have been developed for stationary time series. However, many processes are nonstationary where the distribution changes with time. Several techniques have been proposed for analyzing these nonstationary processes using time deformation. For example, Girardin and Senoussi [5] presented a framework based on harmonic analysis for semigroups, where they defined a semi-group stationary process including a local stationary process and a time-warped process. Clerc and Mallat [6] derived an estimator of the deformation by showing that the deformed autocovariance satisfies a transport equation at small scales. Among the different types of nonstationary processes, special attention has been given to warped nonstationary processes that are obtained by deforming the index set of stationary processes. For example, self-similar processes are obtained by applying the Lamperti operator on the time scale of a stationary process [7]. Hlawatsch and TaubΓΆck [8] and Hlawatsch et al. [9] discuss the use of time frequency displacement operators to displace signals in the time-frequency plane. See also Sampson and Guttorp [10] and Perrin and Senoussi [11].

Gray and Zhang [12] base time deformation on a log transformation of the time axis. They refer to processes that are stationary on the log scale as M-stationary processes and show that the resulting spectral representation is based on the Mellin transform. Flandrin et al. [7] point out that the log transformation of Gray and Zhang [12] is a special case of the Lamperti transform. Gray et al. [13] extend the M-stationary model to analyze data collected at discrete time points. Jiang et al. [3] defined a more general family of nonstationary processes called G-stationary processes whose frequencies monotonically change with time. See also Woodward et al. [14, Chapter  13]. In this paper, we investigate the use of time deformation based on G-stationary models to filter TVF data. We refer to the resulting filters as G-filters.

2.1. G-Stationary Processes

Definition 2.1. Let {𝑋(𝑑)βˆΆπ‘‘βˆˆπ‘†} be a stochastic process defined on π‘†βŠ‚β„, let 𝑒=𝑔(𝑑) be a mapping onto a set π‘…π‘”βŠ‚β„, and let π‘”βˆ’1 denote an inverse. Then 𝑋(𝑑) is said to be a G-stationary process if (i)𝐸[𝑋(𝑑)]=πœ‡, (ii)Var𝑋(𝑑)=𝜎2<∞, (iii)𝐸[(𝑋(𝑑)βˆ’πœ‡)(𝑋(π‘”βˆ’1(𝑔(𝑑)+𝑔(𝜏)))βˆ’πœ‡)]=𝑅𝑋(𝜏).

Although not immediately obvious, this definition basically says the following. Let 𝑔 be a transformation of the time axis and let 𝑔(𝑑)=𝑒and 𝑑=π‘”βˆ’1(𝑒). Consequently, 𝑋(𝑑)=𝑋(π‘”βˆ’1(𝑒))=π‘Œ(𝑒), whereπ‘Œ=π‘‹π‘”βˆ’1. Letting πœ‰=𝑔(𝜏), then π‘‹ξ€·π‘”βˆ’1ξ€Έ(𝑔(𝑑)+𝑔(𝜏))=π‘Œ(𝑔(𝑑)+𝑔(𝜏))=π‘Œ(𝑒+πœ‰).(2.1) So, it follows that 𝑅𝑋𝑋𝑔(𝜏)=𝐸(𝑋(𝑑)βˆ’πœ‡)βˆ’1ξ€Έ[],(𝑔(𝑑)+𝑔(𝜏))βˆ’πœ‡ξ€Έξ€»=𝐸(π‘Œ(𝑒)βˆ’πœ‡)(π‘Œ(𝑒+πœ‰)βˆ’πœ‡)(2.2) that is, Definition 2.1 gives the conditions on 𝑋(𝑑) and 𝑔(𝑑) that are necessary and sufficient in order for the dual process, π‘Œ(𝑒), to be weakly stationary. The implication is that, whereas 𝑋(𝑑) may not be stationary on the original index set (on 𝑑), it can be mapped onto a new (deformed) time scale on which it is stationary. We will refer to π‘Œ(𝑒)as the stationary dual of 𝑋(𝑑),and we let πΆπ‘Œ(πœ‰)=𝐸[(π‘Œ(𝑒)βˆ’πœ‡)(π‘Œ(𝑒+πœ‰)βˆ’πœ‡)] denote the autocovariance of the dual process. 𝑅𝑋(𝜏) is called the G-stationary autocovariance, and clearly πΆπ‘Œ(πœ‰)=𝑅𝑋(𝜏). The G-spectrum is defined in Definition 2.2.

Definition 2.2. Let {𝑋(𝑑),π‘‘βˆˆ(π‘Ž,𝑏)} be a G-stationary process and let {π‘Œ(𝑒),π‘’βˆˆ(βˆ’βˆž,∞)} be the corresponding dual with respect to the time-deformation function, 𝑔(𝑑). The G-spectrum of the G-stationary process,𝑋(𝑑), is defined as the spectrum of the dual process, that is, πΊπ‘‹ξ€œ(𝑓;𝑔)=βˆžβˆ’βˆžπ‘’βˆ’2πœ‹π‘–π‘“πœ‰π‘…π‘Œ(πœ‰)π‘‘πœ‰.(2.3) If the mapping, 𝑔(𝑑), from the original space π‘‘βˆˆ(π‘Ž,𝑏) to the dual space π‘’βˆˆ(βˆ’βˆž,∞)is onto, then πΊπ‘‹ξ€œ(𝑓;𝑔)=βˆžβˆ’βˆžπ‘’βˆ’2πœ‹π‘–π‘“πœ‰π‘…π‘Œ=ξ€œ(πœ‰)π‘‘πœ‰βˆžβˆ’βˆžπ‘’βˆ’2πœ‹π‘–π‘“πœ‰πΆπ‘‹ξ€·π‘”βˆ’1ξ€Έ=ξ€œ(πœ‰);π‘”π‘‘πœ‰π‘π‘Žπ‘’βˆ’2πœ‹π‘–π‘“π‘”(𝜏)𝐢𝑋(𝜏;𝑔)π‘”ξ…ž(𝜏)𝑑(𝜏).(2.4)
When𝑔(𝑑)=π‘Žπ‘‘+𝑏,π‘‘βˆˆ(βˆ’βˆž,∞), the G-stationary process, 𝑋(𝑑), is simply the traditional weakly stationary process, and when𝑔(𝑑)=ln(𝑑), π‘‘βˆˆ(0,∞), 𝑋(𝑑)is called an M-stationary process [13]. When 𝑔(𝑑) is the Box-Cox transformation, 𝑑𝑔(𝑑)=πœ†βˆ’1πœ†(2.5)π‘‘βˆˆ(0,∞), then 𝑋(𝑑)is called a G(πœ†)-stationary process [3]. When 𝑔(𝑑)=π‘Žπ‘‘2+𝑏𝑑, π‘‘βˆˆ(0,∞)withπ‘Ž>0and𝑏β‰₯0, then 𝑋(𝑑)is called a linear chirp stationary process. See Liu [15], and Robertson et al. [16].

2.2. General Strategy for Analyzing G-Stationary Processes

We give the following outline for analyzing G-stationary processes.(1)Transform the time axis to obtain a stationary dual realization, (we will discuss this below).(2)Analyze the transformed (dual) realization using methods for stationary time series. For example, sometimes this is done by fitting an AR(𝑝) or an ARMA(𝑝,π‘ž) model to the dual data and then computing forecasts, spectral estimates, and so forth as desired for the problem at hand (in the current setting we will be filtering the dual data).(3)Transform back to original time scale as needed.

Step 1 (transform the time axis to obtain a stationary dual realization). Finding a time transformation, 𝑒=𝑔(𝑑), is equivalent to determining a sampling scheme such that 𝑋(π‘‘π‘˜)=𝑋(π‘”βˆ’1(π‘’π‘˜))=π‘Œπ‘˜, where π‘Œπ‘˜ is stationary. That is, for a G-stationary process, the stationary dual process is obtained theoretically by sampling 𝑋(𝑑) at the points 𝑑=π‘‘π‘˜=π‘”βˆ’1(π‘’π‘˜). In general, however, the observed data will have been obtained at equally spaced points, and are not available at the points π‘‘π‘˜. Interpolation has primarily been used to deal with this problem. See Gray et al. [13], Jiang et al. [3], and Woodward et al. [14]. For an approach based on Kalman filtering, see Wang et al. [17].
In the following discussion we will discuss techniques for fitting a G(πœ†)-stationary model to observed TVF data. The G(πœ†) model is sufficiently general to include TVF data with either increasing or decreasing frequency behavior. The values, π‘‘π‘˜, needed for obtaining the stationary dual based on the discrete Box-Cox time transformation, π‘˜=(π‘‘πœ†π‘˜βˆ’1)/(Ξ”π‘›πœ†)βˆ’πœ‰, are π‘‘π‘˜=((π‘˜+πœ‰)Ξ”π‘›πœ†+1)1/πœ†, where (πœ‰Ξ”π‘›πœ†+1)1/πœ† is called the realization offset and Δ𝑛 is the sampling increment. Jiang et al. [3] employ a search routine to determine the values of πœ† and offset that produce the β€œmost stationary” dual. For each set of πœ† and offset values considered in the search, the data, 𝑋(π‘‘π‘˜), π‘˜=1,2,…, are approximated at the π‘‘π‘˜'s using interpolation. By then indexing on π‘˜, the dual realization associated with the given πœ† and offset is obtained, and for each combination of πœ† and offset, the resulting dual realization is checked for stationarity. Jiang, et al. [3] suggest measuring stationarity by examining the characteristics (e.g., sample autocorrelations) of the first and second halves of the transformed data. They employ a Q-statistic for measuring the difference between the sample autocorrelations of the two halves. This measure is based on the fact that the correlation structure under stationarity stays constant across the realization. The software, GWS, written in S+, can be used to perform this search procedure, and it is available from the authors at the website http://www.texasoft.com/atsa. This software can be used to fit a G(πœ†) model to a set of data, and it provides methods for spectral analysis, forecasting, and so forth. In the examples here involving analysis of TVF data, we use the GWS software package.

Example 2.3 (G(πœ†)-analysis of the TVF data in Example 1.1). In this example we perform G(πœ†) analysis on the time series discussed in Example 1.1. We also use Wigner-Ville plots, which display the time-frequency behavior in the data by computing inverse Fourier transforms of windowed versions of the autocovariance function (see [14, 18, 19]). As previously noted, 𝑋(𝑑) in (1.2) is an M-stationary process. Figure 2(a) shows the data previously shown in Figure 1(a), and in Figure 2(c) we show the associated Parzen spectral density. Figure 2(c) is a so-called β€œspread spectrum” showing a wide range of frequency behavior between about 𝑓=.06 to 𝑓=.22 which is caused by the frequency changes in the data that were noted in Example 1.1. The frequency change can be visualized in the Wigner-Ville plot in Figure 2(e). In the plot, darker shading corresponds to locations of stronger frequency behavior.For example, Figure 2(a) shows a very pronounced lower TVF component which is illustrated with the lower β€œstrip” that is at about 𝑓=.1 at the beginning of the data whereas by 𝑑=300 the period length associated with the lower TVF component is about 20 (𝑓=.05) which is visually consistent with the component shown in Figure 1(b). The higher TVF component, represented by a lighter strip, indicates frequency at about 𝑓=.22 (periods of about 5) early in the data which decreases to about 𝑓=.1 (periods of about 10) by about 𝑑=300. Again, this is consistent with Figure 1(c). The checkered pattern between these two strips is interference and is not indicative of strong frequency behavior. The Wigner-Ville plot visually illustrates the fact mentioned in Example 1.1, that the lower frequency behavior (the bottom strip) has higher frequency at the beginning of the data than does the higher frequency component (upper strip) at the end of the data set. That is, a horizontal line cannot be drawn that stays entirely within the two strips. Using the GWS software, it is seen that a G(πœ†) transformation with πœ†=0 (M-stationary) and offset 175 is a reasonable choice for producing a stationary dual. The dual data based on this transformation is given in Figure 2(b). The high-frequency and low-frequency behavior is similar to that seen in Figure 2(a) except that neither frequency displays time-varying behavior. The corresponding spectral density in Figure 2(d) shows two distinct peaks (one near .07 and the other near .13). That is, the spectral density of the transformed data (i.e., the G-spectral density) clearly shows that there are two dominant β€œfrequency” components in the data. The Wigner Ville plot in Figure 2(f) shows two parallel lines, one at about 𝑓=.07 and another at about 𝑓=.13. This plot conveys the fact that the frequency behavior is not changing with time, which is consistent with stationary data.

3. The G-Filter

In this section we define the G-filter for filtering G-stationary data. The G-filter can be viewed as a specific example of a unitary equivalent system for filtering G-stationary signals (see [4]).

Definition 3.1. Given a stochastic process{𝑋(𝑑),π‘‘βˆˆ(π‘Ž,𝑏)} and an impulse response function, β„Ž(𝑑), π‘‘βˆˆ(π‘Ž,𝑏), the G-filter or G-convolution, denoted π‘‹βŠ—β„Ž(𝑑), is defined as π‘Œ(𝑑)=π‘‹βŠ—β„Ž(𝑑)β‰‘π‘‹π‘‘βˆ—β„Žπ‘‘(𝑔(𝑑)),(3.1) where 𝑋𝑑(𝑒)=𝑋(π‘”βˆ’1(𝑒)) and β„Žπ‘‘(𝑒)=β„Ž(π‘”βˆ’1(𝑒)), π‘’βˆˆ(βˆ’βˆž,∞) are the duals of 𝑋(𝑑) and β„Ž(𝑑)with respect to the time-deformation function𝑔(𝑑), andπ‘‹π‘‘βˆ—β„Žπ‘‘(𝑑)is the usual convolution, π‘‹π‘‘βˆ—β„Žπ‘‘βˆ«π‘‹(𝑑)=𝑑(𝜏)β„Žπ‘‘(π‘‘βˆ’πœ)π‘‘πœ.

Theorem 3.2. If the mapping 𝑔(𝑑) from the original space π‘‘βˆˆ(π‘Ž,𝑏) to the dual space π‘’βˆˆ(βˆ’βˆž,∞)is onto, then ξ€œπ‘Œ(𝑑)=π‘‹βŠ—β„Ž(𝑑)=π‘π‘Žξ€·π‘”π‘‹(𝑣)β„Žβˆ’1𝑔(𝑔(𝑑)βˆ’π‘”(𝑣))ξ…ž(𝑣)𝑑𝑣.(3.2)

Proof. Consider the following: =ξ€œπ‘Œ(𝑑)=π‘‹βŠ—β„Ž(𝑑)βˆžβˆ’βˆžπ‘‹π‘‘(𝑒)β„Žπ‘‘(=ξ€œπ‘”(𝑑)βˆ’π‘’)π‘‘π‘’π‘π‘Žπ‘‹π‘‘(𝑒)β„Žπ‘‘=ξ€œ(𝑔(𝑑)βˆ’π‘”(𝑣))𝑑𝑔(𝑣)π‘π‘Žξ€·π‘”π‘‹(𝑣)β„Žβˆ’1ξ€Έ=ξ€œ(𝑔(𝑑)βˆ’π‘”(𝑣))𝑑𝑔(𝑣)π‘π‘Žξ€·π‘”π‘‹(𝑣)β„Žβˆ’1𝑔(𝑔(𝑑)βˆ’π‘”(𝑣))ξ…ž(𝑣)𝑑𝑣.(3.3)

Corollary 3.3. Since π‘‹π‘‘βˆ—β„Žπ‘‘(𝑒)=β„Žπ‘‘βˆ—π‘‹π‘‘(𝑒), where 𝑒=𝑔(𝑑), it follows immediately that π‘‹βŠ—β„Ž(𝑑)=β„ŽβŠ—π‘‹(𝑑).

Theorem 3.4. Let {𝑋(𝑑),π‘‘βˆˆ(π‘Ž,𝑏)} be a G-stationary input process with the time-deformation function𝑔(𝑑), then(a)the output process, π‘Œ(𝑑)=π‘‹βŠ—β„Ž(𝑑), π‘‘βˆˆ(π‘Ž,𝑏), where β„Ž(t)is the impulse response function, is also G-stationary, (b)the G-spectra of 𝑋(𝑑)and π‘Œ(𝑑)satisfy πΊπ‘Œ||π‘Š(𝑓;𝑔)=β„Ž||(𝑓;𝑔)2𝐺𝑋(𝑓;𝑔),(3.4) where π‘Šβ„Ž(𝑓;𝑔)=π΄β„Žπ‘‘βˆ«(𝑓)=βˆžβˆ’βˆžπ‘’βˆ’2πœ‹π‘–π‘“π‘’β„Žπ‘‘(𝑒)𝑑𝑒 is the frequency response function of β„Žπ‘‘(𝑒).

Proof. (a) Let {𝑋𝑑(𝑒),π‘’βˆˆ(βˆ’βˆž,∞)} be the stationary dual process of 𝑋(𝑑) and {β„Žπ‘‘(𝑒),π‘’βˆˆ(βˆ’βˆž,∞)} be the dual of β„Ž(t). Since 𝑋𝑑(𝑒) is stationary, then π‘Œπ‘‘(𝑒)=π‘‹π‘‘βˆ—β„Žπ‘‘(𝑒) is stationary. Thus π‘Œ(𝑑)=π‘‹βŠ—β„Ž(𝑑)=π‘‹π‘‘βˆ—β„Žπ‘‘(𝑔(𝑑))=π‘Œπ‘‘(𝑔(𝑑)),π‘‘βˆˆ(π‘Ž,𝑏)(3.5) is G-stationary.
(b) Let 𝑃𝑋𝑑(𝑓) andπ‘ƒπ‘Œπ‘‘(𝑓) be the spectra of the stationary dual of the processes 𝑋𝑑(𝑒)andπ‘Œπ‘‘(𝑒), that is, the G-spectra of 𝑋(𝑑) and π‘Œ(𝑑). Since π΄β„Žπ‘‘(𝑓) is the frequency response function ofβ„Žπ‘‘(𝑒), then based on the fundamental linear filtering theorem, it follows thatπ‘ƒπ‘Œπ‘‘(||𝐴𝑓)=β„Žπ‘‘(||𝑓)2𝑃𝑋𝑑(𝑓),thatis,πΊπ‘Œ(||π‘Šπ‘“;𝑔)=β„Ž(||𝑓;𝑔)2𝐺𝑋(𝑓;𝑔).(3.6) If the mapping 𝑔(𝑑) from the original space π‘‘βˆˆ(π‘Ž,𝑏) to the dual space π‘’βˆˆ(βˆ’βˆž,∞) is onto, then π‘Šβ„Ž(𝑓;𝑔)=π΄β„Žπ‘‘βˆ«(𝑓)=βˆžβˆ’βˆžπ‘’βˆ’2πœ‹π‘–π‘“π‘’β„Žπ‘‘βˆ«(𝑒)𝑑𝑒=π‘π‘Žπ‘’βˆ’2πœ‹π‘–π‘“π‘”(𝑣)β„Ž(𝑣)π‘”ξ…ž(𝑣)𝑑𝑣.

3.1. The M-Filter

Gray and Zhang [12] introduced the M-convolution or M-filter for filtering the M-stationary process. When 𝑔(𝑑)=ln(𝑑), π‘‘βˆˆ(0,∞), then it follows that ξ€œπ‘Œ(𝑑)=∞0𝑔𝑋(𝑣)β„Žβˆ’1ξ€Έξ€œ(𝑔(𝑑)βˆ’π‘”(𝑣))𝑑𝑔(𝑣)=∞0𝑑𝑋(𝑣)β„Žπ‘£ξ‚π‘‘ln(𝑣)=𝑋#β„Ž(𝑑),(3.7) which is the M-convolution defined by Gray and Zhang (1988).

4. Filtering Data Using the G-Filter

Definition 3.1 shows that the G-filter is not a linear filter if viewed from the original space, but it is a linear filter if viewed from the dual space. Consequently, based on the definitions and results concerning G-filters in Section 3, we use the strategy proposed by Baraniuk and Jones [4]: (1)Make an appropriate time deformation, 𝑒=𝑔(𝑑), on the original time space to obtain a stationary dual process. (2)Apply a traditional linear filter on the dual space to extract components. (3)Transform the filtered dual components back to the original time space.

In the following examples we illustrate the implementation of the G-filter using the steps outlined above.

Example 4.1 (Examples 1.1 and 2.3 revisited). In this example we revisit the TVF data set shown in Figure 1(a) that was discussed in Examples 1.1 and 2.3. We will G-filter the data (based on a G(πœ†) model fit to the data) following the steps outlined above.
(1)The first step was done in Example 2.3 yielding the dual data in Figure 2(b).(2)Based on the discussion in Example 2.3 and the Wigner-Ville plot in Figure 2(f) it seems that the low and high frequency components in the dual data could be separated using a cutoff frequency of about 𝑓𝑐=.1. Figures 3(a) and 3(b) show the results of a low pass and high pass Butterworth filter, respectively, applied to the dual data in Figure 2(b). It can be seen that these filters do a good job of separating the low and high frequency components in the dual data.(3)Figures 3(c) and 3(d) show the data in 3(a) and 3(b), respectively, plotted on the original time scale (using linear interpolation). Comparing Figures 3(c) and 3(d) with Figures 1(b) and 1(c), respectively, shows that the procedure did a good job of separating the two TVF frequencies that we were previously unable to separate in Example 1.1 using standard methods. The somewhat jagged behavior of the peak heights is due to the interpolation. Alternatives to linear interpolation are under current investigation.

Example 4.2 (filtering example). In this example we consider the G(πœ†) stationary model ξ€·1βˆ’1.62𝐡+0.995𝐡2ξ€Έξ€·1βˆ’1.96𝐡+0.99𝐡2𝑋(𝑑+30)3βˆ’1ξ‚Άξ‚΅3β„Ž=π‘Ž(𝑑+30)3βˆ’1ξ‚Ά3β„Ž.(4.1) [𝑋(𝑑) is referred to by Jiang et al. [3] as a G(4;3) processes.] Clearly, after the time transformation 𝑔(𝑑)=((𝑑+30)3βˆ’1)/(3β„Ž)=𝑒, the dual model for π‘Œ(𝑒)=𝑋(𝑔(𝑑)) is the stationary AR(4) model ξ€·1βˆ’1.62𝐡+0.995𝐡2ξ€Έξ€·1βˆ’1.96𝐡+0.99𝐡2ξ€Έπ‘Œπ‘‘=π‘Žπ‘‘,(4.2) which has a characteristic equation (1βˆ’1.62𝑧+.995𝑧2)(1βˆ’1.96𝑧+.99𝑧2)=0 that has two pairs of complex conjugate roots quite close to the unit circle. Associated with each component is a system frequency. Realizations from this model will be characterized by frequency behavior in the neighborhoods of the two system frequencies, and consequently spectral estimates will have peaks close to these two frequency values. (See [14, 17]). Figure 4(a) shows a realization of length 𝑛=200 from the model in (4.1), and it is seen that there is a dramatic increase in frequency with time (i.e., period lengths decrease). There is also an indication of two TVFs. Using the GWS code, a time transformation with πœ†=2.9 and offset 14 was selected. Based on this time transformation, the dual data are obtained as in Figure 4(b) where the data appear stationary, again with two underlying frequencies. The extreme increase in frequencies in the original data set is illustrated in the Wigner-Ville plot in Figure 4(c). The lower strip going from near zero frequency for small values of 𝑑 to about 𝑓=.1 at𝑑=200. The upper (less visible) strip is associated with a higher-frequency TVF component that also begins at near zero frequency and increases to about 𝑓=.25 toward 𝑑=200. The Wigner-Ville plot for the dual data indicates that the two frequencies have been stabilized and further support the stationary nature of the dual. The dual model is dominated by two frequencies, a lower one at about 𝑓=.03 and an upper (again less visible) frequency of about 𝑓=.1. The results of applying 3rd order low-pass and high-pass Butterworth filters with a cutoff frequency𝑓𝑐=.065 are shown in Figures 4(e) and 4(f), respectively. Figures 4(g) and 4(h) show the filtered data sets plotted on the original time axis.
The behavior of the two filtered components is consistent with that in the original data as is shown in Figures 5(a) and 5(b). Also, the TVF components are close to those in the original data as seen by comparing Figures 5(c) and 5(d) with Figure 4(c).

Example 4.3 (bat echolocation data). In this example, we consider echolocation data from a large brown bat. The data were obtained courtesy of Al Feng of the Beckman Institute at the University of Illinois. The data shown in Figure 6(a) consist of 381 data points taken at 7-microsecond intervals with a total duration of .00266 seconds. This signal is quite complex and appears to be made up of possibly two different signals. The Wigner-Ville plot in Figure 6(c) suggests that the data contain two TVF components with a suggestion of possibly another TVF associated with higher frequencies than the two main components. Gray et al. [13] analyzed this data set as an M-stationary model with offset 203. Their analysis suggests that there are three dominant TVF components and that the highest component is sufficiently high frequency at the beginning of the time series that it is too rapid for the sampling rate to detect until about 𝑑=100. In our analysis here we will use the time transformation suggested by Gray, et al. [13] to compute the dual data. Using this time transformation produces the dual data in Figure 6(b). The Wigner-Ville plot in Figure 6(c) shows overlap between the TVF strips, for example, the frequency associated with the lower TVF strip near 𝑑=0 is similar to that of the upper strip at around 𝑑=250. The Wigner-Ville plot of the dual data is given in Figure 6(d) where it can be seen that the two dominant frequencies have been stabilized, and that the two dominant dual frequencies are well separated and fall at about 𝑓=.18 and 𝑓=.3. Figures 6(e) and 6(f) show the low-pass and high-pass filter results, respectively, plotted on the original time axis.

5. Concluding Remarks

In this paper, we show that the classical linear filtering methods cannot be directly applied to TVF data where the TVF components overlap over time. We discussed a family of models, G-stationary models, which have proven to be a useful extension of the usual stationary models, and which can be used to model a certain range of nonstationary time series with TVF. Then we introduce an easy and straightforward method for filtering the data from G-stationary models. This G-filtering method extends the standard linear filter, and provides techniques for filtering data with time varying frequencies that overlap in time. Simulation examples and a real data example are given to show that the effectiveness of G-filter in filtering data which are from, or can be approximated by, G-stationary models.