Most economic time series exhibit behaviour that is repeated over time. This allows for these processes to be modelled with the aid of techniques that consider the evolution of the process over a period of time. The application of this methodology would usually refer to the use of techniques that were developed in the time domain and include the general specification of ARIMA and state-space models. Another important phenomena of time series variables is that they may be decomposed into different periodic variations.

For example, we may wish to extract the cyclical component of the time series, which may be regarded as the part that exhibits higher periodic variation than the trend. When we are looking to identify the cyclical component of economic output, we are able to conduct an investigation into the behaviour of the business cycle, which is the topic of a number of empirical macroeconomic research investigations. These decompositions may be used to describe the stylized facts of the business cycle literature, such as the persistence of economic fluctuations and correlations (or lack thereof) between the cyclical components of economic aggregates in different countries.1

It is important to note that when seeking to derive a measure of the business cycle, there are no unique periodicities that may be used to identify this particular part of the process. This is partly due to the fact that it may have a relatively vague definition (from a purely statistical perspective) and may be considerably different when comparing the exact definition of the business cycle in a selection of countries. In addition, most economic time series are both fluctuating and growing, which makes these types of decompositions quite challenging, particularly when trying to perform this decomposition on the data of an emerging market economy.2 In addition, these procedures may also be used to identify the output gap, which is the difference between potential and actual output.

When seeking to decompose a time series into different periodic variations, we could imagine that the process is responding to various driving frequencies that are produced from linear combinations of sine and cosine functions, where each of these sine and cosine functions would represent a different frequency or amplitude. Expressed in these terms, the application of frequency domain techniques may be regarded as a regression of periodic sine and cosines on the respective values of the time series.3

Therefore, this chapter considers the application of a few widely used methods that are used to decompose economic time series. It includes a comparison of several detrending methods that are used to extract the business cycle, including deterministic detrending, stochastic detrending and frequency filtering techniques. The appendix includes a discussion of techniques that have been developed more recently, which exist in the time-frequency domain.

# 1 Spectral analysis

Consider an example where we are given three quarterly time series variables, $$y_t$$, $$x_t$$ and $$\upsilon_t$$, and the term $$\upsilon_t \sim \mathsf{i.i.d.} \mathcal{N} (0,\sigma)$$. If it is the case that $$x_t = y_t + \upsilon_t$$, then after regressing $$y_t$$ on $$x_t$$, we would expect to find that the coefficient would be large and significant, provided that $$\sigma$$ is not too large. The rationale for this is that $$x_t$$ contains information about $$y_t$$, which is reflected by the coefficient value.

As noted in the introduction, an observed time series could be viewed as the weighted sum of a number of underlying series that have different periodic behaviour. Hence, the cumulative variation in an observed time series will then be the sum of the contributions of these underlying series, which may vary in different frequencies. Spectral analysis is a tool that can be used to decompose a time series into different frequency components, where we may be interested in the respective contribution that has been made by various periodic components. When compared to the initial example that we used, a frequency domain analysis would involve a regression of a time series variable, $$x_t$$, on a number of periodic components that have different frequencies to investigate whether or not information relating to each of the frequency components is contained in $$x_t$$.

The general notion of periodicity can be made more precise by introducing some terminology. In order to define the rate at which a series oscillates, we first define a cycle as one complete period of a sine or cosine function defined over a unit time interval. We can then consider the periodic process

$\begin{eqnarray} y_t =A \cos (2 \pi \omega t+\phi) \tag{1.1} \end{eqnarray}$

for $$t= 0,\pm1,\pm2, \ldots$$, where $$\omega$$ is a frequency index, defined in cycles per unit time, while $$A$$ determines the height or amplitude of the function. The starting point of the cosine function is termed the phase and is denoted $$\phi$$. We could then introduce random variation in this time series by allowing the amplitude and phase to vary randomly. When seeking to conduct some form of data analysis, it is usually easier to use a trigonometric identity of this expression which may be written as,4

$\begin{eqnarray} y_t =U_1 \cos (2 \pi \omega t) +U_2 \sin (2 \pi \omega t) \tag{1.2} \end{eqnarray}$

where $$U_1 =A \cos \phi$$ and $$U_2 =-A\sin \phi$$ are often taken to be normally distributed random variables. In this case, the amplitude is $$A=\sqrt{U^2_1+U^2_2}$$ and the phase is $$\phi= \tan^{-1}(-U_2/U_1)$$. From these facts we can show that if, and only if, in (1.1), $$A$$ and $$\phi$$ are independent random variables, where $$A^2$$ is chi-squared with 2 degrees of freedom, and $$\phi$$ is uniformly distributed on $$[-\pi, \pi]$$, then $$U_1$$ and $$U_2$$ are independent, standard normal random variables.

The above random process is also a function of its frequency, defined by the parameter $$\omega$$. The frequency is measured in cycles per unit time, or in cycles per point in the above illustration. For $$\omega= 1$$, the series makes one cycle per time unit; for $$\omega=.50$$, the series makes a cycle every two time units; for $$\omega=.25$$, every four units, and so on. In general, for data that occur at discrete time points will need at least two points to determine a cycle, so the highest frequency of interest is $$\frac{1}{2}$$ cycles per point. This frequency is called the folding frequency and defines the highest frequency that can be seen in discrete sampling. Higher frequencies sampled this way will appear at lower frequencies, called aliases; an example is the way a camera samples a rotating wheel on a moving vehicle in a movie, in which the wheel appears to be rotating at a different rate.

Consider a generalization of (1.2) that allows mixtures of periodic series with multiple frequencies and amplitudes,

$\begin{eqnarray} y_t = \sum_{k=1}^{q} \big[U_{k1}\cos(2 \pi \omega_k t) + U_{k2}\sin(2 \pi \omega_k t)\big], \tag{1.3} \end{eqnarray}$

where $$U_{k1}, U_{k2}$$, for $$k= 1,2, \ldots , q$$, are independent zero-mean random variables with variances $$\sigma^2_k$$, and the $$\omega_k$$ are distinct frequencies. Notice that (1.3) exhibits the process as a sum of independent components, with variance $$\sigma^2_k$$ for frequency $$\omega_k$$. Using the independence of the $$U_s$$ and the trigonometry identities, it is easy to show that the autocovariance function of the process is

$\begin{eqnarray} \gamma(h) = \sum_{k=1}^{q} \sigma^2_k \cos(2 \pi \omega_k h), \tag{1.4} \end{eqnarray}$

and we note the autocovariance function is the sum of periodic components with weights proportional to the variances $$\sigma^2_k$$. Hence, $$y_t$$ is a mean-zero stationary processes with variance which exhibits the overall variance as a sum of variances of each of the component parts.

To see how the spectral techniques can be used to interpret the regular frequencies in the series, consider the following four periodic time series

$\begin{eqnarray*} x_{1,t} &=&2\cos (2\pi t6/100)+3\sin (2\pi t6/100) \\ x_{2,t} &=&4\cos (2\pi t30/100)+5\sin (2\pi t30/100) \\ x_{3,t} &=&6\cos (2\pi t40/100)+7\sin (2\pi t40/100) \\ y_{t} &=&x_{1,t}+x_{2,t}+x_{3,t} \end{eqnarray*}$

Figure 1: Different frequency components

Where the first three series $$\{x_{1,t},x_{2,t}, x_{3,t}\}$$, are generated with different periodicities and amplitude, and the fourth series, $$y_{t}$$, is the sum of the three others. The series are plotted in Figure 1, where we display the actual series over time. These graphs clearly see how $$x_1$$ displays the longer waves (as it has the low frequency component), followed by $$x_2$$ and $$x_3$$. The $$y$$ series is the sum of the other three series, and will therefore inherent the properties of the $$\{x_{1,t},x_{2,t}, x_{3,t}\}$$ series. Note that it is not easy to visualize the periodicities from this time series plot.

The systematic sorting out of the essential frequency components in a time series, including their relative contributions, constitutes one of the main objectives of spectral analysis. One way to accomplish this objective is to regress sinusoids that vary at the different fundamental frequencies on the data. This is represented by the periodogram (or sample spectral density) and may be expressed as,

$\begin{eqnarray*} P(j/n) = \frac{2}{n} \sum^{n}_{t=1} y_t \cos \big(2 \pi t \; j/n \big)^2 + \frac{2}{n} \sum^{n}_{t=1} y_t \sin \big(2 \pi t \; j/n \big)^2 \end{eqnarray*}$

and may be regarded as a measure of the squared correlation of the data with sinusoids oscillating at a frequency of $$\omega_j =j/n$$, or $$j$$ cycles in $$n$$ time points. The periodogram may be computed quickly using the fast Fourier transform (FFT), and there is no need to run repeated regressions. An example of a periodogram is provided in Figure 2, where $$x_1$$ has a peak in the periodogram to the left, followed by $$x_2$$ and $$x_3$$, which has a peak that is to the right of the other two. Hence, $$x_3$$ is the most high frequent of the three series. Note that in this case, it is easy to visualize the three periodic components that are in $$y_t$$, as each corresponds to the original time series component. Hence, we confirm that the $$y$$-series has inherited the periodicity of the three other $$x$$-series.

Figure 2: Periodogram for frequency components

Note that the horizontal scale of the periodogram represents $$P(j/n)$$ for $$j = 0,1, \ldots , n-1$$. These values are related to the size of $$\omega$$, since

$\begin{eqnarray} \nonumber P(6/100) = 0.06, P(30/100) = 0.3, P(40/100) = 0.4, \end{eqnarray}$

Hence, where the specific frequency is present in a time series, the value for $$P(j/n) \ne 0$$ in all other cases, $$P(j/n) = 0$$. Note, that in this example, these values for the respective frequencies of the components are displayed in Figure 1, which suggests that the periodogram may provide some insight into the variance components of any data. In addition, the vertical scale of the periodogram is a function of the amplitude, $$A$$, and may be used to show the relative strength of cosine-sine pairs at various frequencies in the overall behaviour of the time series.

An interesting exercise would be to construct the $$x_1$$ series from $$y_t$$, which may be regarded as actual data. To do so we need to filter out all components that lie outside the chosen frequency band of $$x_1$$. Such a filter could operate with the aid of a regression model that contains that information that relates to a particular frequency (although there are more convenient ways of going about this). Hence, cycles with frequencies corresponding to $$x_2$$ and $$x_3$$ would be excluded, while cycles with frequency corresponding to $$x_1$$ will be maintained (i.e. can pass through the filter).

Figure 3: Filtered result for frequency components

Figure 3 displays the original series y and the new filtered series. From the periodogram we see that the filtered series is almost equal to the original $$x_1$$ series. Hence, the filter removes the high frequent components (corresponding to $$x_2$$ and $$x_3$$), and we are left with the periodic component that corresponds to the original series $$x_1$$.

While this example is particularly intuitive, it is worth bearing in mind that the simulated series were constructed from known cycles with a particular frequency. However, most of the data that we will encounter will consist of many frequencies, making interpretation more difficult. Still one or two frequencies usually dominate most economic time series, and this is typically what we are looking to identify.

# 2 Statistical detrending methods

The various detrending methods that are used on economic data provide different estimates of the cycle, where the most appropriate transformation should be determined by the underlying dynamic properties of the variable that is being decomposed.

If we assume that an economic time series can be decomposed into a trend $$g_{t}$$ and a cycle $$c_{t}$$, then we could use the following expression for a time series,

$\begin{eqnarray} y_{t}=g_{t}+c_{t} \tag{2.1} \end{eqnarray}$ where we abstract from the (irregular) noise and seasonal component. The objective of decomposing a time series would be to obtain estimates of $$g_{t}$$ and $$c_{t}$$, from the respective detrending methods.

It is worth noting that different methods that may be used to decompose a time series would usually produce different estimates of the cycle.5 The ultimate choice of the appropriate transformation of the data should therefore depend on the nature of the underlying dynamic properties of the time series. For example, one may wish to make use of a unit root tests to determine whether the trend in the data is either stochastic or a deterministic. However, such initial testing might not make it obvious which detrending or transformation method should be used, and as such, one way also wish to consider the generally accepted practices that are currently applied in the literature.

In the subsequent sections of this chapter we consider the use of various techniques that may be used to obtain estimates of $$g_{t}$$ and $$c_{t}$$, which may be used to extract a deterministic or stochastic trend. Many of these detrending methods are related, where by way of example the Hodrick-Prescott filter with a high smoothing value could be used to identify a linear trend. Similarly, the use of the Hodrick-Prescott filter with a low smoothing value, would provide an estimate of the trend that is largely equivalent to the result that would have been obtained after performing a stochastic Beveridge-Nelson decomposition.

# 5 Beveridge-Nelson Decomposition

Many economic time series are integrated of the first-order, where the first difference of a process can be represented as a stationary process that has an autoregressive moving-average form. To decompose such a nonstationary series into a permanent and a transitory component one could make use of the decomposition, that is due to Beveridge and Nelson (1981). In this case, the permanent component is assumed to follow a random walk with drift, and the transitory component is a stationary process with zero mean (that is perfectly correlated with the permanent component). If we let $$y_{t}$$ be integrated of first-order, then the first difference, $$\Delta y_{t}$$, is stationary. Where we assume that the stationary component has the following moving-average representation,

$\begin{eqnarray} (1-L)y_{t}=\Delta y_{t}=\mu +B(L)\varepsilon_{t} \tag{5.1} \end{eqnarray}$

To derive estimates for this decomposition, we would firstly define the polynomial,

$\begin{eqnarray*} B^{\ast }(L)=(1-L)^{-1}[B(L)-B(1)] \end{eqnarray*}$

where $$B(1)=\overset{\infty }{\underset{s=0}{\sum }}B_{s}$$. Rewriting this polynomial in terms of $$B(L)$$, provides

$\begin{eqnarray*} B(L)=[B(1)+(1-L)B^{\ast }(L)] \end{eqnarray*}$

and substituting into (5.1) yields the following expression,

$\begin{eqnarray} \Delta y_{t}=\mu +B(L)\varepsilon_{t}=\mu +[B(1)+(1-L)B^{\ast}(L)]\varepsilon_{t} \tag{5.2} \end{eqnarray}$

For the decomposition, $$y_{t}=g_{t}+c_{t}$$, it follows that $$\Delta y_{t}=\Delta g_{t}$$ $$+\Delta c_{t}$$. Using this together with equation (5.2), provides an expression for the change in the trend component of $$y_{t}$$,

$\begin{eqnarray} \Delta g_{t}=\mu +B(1)\varepsilon_{t} \tag{5.3} \end{eqnarray}$

and the change in the cyclical component of $$y_{t}$$ is then provided by

$\begin{eqnarray*} \Delta c_{t}=(1-L)B^{\ast }(L)\varepsilon_{t} \end{eqnarray*}$

From equation (5.3) we see that the trend follows a random walk with drift. This expression can be solved to yield,

$\begin{eqnarray} g_{t}=g_{0}+\mu t+B(1)\overset{t}{\underset{s=1}{\sum }}\varepsilon_{s} \tag{5.4} \end{eqnarray}$

As such, the trend consist of both a deterministic term

$\begin{eqnarray*} g_{0}+\mu t \end{eqnarray*}$

and a stochastic term

$\begin{eqnarray*} B(1)\overset{t}{\underset{s=1}{\sum }}\varepsilon_{s} \end{eqnarray*}$

For $$B(1)=0$$, the trend reduces to a deterministic case, where for $$B(1)\neq 0$$, the stochastic part indicates the long-run impact of a shock $$\varepsilon_{t}$$ on the level of $$y_{t}$$. The cyclical component is stationary and is given by

$\begin{eqnarray} c_{t}=B^{\ast }(L)\varepsilon_{t}=(1-L)^{-1}[B(L)-B(1)]\varepsilon_{t} \tag{5.5} \end{eqnarray}$

Beveridge and Nelson (1981) are able to show that the stochastic trend defined in equation (5.4) could also be interpreted as the long-term forecast of the series adjusted for the mean rate of change. In addition, the cycle that is defined in equation (5.5) as the stationary process that reflects the deviations of the trend from the observed series.

This decomposition implies that an innovation in $$g_{t}$$ and $$c_{t}$$ are proportional to $$\varepsilon_{t}$$. Hence they will be perfectly correlated and the permanent component will have the same drift, $$\mu$$, as the observed series. Further, the variance of the innovations in the permanent component, $$(B(L)\sigma^{2})$$ will be larger (smaller) than the variance of the innovations in the observed data $$y_{t}$$, if $$B(1)$$ is larger (smaller) than one. Note also that when the permanent component is restricted to be a random walk with drift, $$B_{0}=1$$ and all the $$B_{i}=0$$ for $$i>0$$, the variance of the permanent component equals the variance in the observed series, and the cyclical component will be zero for all $$t$$.

To be able to identify the cyclical and permanent component, one must specify models that can be written as the stationary moving average processes. There are two stages involved in this trend-cycle decomposition. First an ARIMA model $$(p,d,q)$$ has to be estimated to the series $$y_{t}$$ where $$p$$ is the number of AR lags, $$d$$ is the number of differencing and $$q$$ is the number of MA lags. Then $$c_{t}$$ has to be numerically estimated.

To see how one can estimate the Beveridge-Nelson (BN) decomposition in practice, assume that an AR(1) process is responsible for the growth rate of output, $$\Delta y_{t}=\phi \Delta y_{t-1}$$ $$+$$ $$\varepsilon_{t}$$, where we have ignored the constant term. Assuming $$\phi <1$$, the AR(1) process can be written in terms of the infinite MA($$\infty$$) process where we find $$B(L)$$, $$B(1)$$ and $$B^{\ast }(L)$$ from,

$\begin{eqnarray*} B(L) &=&\frac{1}{1-\phi L} \\ B(1) &=&\frac{1}{1-\phi } \\ B^{\ast }(L) &=&(1-L)^{-1}[B(L)-B(1)]=\frac{\phi }{(1-\phi )(1-\phi L)} \end{eqnarray*}$

Solving in terms of $$y_{t}$$, using equation (5.2), but now without a constant, provides us with

$\begin{eqnarray*} y_{t}=(1-L)^{-1}[B(1)+(1-L)B^{\ast }(L)]\varepsilon_{t} \end{eqnarray*}$

This can be rewritten as,

$\begin{eqnarray*} y_{t}=B(1)(1-L)^{-1}\varepsilon_{t}+(1-L)^{-1}[B(L)-B(1)]\varepsilon_{t} \end{eqnarray*}$

Substituting in for the AR(1) solution derived above, we have

$\begin{eqnarray*} y_{t} &=&g_{t}+c_{t} \\ &\Downarrow & \\ y_{t} &=&\frac{1}{1-\phi }(1-L)^{-1}\varepsilon_{t}+\frac{-\phi }{(1-\phi L)(1-\phi )}\varepsilon_{t} \end{eqnarray*}$

A quick computational approach was suggested by Cuddington and Winters (1987), where $$g_{t}$$ is calculated directly from the expression in (5.4) by estimating $$B(1)$$ from a truncated Wold representation of $$\Delta y_{t}$$. The obvious difficulty is that the initial value of $$g_{t}$$ in (5.4) is known, so the procedure is only correct up to an additive factor.

The advantage of Beveridge-Nelson is that it is appropriate for the extraction of cycles when a series is difference-stationary. Moreover, it may be used one those series that not only contain a unit root, but are also highly volatile.

One disadvantage, however, is that it is rather time-consuming, as one has to choose between different ARMA models. In addition, Cochrane (1988) notes that these different ARMA specifications may give very different trend-cycle decompositions, where low-order ARMA models will systematically overestimate the random walk component in the trend. hence, the parameter estimates will match the short-run behaviour and misrepresent the long-run behaviour. As the innovative variance of the random walk is a property of the very long-run behaviour of the series, one should therefore estimate higher-order models that adequately capture this long-run behaviour. Finally, misrepresenting an $$I(2)$$ process as an $$I(1)$$ may be problematic in this setting.

# 6 Stylized facts of business cycles

The description of the aggregate fluctuations in a set of economic variables is an important exercise as it establishes the stylized facts of the business cycles in a particular country. However, as the cycles will not be invariant to the method that is used to describe them, the results should be tested against alternative trend specifications.

Figure 8 compares the cycles that are derived from gross domestic product in South Africa. These were obtained from the different de-trending methods that were presented above. The results suggest that a linear filter provides peculiar results that do not agree with what we know of economic events. The stochastic filters suggest a fairly similar pattern for the business cycles although there are parts where the differences are more pronounced. The Beveridge-Nelson decomposition would appear to provide slightly divergent results and as such it would be important to compare how these results compare to the official dating procedure for recessions, which is provided by the South African Reserve Bank.

Figure 8: Business Cycles - SA output (1960Q1-2015Q1)

In addition to this comparison, it is usually a good idea to consider the stylized facts that relate to the cycles of other macroeconomic aggregates, such as: GDP, consumption (Cons), exports (Exp), import (Imp), productivity (Prod), investment (Invest) and employment (EMP). These should be compared to our knowledge of economic events and where relevant, we should also consider the correlation between these measures, as well as whether or not they are leading or lagging one another. This will help us to judge the plausibility of using a particular technique to decompose the business cycle.

# 7 Conclusion

Many economic and financial applications make use of decompositions for nonstationary time series, which are transformed into a permanent and a transitory component. To complete this task one could make use of a linear filter for the trend, which may be perturbed by transitory cyclical fluctuations. alternatively one could make use of a Hodrick-Prescott filter, which is the most popular method for extracting extracting business cycles. The filter extracts a stochastic trend which for a given value of the parameter $$\lambda$$. This trend moves smoothly over time and in uncorrelated with the cycle. It is worth noting that the results from this filter are not robust to changes in the value of the smoothness parameter. Another popular method that is used to measure the business cycle is that of the band pass filter. The filter removes all the components in a series, except for those that correspond to the chosen frequency band. In the Beveridge-Nelson decomposition, the permanent component is shown to be a random walk with drift, and the transitory component is stationary process with zero mean, which is perfectly correlated with the permanent component. When applying these methods, one should interrogate the results for robustness by applying many alternative filters when analysing business cycles.

Then lastly, it is perhaps worth noting that a filtered time series is usually stationary but somewhat persistent; which is the case for the HP filtered measure of the output-gap. Given these properties of the data, one would usually be able to generate reasonable forecasts for output-gap (particularly when compared to the forecast for output growth, which is less persistent). However, one may wish to suggest that a forecast for the output-gap is less useful than a forecast for output growth and the end-of-sample problem that is encountered with the HP filter would detract further from the usefulness of the forecast for the output-gap. However, this would in no way prevent one from using the measure of the output gap in a multivariate model that is concerned with forecasting some other variable (such as inflation).

# 8 Appendix: Wavelet decompositions

Most of the commonly used decompositions in economics, such as those that were designed by Hodrick and Prescott (1997), Baxter and King (1999) and Christiano and Fitzgerald (2003), seek to approximate ideal filters, where one is able to identify the trend, cycle and noise components that are located at different periodicities. When applying these techniques within the frequency domain, one would effectively decompose a time series with a number of sine and cosine functions, which define the rate at which the time series oscillates. It is important to note, that this transformation results in the loss of all time-based information, where it is assumed that the periodicity of all the components are consistent throughout the entire sample.

To allow for changes in the periodicity of the respective components, Gabor (1946) developed the Short-Time Fourier Transform (STFT) technique, which involves applying a number of Fourier transforms to different subsamples of the data. Gencay, Selcuk, and Whitcher (2010) refer to the subsample as a data window, where the technique involves sliding the window across the time series and taking a Fourier transform of each subsample. Although this technique would provide potentially useful information on the timing of an event that may have arisen at a particular frequency, it is limited in that the precision of the analysis is affected by the size of the subsample. For instance, one would need a large subsample to identify changes that arise at a low frequency, and small subsamples to identify changes in the higher frequency components.

To overcome the limitations of the above frequency domain techniques, wavelet transformations were developed to capture features of time-series data across a wide range of frequencies that may arise at different points in time. This technique makes use of a number of wavelet functions that are stretched and shifted to describe features that are localised in frequency and time. For example, the wavelet function would be expanded over a relatively long period of time when identifying low-frequency events, and it would be relatively narrow when describing high frequency events.8 After shifting all of these wavelet functions that have different amplitudes over the entire sample of data, one is able to associate the components with specific time horizons that occur at different locations in time.

Early work with wavelet functions dates back to Haar (1910), who used a number of square-wave functions to decompose time-series data. Unfortunately, the properties of square-wave functions were found to be limited, and as such, a number of alternatives were developed, including those that are discussed in Grossmann and Morlet (1984) and Daubechies (1992).9 For the computation of these transformations, which make use of various wavelet functions at different scales, most studies currently employ the multiresolution decomposition of Mallat (1989) and Strang and Nguyen (1996).

To describe the use of this technique, one could allow for the case where a variable is composed of a trend and a number of higher-frequency components. In this instance, the trend may be represented by a father wavelet, $$\phi(t)$$, while the mother wavelets, $$\psi(t)$$, are used to describe information at lower scales (i.e. higher frequencies). Using an orthogonal wavelet transformation, one could then describe variable $$y_t$$ as

$\begin{eqnarray} y_t = \sum_k s_{0,k} \phi_{0,k} (t) + \sum_{j=0}^{J} \sum_k d_{j,k} \psi_{j,k} (t) \; , \end{eqnarray}$

where $$J$$ refers to the number of scales, and $$k$$ refers to the location of the wavelet in time. The $s_{0,k}$ coefficients are termed smooth coefficients, since they represent the trend, and the $$d_{j,k}$$ coefficients are termed the detailed coefficients, since they represent finer details in the data.

The mother wavelet functions, $$\psi_{1,k} (t), \dots, \psi_{J,k} (t)$$, are then generated by shifts in the location of the wavelet in time and scale, such that

$\begin{eqnarray} \psi_{j,k} (t) = 2^{-j/2} \psi \left(\frac{t-2^{j}k}{2^j}\right), \;\; j=1,\dots,J \; , \end{eqnarray}$

where the shift parameter is represented by $$2^{j}k$$ and the scale parameter is $$2^{j}$$. This choice of dyadic scaling factors is arbitrary but efficient (Daubechies 1992). As depicted in the daublet wavelet functions in Figure 9 , smaller values of $$j$$ (which produce a smaller scale parameter $$2^{j}$$), would provide the relatively tall and narrow wavelet function on the left. For larger values of $$j$$, the wavelet function is more spread out and of lower amplitude. In addition, after shifting this function by one period, we produce the function that is depicted on the right of Figure 9 .

Figure 9: Daublet (4) wavelet functions - $$\psi_{1,0}(t)$$ and $$\psi_{2,1}(t)$$

Early applications of wavelet methods in economics include the work of Ramsey and Zhang (1997), which made use of a wavelet decomposition of exchange rate data to describe the distribution of this data at different frequencies. In addition, Ramsey and Lampart (1998a) made use of a decomposition of money and income data to describe the relationship between these variables at different frequencies, while Ramsey and Lampart (1998b) considered the relationship between income and expenditure (i.e. permanent income hypothesis) at different time scales.10

Modern wavelet functions may take various forms that could be summarized by smoothed functions, that may be used to decompose a series into trend and cycle, peaked functions, that may be used to identify the peak of cycle, or square functions, that are used to identify structural breaks. For the purposes of identifying the business cycle one would use smoothed functions that include daublets, coiflets and symlets. There are also a number of transformations that may be used. Many studies make use of a maximum overlap discrete wavelet transform (MODWT), which does not restrict the sample size to a multiple of $$2^j$$. In addition, this technique is also able to preserve the phase properties of the data, where it can match the smoothed terms to the underlying data.

Figure 10: Daublet (4) wavelet decomposition - South African inflation

Figure 11: Daublet (4) wavelet decomposition - South African inflation

Figure 10 and 11 contain the results of a decomposition that was applied to South African Consumer Price inflation, where we are interested in removing the noise from the data. In this example, we make use of various smoothed wavelet functions that include daublets 3-4. We also make use of both three scales, where $$J$$ is set at $$3$$.11

The decompositions at various different scales are contained in Figure 10, where we note that there is significant change in the periodicity of the variables over time. The results of the filtered trend are contained in Figure 11, which could be used as a measure of core inflation, as in Du$$\;$$Plessis, Du$$\;$$Rand, and Kotzé (2015).

To conclude this section, there are a number of advantages that are inherent in the application of wavelet decompositions, as they can be applied to data of any integration order and allow for changes in the distribution of the frequency over time. In addition, they have many of the benefits that are associated with spectral techniques, but they do not lose the time support, which is useful when seeking to identify changes in the process at different frequencies. Then lastly, as one is able to include a number of bands, which are additive, one is able to focus attention on many possible periodic components.

# 9 References

Baxter, Marianne, and Robert G. King. 1993. “Fiscal Policy in General Equilibrium.” American Economic Review 83 (3): 315–34.

———. 1999. “Measuring Business Cycles: Approximate Band-Pass Filters for Economic Time Series.” Review of Economics and Statistics 81: 575–93.

Beveridge, S., and C. R. Nelson. 1981. “A New Approach to Decomposition of Economic Time Series into Permanent and Transitory Components with Particular Attention to Measurement of the Business Cycle.” Journal of Monetary Economics 7 (2): 151–74.

Burns, Arthur F., and Wesley C. Mitchell. 1947. Measuring Business Cycles. New York: National Bureau of Economic Research.

Canova, Fabio. 1998. “Detrending and Business Cycle Facts: A Users Guide.” Journal of Monetary Economics 41 (3): 475–512.

Christiano, L., and T. Fitzgerald. 2003. “The Band Pass Filter.” International Economic Review 44 (435-465).

Cochrane, John H. 1988. “How Big Is the Random Walk in GNP?” The Journal of Political Economy, 893–920.

Crowley, Patrick M. 2007. “A Guide to Wavelets for Economists.” Journal of Economic Surveys 21 (2): 207–67.

Cuddington, J. T., and L. A. Winters. 1987. “The Beveridge-Nelson Decomposition of Economic Time Series: A Quick Computational Method.” Journal of Monetary Economics 19 (1): 125–27.

Daubechies, Ingrid. 1992. Ten Lectures on Wavelets. Vol. 61. SIAM.

Du$$\;$$Plessis, Stan, Gideon Du$$\;$$Rand, and Kevin Kotzé. 2015. “Measuring Core Inflation in South Africa.” South African Journal of Economics 83 (4): 527–48.

Gabor, Dennis. 1946. “Theory of Communication.” Journal of the I.E.E 93(3): 429–57.

Gencay, Ramazan, Faruk Selcuk, and Brandon Whitcher. 2010. An Introdution to Wavelet and Other Filtering Methods in Finance and Economics. San Diego: Academic Press.

Grossmann, Alexander, and Jean Morlet. 1984. “Decomposition of Hardy Functions into Square Integrable Wavelets of Constant Shape.” SIAM Journal on Mathematical Analysis 15 (4). SIAM: 723–36.

Haar, Alfred. 1910. “Zur Theorie Der Orthogonalen Funktionensysteme.” Mathematische Annalen 69 (3): 331–71.

Harvey, Andrew C., and A. Jaeger. 1993. “Detrending, Stylised Facts and the Business Cycle.” Journal of Applied Econometrics 8: 231–47.

Heil, C., and D. F. Walnut. 2006. Fundamental Papers in Wavelet Theory. Princeton, New Jersey: Princeton University Press.

Hodrick, Robert J., and Edward C. Prescott. 1980. “Postwar U.S. Business Cycles: An Empirical Investigation.” Carnegie Mellon University Discussion Paper 451.

———. 1997. “Postwar U.S. Business Cycles: An Empirical Investigation.” Journal of Money, Credit and Banking 29 (1): 1–16.

Hubbard, Barbara B. 1998. The World According to Wavelets: The Story of a Mathematical Technique in the Making. Second. A K Peters/CRC Press.

King, Robert G., and Sergio T. Rebelo. 1993. “Low Frequency Filtering and Real Business Cycles.” Journal of Economic Dynamics and Control 17 (1-2): 207–31.

Mallat, Stephane G. 1989. “A Theory for Multiresolution Signal Decomposition: The Wavelet Representation.” Pattern Analysis and Machine Intelligence, IEEE Transactions on 11 (7). Ieee: 674–93.

Nelson, C. R., and H. King. 1981. “Spurious Periodicity in Anappropriately Detrended Time Series.” Econometrica, 741–51.

Ramsey, James B. 2002. “Wavelets in Economics and Finance: Past and Future.” Studies in Nonlinear Dynamics & Econometrics 6 (3).

Ramsey, James B., and Camille Lampart. 1998a. “Decomposition of Economic Relationships by Timescale Using Wavelets.” Macroeconomic Dynamics 2 (1). Cambridge Univ Press: 49–71.

———. 1998b. “The Decomposition of Economic Relationships by Time Scale Using Wavelets: Expenditure and Income.” Studies in Nonlinear Dynamics & Econometrics 3 (1): 23–42.

Ramsey, James B., and Zhifeng Zhang. 1997. “The Analysis of Foreign Exchange Data Using Waveform Dictionaries.” Journal of Empirical Finance 4 (4). Elsevier: 341–72.

Schleicher, Christoph. 2002. “An Introduction to Wavelets for Economists.” Working Paper 2002-3. Bank of Canada.

Shumway, R. H., and D. S. Stoffer. 2011. Time Series Analysis and Its Applications: With R Examples. Edited by 3rd. New York: Springer.

Strang, Gilbert, and Truong Nguyen. 1996. Wavelets and Filter Banks. Wellesley Cambridge Press.

1. Burns and Mitchell (1947) conducted an early empirical study into the behaviour of business cycles in the United States, with the aid of a systematic approach where they observed that a cycle would consist of expansions occurring at about the same time in many economic activities, followed by a contraction in many of the variables. They suggested that this sequence of changes is recurrent but not periodic. Similar work is currently performed by the South African Reserve Bank, which seeks to establish the dates for economic expansions and contractions.

2. In terms of a formal definition, the business cycle refers to the regular periods of expansion and contraction in major economic aggregate variables (Burns and Mitchell 1947). We would infer that a turning point occurs when the business cycle reaches local maximum (peak) or local minimum (trough).

3. Shumway and Stoffer (2011) suggest that most investigations into the cyclical component of a time series should be expressed with the aid of frequency domain techniques, which employ Fourier transformations that are driven by sine and cosine functions.

4. The general rules of trigonometry ensure that $$\cos(0)=1$$, $$\sin(0)=0 \cos(-x)=\cos (x)$$ and $$\sin(-x)=\sin(-x)$$.

5. See Canova (1998) amongst others.