The areas where SSA can be applied are very broad: climatology, marine science, geophysics, engineering, image processing, medicine,
econometrics among them. Hence different modifications of SSA have been proposed and different methodologies of SSA are used in practical applications such as
trend extraction,
periodicity detection,
seasonal adjustment,
smoothing,
noise reduction (Golyandina, et al, 2001).
Basic SSA SSA can be used as a model-free technique so that it can be applied to arbitrary time series including non-stationary time series. The basic aim of SSA is to decompose the time series into the sum of interpretable components such as trend, periodic components and noise with no a-priori assumptions about the parametric form of these components. Consider a real-valued time series \mathbb{X}=(x_1,\ldots,x_{N}) of length N. Let L \ (1 be some integer called the
window length and K=N-L+1.
Main algorithm 1st step: Embedding. Form the
trajectory matrix of the series \mathbb{X}, which is the L\!\times\! K matrix : \mathbf{X}=[X_1:\ldots:X_K]=(x_{ij})_{i,j=1}^{L,K}= \begin{bmatrix} x_1&x_2&x_3&\ldots&x_{K}\\ x_2&x_3&x_4&\ldots&x_{K+1}\\ x_3&x_4&x_5&\ldots&x_{K+2}\\ \vdots&\vdots&\vdots&\ddots&\vdots\\ x_{L}&x_{L+1}&x_{L+2}&\ldots&x_{N}\\ \end{bmatrix} where X_i=(x_{i},\ldots,x_{i+L-1})^\mathrm{T} \; \quad (1\leq i\leq K) are
lagged vectors of size L. The matrix \mathbf{X} is a
Hankel matrix which means that \mathbf{X} has equal elements x_{ij} on the anti-diagonals i+j =\,{\rm const}.
2nd step: Singular Value Decomposition (SVD). Perform the singular value decomposition (SVD) of the trajectory matrix \mathbf{X}. Set \mathbf{S}=\mathbf{X} \mathbf{X}^\mathrm{T} and denote by \lambda_1, \ldots,\lambda_L the
eigenvalues of \mathbf{S} taken in the decreasing order of magnitude (\lambda_1\geq \ldots \geq \lambda_L\geq 0) and by U_1,\ldots,U_L the orthonormal system of the
eigenvectors of the matrix \mathbf{S} corresponding to these eigenvalues. Set d= \mathop{\mathrm{rank}} \mathbf{X} =\max\{i,\ \mbox{such that}\ \lambda_i >0\} (note that d=L for a typical real-life series) and V_i=\mathbf{X}^\mathrm{T} U_i/\sqrt{\lambda_i} (i=1,\ldots,d). In this notation, the SVD of the trajectory matrix \mathbf{X} can be written as : \mathbf{X} = \mathbf{X}_1 + \ldots + \mathbf{X}_d, where :\mathbf{X}_i=\sqrt{\lambda_i}U_i V_i^\mathrm{T} are matrices having rank 1; these are called
elementary matrices. The collection (\sqrt{\lambda_i},U_i,V_i) will be called the ith
eigentriple (abbreviated as ET) of the SVD. Vectors U_i are the left singular vectors of the matrix \mathbf{X}, numbers \sqrt{\lambda_i} are the singular values and provide the singular spectrum of \mathbf{X}; this gives the name to SSA. Vectors \sqrt{\lambda_i}V_i=\mathbf{X}^\mathrm{T} U_i are called vectors of principal components (PCs).
3rd step: Eigentriple grouping. Partition the set of indices \{1,\ldots,d\} into m disjoint subsets I_1,\ldots,I_m. Let I=\{i_1,\ldots,i_p\}. Then the resultant matrix \mathbf{X}_I corresponding to the group I is defined as \mathbf{X}_I=\mathbf{X}_{i_1}+\ldots+\mathbf{X}_{i_p}. The resultant matrices are computed for the groups I=I_1, \ldots, I_m and the grouped SVD expansion of \mathbf{X} can now be written as : \mathbf{X}=\mathbf{X}_{I_1}+\ldots+\mathbf{X}_{I_m}.
4th step: Diagonal averaging. Each matrix \mathbf{X}_{I_j} of the grouped decomposition is hankelized and then the obtained
Hankel matrix is transformed into a new series of length N using the one-to-one correspondence between Hankel matrices and time series. Diagonal averaging applied to a resultant matrix \mathbf{X}_{I_k} produces a
reconstructed series \widetilde{\mathbb{X}}^{(k)}=(\widetilde{x}^{(k)}_1,\ldots,\widetilde{x}^{(k)}_N). In this way, the initial series x_1,\ldots,x_N is decomposed into a sum of m reconstructed subseries: : x_n = \sum\limits_{k=1}^m \widetilde{x}^{(k)}_n \ \ (n=1,2, \ldots, N). This decomposition is the main result of the SSA algorithm. The decomposition is meaningful if each reconstructed subseries could be classified as a part of either trend or some periodic component or noise.
Theory of SSA separability The two main questions which the theory of SSA attempts to answer are: (a) what time series components can be separated by SSA, and (b) how to choose the window length L and make proper grouping for extraction of a desirable component. Many theoretical results can be found in Golyandina et al. (2001, Ch. 1 and 6). Trend (which is defined as a slowly varying component of the time series), periodic components and noise are asymptotically separable as N\rightarrow \infty. In practice N is fixed and one is interested in approximate separability between time series components. A number of indicators of approximate separability can be used, see Golyandina et al. (2001, Ch. 1). The window length L determines the resolution of the method: larger values of L provide more refined decomposition into elementary components and therefore better separability. The window length L determines the longest periodicity captured by SSA. Trends can be extracted by grouping of eigentriples with slowly varying eigenvectors. A sinusoid with frequency smaller than 0.5 produces two approximately equal eigenvalues and two sine-wave eigenvectors with the same frequencies and \pi/2-shifted phases. Separation of two time series components can be considered as extraction of one component in the presence of perturbation by the other component. SSA perturbation theory is developed in Nekrutkin (2010) and Hassani et al. (2011).
Forecasting by SSA If for some series \mathbb{X} the SVD step in Basic SSA gives d, then this series is called
time series of rank d (Golyandina et al., 2001, Ch.5). The subspace spanned by the d leading eigenvectors is called
signal subspace. This subspace is used for estimating the signal parameters in
signal processing, e.g.
ESPRIT for high-resolution frequency estimation. Also, this subspace determines the
linear homogeneous recurrence relation (LRR) governing the series, which can be used for forecasting. Continuation of the series by the LRR is similar to forward
linear prediction in signal processing. Let the series be governed by the minimal LRR x_{n}=\sum_{k=1}^d b_k x_{n-k}. Let us choose L>d, U_1,\ldots,U_d be the eigenvectors (left singular vectors of the L-trajectory matrix), which are provided by the SVD step of SSA. Then this series is governed by an LRR x_{n}=\sum_{k=1}^{L-1} a_k x_{n-k}, where (a_{L-1},\ldots,a_1)^\mathrm{T} are expressed through U_1,\ldots,U_d (Golyandina et al., 2001, Ch.5), and can be continued by the same LRR. This provides the basis for SSA recurrent and vector forecasting algorithms (Golyandina et al., 2001, Ch.2). In practice, the signal is corrupted by a perturbation, e.g., by noise, and its subspace is estimated by SSA approximately. Thus, SSA forecasting can be applied for forecasting of a time series component that is approximately governed by an LRR and is approximately separated from the residual.
Multivariate extension Multi-channel, Multivariate SSA (or M-SSA) is a natural extension of SSA to for analyzing multivariate time series, where the size of different univariate series does not have to be the same. The trajectory matrix of multi-channel time series consists of linked trajectory matrices of separate times series. The rest of the algorithm is the same as in the univariate case. System of series can be forecasted analogously to SSA recurrent and vector algorithms (Golyandina and Stepanov, 2005). MSSA has many applications. It is especially popular in analyzing and forecasting economic and financial time series with short and long series length (Patterson et al., 2011, Hassani et al., 2012, Hassani and Mahmoudvand, 2013). Other multivariate extension is 2D-SSA that can be applied to two-dimensional data like digital images (Golyandina and Usevich, 2010). The analogue of trajectory matrix is constructed by moving 2D windows of size L_x \times L_y.
MSSA and causality A question that frequently arises in time series analysis is whether one economic variable can help in predicting another economic variable. One way to address this question was proposed by Granger (1969), in which he formalized the causality concept. A comprehensive causality test based on MSSA has recently introduced for causality measurement. The test is based on the forecasting accuracy and predictability of the direction of change of the MSSA algorithms (Hassani et al., 2011 and Hassani et al.,2012).
MSSA and EMH The MSSA forecasting results can be used in examining the
efficient-market hypothesis controversy (EMH). The EMH suggests that the information contained in the price series of an asset is reflected “instantly, fully, and perpetually” in the asset’s current price. Since the price series and the information contained in it are available to all market participants, no one can benefit by attempting to take advantage of the information contained in the price history of an asset by trading in the markets. This is evaluated using two series with different series length in a multivariate system in SSA analysis (Hassani et al. 2010).
MSSA, SSA and business cycles Business cycles plays a key role in macroeconomics, and are interest for a variety of players in the economy, including central banks, policy-makers, and financial intermediaries. MSSA-based methods for tracking business cycles have been recently introduced, and have been shown to allow for a reliable assessment of the cyclical position of the economy in real-time (de Carvalho et al., 2012 and de Carvalho and Rua, 2017).
MSSA, SSA and unit root SSA's applicability to any kind of stationary or deterministically trending series has been extended to the case of a series with a stochastic trend, also known as a series with a unit root. In Hassani and Thomakos (2010) and Thomakos (2010) the basic theory on the properties and application of SSA in the case of series of a unit root is given, along with several examples. It is shown that SSA in such series produces a special kind of filter, whose form and spectral properties are derived, and that forecasting the single reconstructed component reduces to a moving average. SSA in unit roots thus provides an `optimizing' non-parametric framework for smoothing series with a unit root. This line of work is also extended to the case of two series, both of which have a unit root but are cointegrated. The application of SSA in this bivariate framework produces a smoothed series of the common root component.
Gap-filling The gap-filling versions of SSA can be used to analyze data sets that are unevenly sampled or contain
missing data (Schoellhamer, 2001; Golyandina and Osipov, 2007). Schoellhamer (2001) shows that the straightforward idea to formally calculate approximate inner products omitting unknown terms is workable for long stationary time series. Golyandina and Osipov (2007) uses the idea of filling in missing entries in vectors taken from the given subspace. The recurrent and vector SSA forecasting can be considered as particular cases of filling in algorithms described in the paper.
Detection of structural changes SSA can be effectively used as a non-parametric method of time series monitoring and
change detection. To do that, SSA performs the subspace tracking in the following way. SSA is applied sequentially to the initial parts of the series, constructs the corresponding signal subspaces and checks the distances between these subspaces and the lagged vectors formed from the few most recent observations. If these distances become too large, a structural change is suspected to have occurred in the series (Golyandina et al., 2001, Ch.3; Moskvina and Zhigljavsky, 2003). In this way, SSA could be used for
change detection not only in trends but also in the variability of the series, in the mechanism that determines dependence between different series and even in the noise structure. The method have proved to be useful in different engineering problems (e.g. Mohammad and Nishida (2011) in robotics), and has been extended to the multivariate case with corresponding analysis of detection delay and false positive rate. == Relation between SSA and other methods ==