MarketVariance-based sensitivity analysis
Company Profile

Variance-based sensitivity analysis

Variance-based sensitivity analysis is a form of global sensitivity analysis. Working within a probabilistic framework, it decomposes the variance of the output of the model or system into fractions which can be attributed to inputs or sets of inputs. For example, given a model with two inputs and one output, one might find that 70% of the output variance is caused by the variance in the first input, 20% by the variance in the second, and 10% due to interactions between the two. These percentages are directly interpreted as measures of sensitivity. Variance-based measures of sensitivity are attractive because they measure sensitivity across the whole input space, they can deal with nonlinear responses, and they can measure the effect of interactions in non-additive systems.

Decomposition of variance
From a black box perspective, any model may be viewed as a function Y=f(X), where X is a vector of d uncertain model inputs {X1, X2, ... Xd}, and Y is a chosen univariate model output (note that this approach examines scalar model outputs, but multiple outputs can be analysed by multiple independent sensitivity analyses). Furthermore, it will be assumed that the inputs are independently and uniformly distributed within the unit hypercube, i.e. X_i \in [0,1] for i=1,2,...,d . This incurs no loss of generality because any input space can be transformed onto this unit hypercube. f(X) may be decomposed in the following way, : Y = f_0 + \sum_{i=1}^d f_i(X_i) + \sum_{i where f0 is a constant and fi is a function of Xi, fij a function of Xi and Xj, etc. A condition of this decomposition is that, : \int_0^1 f_{i_1 i_2 \dots i_s}(X_{i_1},X_{i_2},\dots,X_{i_s}) dX_{k}=0, \text{ for } k = i_1,...,i_s i.e. all the terms in the functional decomposition are orthogonal. This leads to definitions of the terms of the functional decomposition in terms of conditional expected values, : f_0 = E(Y) : f_i(X_i) = E(Y|X_i) - f_0 : f_{ij}(X_i,X_j) = E(Y|X_i,X_j) - f_0 - f_i - f_j From which it can be seen that fi is the effect of varying Xi alone (known as the main effect of Xi), and fij is the effect of varying Xi and Xj simultaneously, additional to the effect of their individual variations. This is known as a second-order interaction. Higher-order terms have analogous definitions. Now, further assuming that the f(X) is square-integrable, the functional decomposition may be squared and integrated to give, : \int f^2(\mathbf{X}) d\mathbf{X} - f_0^2 = \sum_{s=1}^d \sum_{i_1 Notice that the left hand side is equal to the variance of Y, and the terms of the right hand side are variance terms, now decomposed with respect to sets of the Xi. This finally leads to the decomposition of variance expression, : \operatorname{Var}(Y) = \sum_{i=1}^d V_i + \sum_{i where : V_{i} = \operatorname{Var}_{X_i} \left( E_{\textbf{X}_{\sim i}} (Y \mid X_{i}) \right) , : V_{ij} = \operatorname{Var}_{X_{ij}} \left( E_{\textbf{X}_{\sim ij}} \left( Y \mid X_i, X_j\right)\right) - V_{i} - V_{j} and so on. The X~i notation indicates the set of all variables except Xi. The above variance decomposition shows how the variance of the model output can be decomposed into terms attributable to each input, as well as the interaction effects between them. Together, all terms sum to the total variance of the model output. ==First-order indices==
First-order indices
A direct variance-based measure of sensitivity Si, called the "first-order sensitivity index", or "main effect index" is stated as follows, : S_i = \frac{V_i}{\operatorname{Var}(Y)} This is the contribution to the output variance of the main effect of Xi, therefore it measures the effect of varying Xi alone, but averaged over variations in other input parameters. It is standardised by the total variance to provide a fractional contribution. Higher-order interaction indices Sij, Sijk and so on can be formed by dividing other terms in the variance decomposition by Var(Y). Note that this has the implication that, : \sum_{i=1}^d S_i + \sum_{i ==Total-effect index==
Total-effect index
Using the Si, Sij and higher-order indices given above, one can build a picture of the importance of each variable in determining the output variance. However, when the number of variables is large, this requires the evaluation of 2d-1 indices, which can be too computationally demanding. For this reason, a measure known as the "Total-effect index" or "Total-order index", STi, is used. This measures the contribution to the output variance of Xi, including all variance caused by its interactions, of any order, with any other input variables. It is given as, : S_{Ti} = \frac{E_{\textbf{X}_{\sim i}} \left(\operatorname{Var}_{X_i} (Y \mid \mathbf{X}_{\sim i}) \right)}{\operatorname{Var}(Y)} = 1 - \frac{\operatorname{Var}_{\textbf{X}_{\sim i}} \left(E_{X_i} (Y \mid \mathbf{X}_{\sim i}) \right)}{\operatorname{Var}(Y)} Note that unlike the Si, : \sum_{i=1}^d S_{Ti} \geq 1 due to the fact that the interaction effect between e.g. Xi and Xj is counted in both STi and STj. In fact, the sum of the STi will only be equal to 1 when the model is purely additive. ==Calculation of indices==
Calculation of indices
For analytically tractable functions, the indices above may be calculated analytically by evaluating the integrals in the decomposition. However, in the vast majority of cases they are estimated – this is usually done by the Monte Carlo method. Sampling sequences The Monte Carlo approach involves generating a sequence of randomly distributed points inside the unit hypercube (strictly speaking these will be pseudorandom). In practice, it is common to substitute random sequences with low-discrepancy sequences to improve the efficiency of the estimators. This is then known as the quasi-Monte Carlo method. Some low-discrepancy sequences commonly used in sensitivity analysis include the Sobol’ sequence and the Latin hypercube design. Procedure To calculate the indices using the (quasi) Monte Carlo method, the following steps are used: : \operatorname{Var}_{X_i}(E_{\mathbf{X}_{\sim i}}(Y|X_i)) \approx { \frac {1}{N} \sum_{j=1}^{N} f \left ( \mathbf{B} \right )_{j} \left ( f \left ( \mathbf{A}^i_B \right )_{j} - f \left ( \mathbf{A} \right )_{j} \right ) } and : E_{\mathbf{X}_{\sim i}}\left(\operatorname{Var}_{X_i}\left(Y \mid \mathbf{X}_{\sim i} \right) \right) \approx {\frac {1}{2N}\sum_{j=1}^{N} \left( f \left ( \mathbf{A} \right )_{j} - f \left (\mathbf{A}^i_B \right )_{j}\right )^2 } for the estimation of the Si and the STi respectively. Computational expense For the estimation of the Si and the STi for all input variables, N(d+2) model runs are required. Since N is often of the order of hundreds or thousands of runs, computational expense can quickly become a problem when the model takes a significant amount of time for a single run. In such cases, there are a number of techniques available to reduce the computational cost of estimating sensitivity indices, such as emulators, HDMR and FAST. ==See also==
tickerdossier.comtickerdossier.substack.com