When
f is a set of non-linear combination of the variables
x, an
interval propagation could be performed in order to compute intervals which contain all consistent values for the variables. In a probabilistic approach, the function
f must usually be linearised by approximation to a first-order
Taylor series expansion, though in some cases, exact formulae can be derived that do not depend on the expansion as is the case for the exact variance of products. The Taylor expansion would be: f_k \approx f^0_k+ \sum_i^n \frac{\partial f_k}{\partial {x_i}} x_i where \partial f_k/\partial x_i denotes the
partial derivative of
fk with respect to the
i-th variable, evaluated at the mean value of all components of vector
x. Or in
matrix notation, \mathrm{f} \approx \mathrm{f}^0 + \mathrm{J} \mathrm{x}\, where J is the
Jacobian matrix. Since f0 is a constant it does not contribute to the error on f. Therefore, the propagation of error follows the linear case, above, but replacing the linear coefficients,
Aki and
Akj by the partial derivatives, \frac{\partial f_k}{\partial x_i} and \frac{\partial f_k}{\partial x_j}. In matrix notation, \mathrm{\Sigma}^\mathrm{f} = \mathrm{J} \mathrm{\Sigma}^\mathrm{x} \mathrm{J}^\top. That is, the Jacobian of the function is used to transform the rows and columns of the variance-covariance matrix of the argument. Note this is equivalent to the matrix expression for the linear case with \mathrm{J = A}.
Simplification Neglecting correlations or assuming independent variables yields a common formula among engineers and experimental scientists to calculate error propagation, the variance formula: s_f = \sqrt{ \left(\frac{\partial f}{\partial x}\right)^2 s_x^2 + \left(\frac{\partial f}{\partial y} \right)^2 s_y^2 + \left(\frac{\partial f}{\partial z} \right)^2 s_z^2 + \cdots} where s_f represents the standard deviation of the function f, s_x represents the standard deviation of x, s_y represents the standard deviation of y, and so forth. This formula is based on the linear characteristics of the gradient of f and therefore it is a good estimation for the standard deviation of f as long as s_x, s_y, s_z,\ldots are small enough. Specifically, the
linear approximation of f has to be close to f inside a neighbourhood of radius s_x, s_y, s_z,\ldots.
Example Any non-linear
differentiable function, f(a,b), of two variables, a and b, can be expanded as f\approx f^0+\frac{\partial f}{\partial a}a+\frac{\partial f}{\partial b}b. If we take the variance on both sides and use the formula for the variance of a
linear combination of variables \operatorname{Var}(aX + bY) = a^2\operatorname{Var}(X) + b^2\operatorname{Var}(Y) + 2ab \operatorname{Cov}(X, Y), then we obtain \sigma^2_f\approx\left| \frac{\partial f}{\partial a}\right| ^2\sigma^2_a+\left| \frac{\partial f}{\partial b}\right|^2\sigma^2_b+2\frac{\partial f}{\partial a}\frac{\partial f} {\partial b}\sigma_{ab}, where \sigma_{f} is the standard deviation of the function f, \sigma_{a} is the standard deviation of a, \sigma_{b} is the standard deviation of b and \sigma_{ab} = \sigma_{a}\sigma_{b} \rho_{ab} is the covariance between a and b. In the particular case that {{nowrap|\frac{\partial f}{\partial a} = b,}} {{nowrap|\frac{\partial f}{\partial b} = a.}} Then \sigma^2_f \approx b^2\sigma^2_a+a^2 \sigma_b^2+2ab\,\sigma_{ab} or \left(\frac{\sigma_f}{f}\right)^2 \approx \left(\frac{\sigma_a}{a} \right)^2 + \left(\frac{\sigma_b}{b}\right)^2 + 2\left(\frac{\sigma_a}{a}\right)\left(\frac{\sigma_b}{b}\right)\rho_{ab} where \rho_{ab} is the correlation between a and b. When the variables a and b are uncorrelated, \rho_{ab}=0. Then \left(\frac{\sigma_f}{f}\right)^2 \approx \left(\frac{\sigma_a}{a} \right)^2 + \left(\frac{\sigma_b}{b}\right)^2.
Caveats and warnings Error estimates for non-linear functions are
biased on account of using a truncated series expansion. The extent of this bias depends on the nature of the function. For example, the bias on the error calculated for log(1+
x) increases as
x increases, since the expansion to
x is a good approximation only when
x is near zero. For highly non-linear functions, there exist five categories of probabilistic approaches for uncertainty propagation; see
Uncertainty quantification for details.
Reciprocal and shifted reciprocal In the special case of the inverse or reciprocal 1/B, where B=N(0,1) follows a
standard normal distribution, the resulting distribution is a reciprocal standard normal distribution, and there is no definable variance. However, in the slightly more general case of a shifted reciprocal function 1/(p-B) for B=N(\mu,\sigma) following a general normal distribution, then mean and variance statistics do exist in a
principal value sense, if the difference between the pole p and the mean \mu is real-valued.
Ratios Ratios are also problematic; normal approximations exist under certain conditions. ==Example formulae==