Two very commonly used loss functions are the
squared loss, L(a) = a^2, and the
absolute loss, L(a)=|a|. The squared loss function results in an
arithmetic mean-
unbiased estimator, and the absolute-value loss function results in a
median-unbiased estimator (in the one-dimensional case, and a
geometric median-unbiased estimator for the multi-dimensional case). The squared loss has the disadvantage that it has the tendency to be dominated by outliers—when summing over a set of a's (as in \sum_{i=1}^n L(a_i) ), the sample mean is influenced too much by a few particularly large a-values when the distribution is heavy tailed: in terms of
estimation theory, the asymptotic relative efficiency of the mean is poor for heavy-tailed distributions. As defined above, the Huber loss function is
strongly convex in a uniform neighborhood of its minimum a=0; at the boundary of this uniform neighborhood, the Huber loss function has a differentiable extension to an affine function at points a=-\delta and a = \delta . These properties allow it to combine much of the sensitivity of the mean-unbiased, minimum-variance estimator of the mean (using the quadratic loss function) and the robustness of the median-unbiased estimator (using the absolute value function). ==Pseudo-Huber loss function==