In nonlinear regression, a
statistical model of the form, \mathbf{y} \sim f(\mathbf{x}, \boldsymbol\beta) relates a vector of
independent variables, \mathbf{x}, and its associated observed
dependent variables, \mathbf{y}. The function f is nonlinear in the components of the vector of parameters \beta, but otherwise arbitrary. For example, the
Michaelis–Menten model for enzyme kinetics has two parameters and one independent variable, related by f by:{{efn|This model can also be expressed in the conventional biological notation: v = \frac{V_\max\ [\mathrm{S}]}{K_m + [\mathrm{S}]} }} f(x,\boldsymbol\beta)= \frac{\beta_1 x}{\beta_2 + x} This function, which is a rectangular hyperbola, is because it cannot be expressed as a
linear combination of the two \betas.
Systematic error may be present in the independent variables but its treatment is outside the scope of regression analysis. If the independent variables are not error-free, this is an
errors-in-variables model, also outside this scope. Other examples of nonlinear functions include
exponential functions,
logarithmic functions,
trigonometric functions,
power functions,
Gaussian function, and
Lorentz distributions. Some functions, such as the exponential or logarithmic functions, can be transformed so that they are linear. When so transformed, standard linear regression can be performed but must be applied with caution. See , below, for more details. In microbiology and biotechnology, nonlinear regression is used to model complex microbial growth kinetics. While simple growth follows monoauxic functions (such as the Gompertz or Boltzmann models), multiphasic (polyauxic) growth is modeled using linear combinations of these nonlinear functions. Estimating parameters for these complex models often requires robust regression techniques (e.g., using a Lorentzian loss function to mitigate outliers) and global optimization algorithms (such as
Differential evolution with
L-BFGS-B) to avoid local minima and ensure biologically interpretable results. In general, there is no closed-form expression for the best-fitting parameters, as there is in
linear regression. Usually numerical
optimization algorithms are applied to determine the best-fitting parameters. Again in contrast to linear regression, there may be many
local minima of the function to be optimized and even the global minimum may produce a
biased estimate. In practice,
estimated values of the parameters are used, in conjunction with the optimization algorithm, to attempt to find the global minimum of a sum of squares. For details concerning nonlinear data modeling see
least squares and
non-linear least squares. ==Regression statistics==