Basic notions and notation Let be a complex -dimensional vector, and :S''{}_{xx}(x) \equiv \left( \frac{\partial^2 S(x)}{\partial x_i \partial x_j} \right), \qquad 1\leqslant i,\, j\leqslant n, denote the
Hessian matrix for a function . If :\boldsymbol{\varphi}(x) = (\varphi_1(x), \varphi_2(x), \ldots, \varphi_k(x)) is a vector function, then its
Jacobian matrix is defined as :\boldsymbol{\varphi}_x' (x) \equiv \left( \frac{\partial \varphi_i (x)}{\partial x_j} \right), \qquad 1 \leqslant i \leqslant k, \quad 1 \leqslant j \leqslant n. A
non-degenerate saddle point, , of a holomorphic function is a critical point of the function (i.e., ) where the function's Hessian matrix has a non-vanishing
determinant (i.e., \det S''{}_{zz}(z^0) \neq 0). The following is the main tool for constructing the asymptotics of integrals in the case of a non-degenerate saddle point:
Complex Morse lemma The
Morse lemma for real-valued functions generalizes as follows for
holomorphic functions: near a non-degenerate saddle point of a holomorphic function , there exist coordinates in terms of which is exactly quadratic. To make this precise, let be a holomorphic function with domain , and let in be a non-degenerate saddle point of , that is, and \det S''{}_{zz}(z^0) \neq 0. Then there exist neighborhoods of and of , and a
bijective holomorphic function with such that :\forall w \in V: \qquad S(\boldsymbol{\varphi}(w)) = S(z^0) + \frac{1}{2} \sum_{j=1}^n \mu_j w_j^2, \quad \det\boldsymbol{\varphi}_w'(0) = 1, Here, the are the
eigenvalues of the matrix S''{}_{zz}(z^0). {{math proof|title=Proof of complex Morse lemma|proof= The following proof is a straightforward generalization of the proof of the real
Morse Lemma, which can be found in . We begin by demonstrating :
Auxiliary statement. Let be
holomorphic in a neighborhood of the origin and . Then in some neighborhood, there exist functions such that f(z) = \sum_{i=1}^n z_i g_i(z), where each is
holomorphic and g_i(0) = \left. \tfrac{\partial f(z)}{\partial z_i} \right|_{z=0}. From the identity :f(z) = \int_0^1 \frac{d}{dt} f \left (t z_1,\cdots, t z_n \right ) dt = \sum_{i=1}^n z_i \int_0^1 \left. \frac{\partial f(z)}{\partial z_i}\right|_{z=(t z_1, \ldots, t z_n)} dt, we conclude that :g_i(z) = \int_0^1 \left. \frac{\partial f(z)}{\partial z_i}\right|_{z=(t z_1, \ldots, t z_n)} dt and :g_i(0) = \left. \frac{\partial f(z)}{\partial z_i} \right|_{z=0}. Without loss of generality, we translate the origin to , such that and . Using the Auxiliary Statement, we have :S(z) = \sum_{i=1}^n z_i g_i (z). Since the origin is a saddle point, :\left. \frac{\partial S(z)}{\partial z_i} \right|_{z=0} = g_i(0) = 0, we can also apply the Auxiliary Statement to the functions and obtain {{NumBlk|:|S(z) = \sum_{i,j=1}^n z_i z_j h_{ij}(z).|}} Recall that an arbitrary matrix can be represented as a sum of symmetric and anti-symmetric matrices, :A_{ij} = A_{ij}^{(s)} + A_{ij}^{(a)}, \qquad A_{ij}^{(s)} = \tfrac{1}{2}\left( A_{ij} + A_{ji} \right), \qquad A_{ij}^{(a)} = \tfrac{1}{2}\left( A_{ij} - A_{ji} \right). The contraction of any symmetric matrix
B with an arbitrary matrix is {{NumBlk|:|\sum_{i,j} B_{ij} A_{ij} = \sum_{i,j} B_{ij} A_{ij}^{(s)}, i.e., the anti-symmetric component of does not contribute because :\sum_{i,j} B_{ij} C_{ij} = \sum_{i,j} B_{ji} C_{ji} = - \sum_{i,j} B_{ij} C_{ij} = 0. Thus, in equation (1) can be assumed to be symmetric with respect to the interchange of the indices and . Note that :\left. \frac{\partial^2 S(z)}{\partial z_i \partial z_j} \right|_{z=0} = 2h_{ij}(0); hence, because the origin is a non-degenerate saddle point. Let us show by
induction that there are local coordinates , such that {{NumBlk|:|S(\boldsymbol{\psi}(u)) = \sum_{i=1}^n u_i^2.|}} First, assume that there exist local coordinates , such that {{NumBlk|:|S(\boldsymbol{\phi}(y)) = y_1^2 + \cdots + y_{r-1}^2 + \sum_{i,j = r}^n y_i y_j H_{ij} (y),|}} where is symmetric due to equation (2). By a linear change of the variables , we can assure that . From the
chain rule, we have :\frac{\partial^2 S (\boldsymbol{\phi}(y))}{\partial y_i \partial y_j} = \sum_{l,k=1}^n \left. \frac{\partial^2 S (z)}{\partial z_k \partial z_l} \right|_{z=\boldsymbol{\phi}(y)} \frac{\partial \phi_k}{\partial y_i} \frac{\partial \phi_l}{\partial y_j} + \sum_{k=1}^n \left. \frac{\partial S (z)}{\partial z_k } \right|_{z=\boldsymbol{\phi}(y)} \frac{\partial^2 \phi_k}{\partial y_i \partial y_j} Therefore: :S''{}_{yy} (\boldsymbol{\phi}(0)) = \boldsymbol{\phi}'_y(0)^T S''{}_{zz}(0) \boldsymbol{\phi}'_y(0), \qquad \det \boldsymbol{\phi}'_y(0) \neq 0; whence, :0 \neq \det S''{}_{yy} (\boldsymbol{\phi}(0)) = 2^{r-1} \det \left( 2H_{ij}(0) \right). The matrix can be recast in the
Jordan normal form: , were gives the desired non-singular linear transformation and the diagonal of contains non-zero
eigenvalues of . If then, due to continuity of , it must be also non-vanishing in some neighborhood of the origin. Having introduced \tilde{H}_{ij}(y) = H_{ij}(y)/H_{rr}(y), we write :\begin{align} S(\boldsymbol{\varphi}(y)) =& y_1^2 + \cdots + y_{r-1}^2 + H_{rr}(y) \sum_{i,j = r}^n y_i y_j \tilde{H}_{ij} (y) \\ =& y_1^2 + \cdots + y_{r-1}^2 + H_{rr}(y)\left[ y_r^2 + 2y_r \sum_{j=r+1}^n y_j \tilde{H}_{rj} (y) + \sum_{i,j = r+1}^n y_i y_j \tilde{H}_{ij} (y) \right] \\ =& y_1^2 + \cdots + y_{r-1}^2 + H_{rr}(y)\left[ \left( y_r + \sum_{j=r+1}^n y_j \tilde{H}_{rj} (y)\right)^2 - \left( \sum_{j=r+1}^n y_j \tilde{H}_{rj} (y)\right)^2 \right] + H_{rr}(y) \sum_{i,j = r+1}^n y_i y_j \tilde{H}_{ij}(y) \end{align} Motivated by the last expression, we introduce new coordinates :x_r = \sqrt{ H_{rr}(y) } \left( y_r + \sum_{j=r+1}^n y_j \tilde{H}_{rj} (y)\right), \qquad x_j = y_j, \quad \forall j \neq r. The change of the variables is locally invertible since the corresponding
Jacobian is non-zero, :\left. \frac{\partial x_r}{\partial y_k} \right|_{y=0} = \sqrt{H_{rr}(0)} \left[ \delta_{r,\, k} + \sum_{j=r+1}^n \delta_{j, \, k} \tilde{H}_{jr}(0) \right]. Therefore, {{NumBlk|:|S(\boldsymbol{\eta}(x)) = {x}_1^2 + \cdots + {x}_r^2 + \sum_{i,j = r+1}^n {x}_i {x}_j W_{ij} (x).|}} Comparing equations (4) and (5), we conclude that equation (3) is verified. Denoting the
eigenvalues of S''{}_{zz}(0) by , equation (3) can be rewritten as {{NumBlk|:|S(\boldsymbol{\varphi}(w)) = \frac 12 \sum_{j=1}^n \mu_j w_j^2.|}} Therefore, {{NumBlk|:|S''{}_{ww} (\boldsymbol{\varphi}(0)) = \boldsymbol{\varphi}'_w(0)^T S''{}_{zz}(0) \boldsymbol{\varphi}'_w(0),|}} From equation (6), it follows that \det S
{}_{ww} (\boldsymbol{\varphi}(0)) = \mu_1 \cdots \mu_n. The Jordan normal form of S{}_{zz}(0) reads S
{}_{zz}(0) = P J_z P^{-1}, where is an upper diagonal matrix containing the eigenvalues and ; hence, \det S{}_{zz} (0) = \mu_1 \cdots \mu_n. We obtain from equation (7) :\det S''{}_{ww} (\boldsymbol{\varphi}(0)) = \left[\det \boldsymbol{\varphi}'_w(0) \right]^2 \det S''{}_{zz}(0) \Longrightarrow \det \boldsymbol{\varphi}'_w(0) = \pm 1. If \det \boldsymbol{\varphi}'_w(0) = -1, then interchanging two variables assures that \det \boldsymbol{\varphi}'_w(0) = +1. }}
The asymptotic expansion in the case of a single non-degenerate saddle point Assume • and are
holomorphic functions in an
open,
bounded, and
simply connected set such that the is
connected; • \Re(S(z)) has a single maximum: \max_{z \in I_x} \Re(S(z)) = \Re(S(x^0)) for exactly one point ; • is a non-degenerate saddle point (i.e., and \det S''{}_{xx}(x^0) \neq 0). Then, the following asymptotic holds {{NumBlk|:|I(\lambda) \equiv \int_{I_x} f(x) e^{\lambda S(x)} dx = \left( \frac{2\pi}{\lambda}\right)^{\frac{n}{2}} e^{\lambda S(x^0)} \left(f(x^0)+ O\left(\lambda^{-1}\right) \right) \prod_{j=1}^n (-\mu_j)^{-\frac{1}{2}}, \qquad \lambda \to \infty,|}} where are eigenvalues of the
Hessian S''{}_{xx}(x^0) and (-\mu_j)^{-\frac{1}{2}} are defined with arguments {{NumBlk|:|\left | \arg\sqrt{-\mu_j} \right| This statement is a special case of more general results presented in Fedoryuk (1987). {{math proof|title=Derivation of equation (8)|proof= First, we deform the contour into a new contour I'_x \subset \Omega_x passing through the saddle point and sharing the boundary with . This deformation does not change the value of the integral . We employ the
Complex Morse Lemma to change the variables of integration. According to the lemma, the function maps a neighborhood onto a neighborhood containing the origin. The integral can be split into two: , where is the integral over U\cap I'_x, while is over I'_x \setminus (U\cap I'_x) (i.e., the remaining part of the contour ). Since the latter region does not contain the saddle point , the value of is exponentially smaller than as ; thus, is ignored. Introducing the contour such that U\cap I'_x = \boldsymbol{\varphi}(I_w), we have {{NumBlk|:|I_0(\lambda) = e^{\lambda S(x^0)} \int_{I_w} f[\boldsymbol{\varphi}(w)] \exp\left( \lambda \sum_{j=1}^n \tfrac{\mu_j}{2} w_j^2 \right) \left |\det\boldsymbol{\varphi}_w'(w) \right | dw.|}} Recalling that as well as \det \boldsymbol{\varphi}_w'(0) = 1, we expand the pre-exponential function f[\boldsymbol{\varphi}(w)] into a Taylor series and keep just the leading zero-order term {{NumBlk|:|I_0(\lambda) \approx f(x^0) e^{\lambda S(x^0)} \int_{\mathbf{R}^n} \exp \left( \lambda \sum_{j=1}^n \tfrac{\mu_j}{2} w_j^2 \right) dw = f(x^0)e^{\lambda S(x^0)} \prod_{j=1}^n \int_{-\infty}^{\infty} e^{\frac{1}{2}\lambda \mu_j y^2} dy.|}} Here, we have substituted the integration region by because both contain the origin, which is a saddle point, hence they are equal up to an exponentially small term. The integrals in the r.h.s. of equation (11) can be expressed as {{NumBlk|:|\mathcal{I}_j = \int_{-\infty}^{\infty} e^{\frac{1}{2} \lambda \mu_j y^2} dy = 2\int_0^{\infty} e^{-\frac{1}{2} \lambda \left(\sqrt{-\mu_j} y\right)^2} dy = 2\int_0^{\infty} e^{-\frac{1}{2} \lambda \left |\sqrt{-\mu_j} \right|^2 y^2\exp\left(2i\arg\sqrt{-\mu_j}\right)} dy.|}} From this representation, we conclude that condition (9) must be satisfied in order for the r.h.s. and l.h.s. of equation (12) to coincide. According to assumption 2, \Re \left( S''{}_{xx}(x^0) \right) is a
negatively defined quadratic form (viz., \Re(\mu_j)) implying the existence of the integral \mathcal{I}_j, which is readily calculated :\mathcal{I}_j = \frac{2}{\sqrt{-\mu_j}\sqrt{\lambda}} \int_0^{\infty} e^{-\frac{\xi^2}{2}} d\xi = \sqrt{\frac{2\pi}{\lambda}} (-\mu_j)^{-\frac{1}{2}}. }} Equation (8) can also be written as {{NumBlk|:|I(\lambda) = \left( \frac{2\pi}{\lambda}\right)^{\frac{n}{2}} e^{\lambda S(x^0)} \left ( \det (-S''{}_{xx}(x^0)) \right )^{-\frac{1}{2}} \left (f(x^0) + O\left(\lambda^{-1}\right) \right),|}} where the branch of :\sqrt{\det \left (-S''{}_{xx}(x^0) \right)} is selected as follows :\begin{align} \left (\det \left (-S
{}_{xx}(x^0) \right ) \right)^{-\frac{1}{2}} &= \exp\left( -i \text{ Ind} \left (- S{}_{xx}(x^0) \right ) \right) \prod_{j=1}^n \left| \mu_j \right|^{-\frac{1}{2}}, \\ \text{Ind} \left (-S''{}_{xx}(x^0) \right) &= \tfrac{1}{2} \sum_{j=1}^n \arg (-\mu_j), && |\arg(-\mu_j)| Consider important special cases: • If is real valued for real and in (aka, the
multidimensional Laplace method), then \text{Ind} \left(-S''{}_{xx}(x^0) \right ) = 0. • If is purely imaginary for real (i.e., \Re(S(x)) = 0 for all in ) and in (aka, the
multidimensional stationary phase method), then \text{Ind} \left (-S
{}_{xx}(x^0) \right ) = \frac{\pi}{4} \text{sign }S{}_{xx}(x_0), where \text{sign }S
{}_{xx}(x_0) denotes the signature of matrix S{}_{xx}(x_0), which equals to the number of negative eigenvalues minus the number of positive ones. It is noteworthy that in applications of the stationary phase method to the multidimensional WKB approximation in
quantum mechanics (as well as in optics), is related to the
Maslov index see, e.g., and . == The case of multiple non-degenerate saddle points ==