Proof of basic form We first prove the case of constant limits of integration
a and
b. We use
Fubini's theorem to change the order of integration. For every and , such that and both and are within , we have: \begin{align} \int_x^{x+h} \int_a^b f_x(x,t) \,dt \,dx &= \int_a^b \int_x^{x+h} f_x(x,t) \,dx \,dt \\[2ex] &= \int_a^b \left(f(x+h,t) - f(x,t)\right) \,dt \\[2ex] &= \int_a^b f(x+h,t) \,dt - \int_a^b f(x,t) \,dt \end{align} Note that the integrals at hand are well defined since f_x(x,t) is continuous at the closed rectangle [x_0, x_1] \times [a,b] and thus also uniformly continuous there; thus its integrals by either
dt or
dx are continuous in the other variable and also integrable by it (essentially this is because for uniformly continuous functions, one may pass the limit through the integration sign, as elaborated below). Therefore: \begin{align} \frac{ \int_a^b f(x+h,t) \,dt - \int_a^b f(x,t) \,dt }{h} &= \frac{1}{h}\int_x^{x+h} \int_a^b f_x(x,t) \,dt \,dx \\[2ex] &= \frac{F(x+h)-F(x)}{h} \end{align} Where we have defined: F(u) := \int_{x_0}^{u} \int_a^b f_x(x,t) \,dt \,dx (we may replace
x0 here by any other point between
x0 and
x)
F is differentiable with derivative \int_a^b f_x(x,t) \,dt , so we can take the limit where approaches zero. For the left hand side this limit is: \frac{d}{dx}\int_a^b f(x,t) \, dt For the right hand side, we get: F'(x) = \int_a^b f_x(x,t) \, dt And we thus prove the desired result: \frac{d}{dx}\int_a^b f(x,t) \, dt = \int_a^b f_x(x,t) \, dt
Another proof using the bounded convergence theorem If the integrals at hand are
Lebesgue integrals, we may use the
bounded convergence theorem (valid for these integrals, but not for
Riemann integrals) in order to show that the limit can be passed through the integral sign. Note that this proof is weaker in the sense that it only shows that
fx(
x,
t) is Lebesgue integrable, but not that it is Riemann integrable. In the former (stronger) proof, if
f(
x,
t) is Riemann integrable, then so is
fx(
x,
t) (and thus is obviously also Lebesgue integrable). Let By the definition of the derivative, {{NumBlk||u'(x) = \lim_{h \to 0} \frac{u(x + h) - u(x)}{h}. |}} Substitute equation () into equation (). The difference of two integrals equals the integral of the difference, and 1/
h is a constant, so \begin{align} u'(x) &= \lim_{h \to 0} \frac{\int_a^bf(x + h, t)\,dt - \int_a^b f(x, t)\,dt}{h} \\ &= \lim_{h \to 0} \frac{\int_a^b\left( f(x + h, t) - f(x,t) \right)\,dt}{h} \\ &= \lim_{h \to 0} \int_a^b \frac{f(x + h, t) - f(x, t)}{h} \,dt. \end{align} We now show that the limit can be passed through the integral sign. We claim that the passage of the limit under the integral sign is valid by the bounded convergence theorem (a corollary of the
dominated convergence theorem). For each
δ > 0, consider the
difference quotient f_\delta(x, t) = \frac{f(x + \delta, t) - f(x, t)}{\delta}. For
t fixed, the
mean value theorem implies there exists
z in the interval [
x,
x +
δ] such that f_\delta(x, t) = f_x(z, t). Continuity of
fx(
x,
t) and compactness of the domain together imply that
fx(
x,
t) is bounded. The above application of the mean value theorem therefore gives a uniform (independent of t) bound on f_\delta(x, t). The difference quotients converge pointwise to the partial derivative
fx by the assumption that the partial derivative exists. The above argument shows that for every sequence {
δn} → 0, the sequence \{f_{\delta_n}(x, t)\} is uniformly bounded and converges pointwise to
fx. The bounded convergence theorem states that if a sequence of functions on a set of finite measure is uniformly bounded and converges pointwise, then passage of the limit under the integral is valid. In particular, the limit and integral may be exchanged for every sequence {
δn} → 0. Therefore, the limit as
δ → 0 may be passed through the integral sign. If instead we only know that there is an integrable function \theta \colon \Omega \to \mathbf{R} such that |f_x(x,\omega)| \leq \theta ( \omega), then |f_\delta(x, t)| = |f_x(z, t)|\leq \theta(\omega) and the dominated convergence theorem allows us to move the limit inside of the integral.
Variable limits form For a
continuous real valued function g of one
real variable, and real valued
differentiable functions f_1 and f_2 of one real variable, \frac{d}{dx} \left( \int_{f_1(x)}^{f_2(x)} g(t) \,dt \right )= g\left(f_2(x)\right) {f_2'(x)} - g\left(f_1(x)\right) {f_1'(x)}. This follows from the
chain rule and the
First Fundamental Theorem of Calculus. Define G(x) = \int_{f_1(x)}^{f_2(x)} g(t) \, dt, and \Gamma(x) = \int_{0}^{x} g(t) \, dt. (The lower limit just has to be some number in the domain of g ) Then, G(x) can be written as a
composition: G(x) = (\Gamma \circ f_2)(x) - (\Gamma \circ f_1)(x) . The
Chain Rule then implies that G'(x) = \Gamma'\left(f_2(x)\right) f_2'(x) - \Gamma'\left(f_1(x)\right) f_1'(x). By the
First Fundamental Theorem of Calculus, \Gamma'(x) = g(x) . Therefore, substituting this result above, we get the desired equation: G'(x) = g\left(f_2(x)\right) {f_2'(x)} - g\left(f_1(x)\right) {f_1'(x)}.
Note: This form can be particularly useful if the expression to be differentiated is of the form: \int_{f_1(x)}^{f_2(x)} h(x) \, g(t) \,dt Because h(x) does not depend on the limits of integration, it may be moved out from under the integral sign, and the above form may be used with the
Product rule, i.e., \begin{align} \frac{d}{dx} \left( \int_{f_1(x)}^{f_2(x)} h(x)g(t) \,dt \right ) &= \frac{d}{dx} \left(h(x) \int_{f_1(x)}^{f_2(x)} g(t) \,dt \right ) \\ &= h'(x)\int_{f_1(x)}^{f_2(x)} g(t) \,dt + h(x) \frac{d}{dx} \left(\int_{f_1(x)}^{f_2(x)} g(t) \,dt \right ) \end{align}
General form with variable limits Set \varphi(\alpha) = \int_a^b f(x,\alpha)\,dx, where
a and
b are functions of
α that exhibit increments Δ
a and Δ
b, respectively, when
α is increased by Δ
α. Then, \begin{align} \Delta\varphi &= \varphi(\alpha + \Delta\alpha) - \varphi(\alpha) \\[4pt] &= \int_{a + \Delta a}^{b + \Delta b}f(x, \alpha + \Delta\alpha)\,dx - \int_a^b f(x, \alpha)\,dx \\[4pt] &= \int_{a + \Delta a}^af(x, \alpha + \Delta\alpha)\,dx + \int_a^bf(x, \alpha + \Delta\alpha)\,dx + \int_b^{b + \Delta b} f(x, \alpha+\Delta\alpha)\,dx - \int_a^b f(x, \alpha)\,dx \\[4pt] &= -\int_a^{a + \Delta a} f(x, \alpha + \Delta\alpha)\,dx + \int_a^b [f(x, \alpha + \Delta\alpha) - f(x,\alpha)]\,dx + \int_b^{b + \Delta b} f(x, \alpha + \Delta\alpha)\,dx. \end{align} A form of the
mean value theorem, \int_a^b f(x)\,dx = (b - a)f(\xi), where
a \Delta\varphi = -\Delta a f(\xi_1, \alpha + \Delta\alpha) + \int_a^b [f(x, \alpha + \Delta\alpha) - f(x,\alpha)]\,dx + \Delta b f(\xi_2, \alpha + \Delta\alpha). Divide by Δ
α and let Δ
α → 0. Notice
ξ1 →
a and
ξ2 →
b. We may pass the limit through the integral sign: \lim_{\Delta\alpha\to 0}\int_a^b \frac{f(x,\alpha + \Delta\alpha) - f(x,\alpha)}{\Delta\alpha}\,dx = \int_a^b \frac{\partial}{\partial\alpha}f(x, \alpha)\,dx, again by the bounded convergence theorem. This yields the general form of the Leibniz integral rule, \frac{d\varphi}{d\alpha} = \int_a^b \frac{\partial}{\partial\alpha}f(x, \alpha)\,dx + f(b, \alpha) \frac{db}{d\alpha} - f(a, \alpha)\frac{da}{d\alpha}.
Alternative proof of the general form with variable limits, using the chain rule The general form of Leibniz's Integral Rule with variable limits can be derived as a consequence of the
basic form of Leibniz's Integral Rule, the
multivariable chain rule, and the
first fundamental theorem of calculus. Suppose f is defined in a rectangle in the x - t plane, for x \in [x_1, x_2] and t \in [t_1, t_2] . Also, assume f and the partial derivative \frac{\partial f}{\partial x} are both continuous functions on this rectangle. Suppose a, b are
differentiable real valued functions defined on [x_1, x_2], with values in [t_1, t_2] (i.e. for every x \in [x_1, x_2], a(x) , b(x) \in [t_1, t_2] ). Now, set F(x,y) = \int_{t_1}^{y} f(x,t)\,dt , \qquad \text{for} ~ x \in [x_1, x_2] ~\text{and}~ y \in [t_1, t_2] and G(x) = \int_{a(x)}^{b(x)} f(x,t)\,dt , \quad \text{for} ~ x \in [x_1, x_2] Then, by properties of
definite Integrals, we can write G(x) = \int_{t_1}^{b(x)} f(x,t)\,dt - \int_{t_1}^{a(x)} f(x,t)\,dt = F(x, b(x)) - F(x, a(x)) Since the functions F, a, b are all differentiable (see the remark at the end of the proof), by the
multivariable chain rule, it follows that G is differentiable, and its derivative is given by the formula: G'(x) = \left(\frac{\partial F}{\partial x} (x, b(x)) + \frac{\partial F}{\partial b(x)} (x, b(x) ) b'(x) \right) - \left(\frac{\partial F}{\partial x} (x, a(x)) + \frac{\partial F}{\partial a(x)} (x, a(x)) a'(x) \right) Now, note that for every x \in [x_1, x_2] , and for every y \in [t_1, t_2] , we have that \frac{\partial F}{\partial x}(x, y) = \int_{t_1}^y \frac{\partial f}{\partial x}(x,t) \, dt , because when taking the partial derivative with respect to x of F , we are keeping y fixed in the expression \int_{t_1}^{y} f(x,t)\,dt ; thus the
basic form of Leibniz's Integral Rule with constant limits of integration applies. Next, by the
first fundamental theorem of calculus, we have that \frac{\partial F}{\partial y}(x, y) = f(x,y) ; because when taking the partial derivative with respect to y of F , the first variable x is fixed, so the fundamental theorem can indeed be applied. Substituting these results into the equation for G'(x) above gives: \begin{align} G'(x) &= \left(\int_{t_1}^{b(x)} \frac{\partial f}{\partial x}(x,t) \, dt + f(x, b(x)) b'(x) \right) - \left(\int_{t_1}^{a(x)} \dfrac{\partial f}{\partial x}(x,t) \, dt + f(x, a(x)) a'(x) \right) \\[2pt] &= f(x,b(x)) b'(x) - f(x,a(x)) a'(x) + \int_{a(x)}^{b(x)} \frac{\partial f}{\partial x}(x,t) \, dt, \end{align} as desired. There is a technical point in the proof above which is worth noting: applying the Chain Rule to G requires that F already be
differentiable. This is where we use our assumptions about f . As mentioned above, the partial derivatives of F are given by the formulas \frac{\partial F}{\partial x}(x, y) = \int_{t_1}^y \frac{\partial f}{\partial x}(x,t) \, dt and \frac{\partial F}{\partial y}(x, y) = f(x,y) . Since \dfrac{\partial f}{\partial x} is continuous, its integral is also a continuous function, and since f is also continuous, these two results show that both the partial derivatives of F are continuous. Since continuity of partial derivatives implies differentiability of the function, F is indeed differentiable.
Three-dimensional, time-dependent form At time
t the surface Σ in
Figure 1 contains a set of points arranged about a centroid \mathbf{C}(t). The function \mathbf{F}(\mathbf{r}, t) can be written as \mathbf{F}(\mathbf{C}(t) + \mathbf{r} - \mathbf{C}(t), t) = \mathbf{F}(\mathbf{C}(t) + \mathbf{I}, t), with \mathbf{I} independent of time. Variables are shifted to a new frame of reference attached to the moving surface, with origin at \mathbf{C}(t). For a rigidly translating surface, the limits of integration are then independent of time, so: \frac {d}{dt} \left (\iint_{\Sigma (t)} d \mathbf{A}_{\mathbf{r}}\cdot \mathbf{F}(\mathbf{r}, t) \right) = \iint_\Sigma d \mathbf{A}_{\mathbf{I}} \cdot \frac {d}{dt}\mathbf{F}(\mathbf{C}(t) + \mathbf{I}, t), where the limits of integration confining the integral to the region Σ no longer are time dependent so differentiation passes through the integration to act on the integrand only: \frac {d}{dt}\mathbf{F}( \mathbf{C}(t) + \mathbf{I}, t) = \mathbf{F}_t(\mathbf{C}(t) + \mathbf{I}, t) + \mathbf{v \cdot \nabla F}(\mathbf{C}(t) + \mathbf{I}, t) = \mathbf{F}_t(\mathbf{r}, t) + \mathbf{v} \cdot \nabla \mathbf{F}(\mathbf{r}, t), with the velocity of motion of the surface defined by \mathbf{v} = \frac {d}{dt} \mathbf{C} (t). This equation expresses the
material derivative of the field, that is, the derivative with respect to a coordinate system attached to the moving surface. Having found the derivative, variables can be switched back to the original frame of reference. We notice that (see
article on curl) \nabla \times \left(\mathbf{v} \times \mathbf{F}\right) = (\nabla \cdot \mathbf{F} + \mathbf{F} \cdot \nabla) \mathbf{v}- (\nabla \cdot \mathbf{v} + \mathbf{v} \cdot \nabla) \mathbf{F}, and that
Stokes theorem equates the surface integral of the curl over Σ with a line integral over : \frac{d}{dt} \left(\iint_{\Sigma (t)} \mathbf{F} (\mathbf{r}, t) \cdot d \mathbf{A}\right) = \iint_{\Sigma (t)} \big(\mathbf{F}_t (\mathbf{r}, t) + \left(\mathbf{F \cdot \nabla} \right)\mathbf{v} + \left(\nabla \cdot \mathbf{F} \right) \mathbf{v} - (\nabla \cdot \mathbf{v})\mathbf{F}\big)\cdot d\mathbf{A} - \oint_{\partial \Sigma (t)}\left(\mathbf{v} \times \mathbf{F}\right)\cdot d\mathbf{s}. The sign of the line integral is based on the
right-hand rule for the choice of direction of line element
ds. To establish this sign, for example, suppose the field
F points in the positive
z-direction, and the surface Σ is a portion of the
xy-plane with perimeter ∂Σ. We adopt the normal to Σ to be in the positive
z-direction. Positive traversal of ∂Σ is then counterclockwise (right-hand rule with thumb along
z-axis). Then the integral on the left-hand side determines a
positive flux of
F through Σ. Suppose Σ translates in the positive
x-direction at velocity
v. An element of the boundary of Σ parallel to the
y-axis, say
ds, sweeps out an area
vt ×
ds in time
t. If we integrate around the boundary ∂Σ in a counterclockwise sense,
vt ×
ds points in the negative
z-direction on the left side of ∂Σ (where
ds points downward), and in the positive
z-direction on the right side of ∂Σ (where
ds points upward), which makes sense because Σ is moving to the right, adding area on the right and losing it on the left. On that basis, the flux of
F is increasing on the right of ∂Σ and decreasing on the left. However, the
dot product . Consequently, the sign of the line integral is taken as negative. If
v is a constant, \frac {d}{dt} \iint_{\Sigma (t)} \mathbf{F} (\mathbf{r}, t) \cdot d \mathbf{A} = \iint_{\Sigma (t)} \big(\mathbf{F}_t (\mathbf{r}, t) + \left(\nabla \cdot \mathbf{F} \right) \mathbf{v}\big) \cdot d \mathbf{A} - \oint_{\partial \Sigma (t)}\left(\mathbf{v} \times \mathbf{F}\right) \cdot \,d\mathbf{s}, which is the quoted result. This proof does not consider the possibility of the surface deforming as it moves.
Alternative derivation Lemma. One has: \frac{\partial}{\partial b} \left (\int_a^b f(x) \,dx \right ) = f(b), \qquad \frac{\partial}{\partial a} \left (\int_a^b f(x) \,dx \right )= -f(a).
Proof. From the
proof of the fundamental theorem of calculus, \begin{align} \frac{\partial}{\partial b} \left(\int_a^b f(x) \,dx\right) &= \lim_{\Delta b \to 0} \frac{1}{\Delta b} \left( \int_a^{b+\Delta b} f(x)\,dx - \int_a^b f(x)\,dx \right) \\[1ex] &= \lim_{\Delta b \to 0} \frac{1}{\Delta b} \left( \int_a^b f(x) \, dx + \int_b^{b+\Delta b} f(x)\,dx - \int_a^b f(x)\,dx \right) \\[1ex] &= \lim_{\Delta b \to 0} \frac{1}{\Delta b} \int_b^{b+\Delta b} f(x)\,dx \\[1ex] &= \lim_{\Delta b \to 0} \frac{1}{\Delta b} \left[ f(b) \Delta b + O\left(\Delta b^2\right) \right] \\[1ex] &= f(b), \end{align} and \begin{align} \frac{\partial}{\partial a} \left (\int_a^b f(x) \,dx \right )&= \lim_{\Delta a \to 0} \frac{1}{\Delta a} \left[ \int_{a+\Delta a}^b f(x)\,dx - \int_a^b f(x)\,dx \right] \\[6pt] &= \lim_{\Delta a \to 0} \frac{1}{\Delta a} \int_{a+\Delta a}^a f(x)\,dx \\[6pt] &= \lim_{\Delta a \to 0} \frac{1}{\Delta a} \left[ -f(a) \Delta a + O\left(\Delta a^2\right) \right]\\[6pt] &= -f(a). \end{align} Suppose
a and
b are constant, and that
f(
x) involves a parameter
α which is constant in the integration but may vary to form different integrals. Assume that
f(
x,
α) is a continuous function of
x and
α in the compact set {(
x,
α) :
α0 ≤
α ≤
α1 and
a ≤
x ≤
b}, and that the partial derivative
fα(
x,
α) exists and is continuous. If one defines: \varphi(\alpha) = \int_a^b f(x,\alpha)\,dx, then \varphi may be differentiated with respect to
α by differentiating under the integral sign, i.e., \frac{d\varphi}{d\alpha}=\int_a^b\frac{\partial}{\partial\alpha}f(x,\alpha)\,dx. By the
Heine–Cantor theorem it is uniformly continuous in that set. In other words, for any
ε > 0 there exists Δ
α such that for all values of
x in [
a,
b], |f(x,\alpha+\Delta \alpha)-f(x,\alpha)| On the other hand, \begin{align} \Delta\varphi &=\varphi(\alpha+\Delta \alpha)-\varphi(\alpha) \\[6pt] &=\int_a^b f(x,\alpha+\Delta\alpha)\,dx - \int_a^b f(x,\alpha)\, dx \\[6pt] &=\int_a^b \left (f(x,\alpha+\Delta\alpha)-f(x,\alpha) \right )\,dx \\[6pt] &\leq \varepsilon (b-a). \end{align} Hence
φ(
α) is a continuous function. Similarly if \frac{\partial}{\partial\alpha} f(x,\alpha) exists and is continuous, then for all
ε > 0 there exists Δ
α such that: \forall x \in [a, b], \quad \left|\frac{f(x,\alpha+\Delta \alpha)-f(x,\alpha)}{\Delta \alpha} - \frac{\partial f}{\partial\alpha}\right| Therefore, \frac{\Delta \varphi}{\Delta \alpha}=\int_a^b\frac{f(x,\alpha+\Delta\alpha)-f(x,\alpha)}{\Delta \alpha}\,dx = \int_a^b \frac{\partial f(x,\alpha)}{\partial \alpha}\,dx + R, where |R| Now,
ε → 0 as Δ
α → 0, so \lim_{{\Delta \alpha} \to 0}\frac{\Delta\varphi}{\Delta \alpha}= \frac{d\varphi}{d\alpha} = \int_a^b \frac{\partial}{\partial \alpha} f(x,\alpha)\,dx. This is the formula we set out to prove. Now, suppose \int_a^b f(x,\alpha)\,dx=\varphi(\alpha), where
a and
b are functions of
α which take increments Δ
a and Δ
b, respectively, when
α is increased by Δ
α. Then, \begin{align} \Delta\varphi &=\varphi(\alpha+\Delta\alpha)-\varphi(\alpha) \\[6pt] &=\int_{a+\Delta a}^{b+\Delta b}f(x,\alpha+\Delta\alpha)\,dx -\int_a^b f(x,\alpha)\,dx \\[6pt] &=\int_{a+\Delta a}^af(x,\alpha+\Delta\alpha)\,dx+\int_a^bf(x,\alpha+\Delta\alpha)\,dx+\int_b^{b+\Delta b}f(x,\alpha+\Delta\alpha)\,dx -\int_a^b f(x,\alpha)\,dx \\[6pt] &=-\int_a^{a+\Delta a} f(x,\alpha+\Delta\alpha)\,dx+\int_a^b[f(x,\alpha+\Delta\alpha)-f(x,\alpha)]\,dx+\int_b^{b+\Delta b} f(x,\alpha+\Delta\alpha)\,dx. \end{align} A form of the
mean value theorem, \int_a^b f(x)\,dx=(b-a)f(\xi), where
a \Delta\varphi=-\Delta a\,f(\xi_1,\alpha+\Delta\alpha)+\int_a^b[f(x,\alpha+\Delta\alpha)-f(x,\alpha)]\,dx+\Delta b\,f(\xi_2,\alpha+\Delta\alpha). Dividing by Δ
α, letting Δ
α → 0, noticing
ξ1 →
a and
ξ2 →
b and using the above derivation for \frac{d\varphi}{d\alpha} = \int_a^b\frac{\partial}{\partial \alpha} f(x,\alpha)\,dx yields \frac{d\varphi}{d\alpha} = \int_a^b\frac{\partial}{\partial \alpha} f(x,\alpha)\,dx+f(b,\alpha)\frac{\partial b}{\partial \alpha}-f(a,\alpha)\frac{\partial a}{\partial \alpha}. This is the general form of the Leibniz integral rule. == Examples ==