HOME

TheInfoList



OR:

In
calculus Calculus, originally called infinitesimal calculus or "the calculus of infinitesimals", is the mathematical study of continuous change, in the same way that geometry is the study of shape, and algebra is the study of generalizations of arithm ...
, the Leibniz integral rule for differentiation under the integral sign, named after
Gottfried Leibniz Gottfried Wilhelm (von) Leibniz . ( – 14 November 1716) was a German polymath active as a mathematician, philosopher, scientist and diplomat. He is one of the most prominent figures in both the history of philosophy and the history of mathem ...
, states that for an
integral In mathematics, an integral assigns numbers to functions in a way that describes displacement, area, volume, and other concepts that arise by combining infinitesimal data. The process of finding integrals is called integration. Along wit ...
of the form \int_^ f(x,t)\,dt, where -\infty < a(x), b(x) < \infty and the integral are functions dependent on x, the derivative of this integral is expressible as \frac \left (\int_^ f(x,t)\,dt \right )= f\big(x,b(x)\big)\cdot \frac b(x) - f\big(x,a(x)\big)\cdot \frac a(x) + \int_^\frac f(x,t) \,dt, where the partial derivative \tfrac indicates that inside the integral, only the variation of f(x, t) with x is considered in taking the derivative. In the special case where the functions a(x) and b(x) are constants a(x)=a and b(x)=b with values that do not depend on x, this simplifies to: \frac \left(\int_a^b f(x,t)\,dt \right)= \int_a^b \frac f(x,t) \,dt. If a(x)=a is constant and b(x)=x, which is another common situation (for example, in the proof of Cauchy's repeated integration formula), the Leibniz integral rule becomes: \frac \left (\int_a^x f(x,t) \, dt \right )= f\big(x,x\big) + \int_a^x \frac f(x,t) \, dt, This important result may, under certain conditions, be used to interchange the integral and partial differential
operators Operator may refer to: Mathematics * A symbol indicating a mathematical operation * Logical operator or logical connective in mathematical logic * Operator (mathematics), mapping that acts on elements of a space to produce elements of another sp ...
, and is particularly useful in the differentiation of
integral transform In mathematics, an integral transform maps a function from its original function space into another function space via integration, where some of the properties of the original function might be more easily characterized and manipulated than in ...
s. An example of such is the
moment generating function In probability theory and statistics, the moment-generating function of a real-valued random variable is an alternative specification of its probability distribution. Thus, it provides the basis of an alternative route to analytical results compare ...
in
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set ...
, a variation of the
Laplace transform In mathematics, the Laplace transform, named after its discoverer Pierre-Simon Laplace (), is an integral transform that converts a function of a real variable (usually t, in the '' time domain'') to a function of a complex variable s (in the ...
, which can be differentiated to generate the moments of a random variable. Whether Leibniz's integral rule applies is essentially a question about the interchange of
limits Limit or Limits may refer to: Arts and media * ''Limit'' (manga), a manga by Keiko Suenobu * ''Limit'' (film), a South Korean film * Limit (music), a way to characterize harmony * "Limit" (song), a 2016 single by Luna Sea * "Limits", a 2019 ...
.


General form: differentiation under the integral sign

Stronger versions of the theorem only require that the partial derivative exist almost everywhere, and not that it be continuous. This formula is the general form of the Leibniz integral rule and can be derived using the
fundamental theorem of calculus The fundamental theorem of calculus is a theorem that links the concept of differentiating a function (calculating its slopes, or rate of change at each time) with the concept of integrating a function (calculating the area under its graph, or ...
. The (first) fundamental theorem of calculus is just the particular case of the above formula where a(x)=a\in\mathbb, b(x) = x, and f(x, t) = f(t). If both upper and lower limits are taken as constants, then the formula takes the shape of an operator equation: \mathcal_t \partial_x = \partial_x \mathcal_t where \partial_x is the partial derivative with respect to x and \mathcal_t is the integral operator with respect to t over a fixed interval. That is, it is related to the symmetry of second derivatives, but involving integrals as well as derivatives. This case is also known as the Leibniz integral rule. The following three basic theorems on the interchange of limits are essentially equivalent: * the interchange of a derivative and an integral (differentiation under the integral sign; i.e., Leibniz integral rule); * the change of order of partial derivatives; * the change of order of integration (integration under the integral sign; i.e.,
Fubini's theorem In mathematical analysis Fubini's theorem is a result that gives conditions under which it is possible to compute a double integral by using an iterated integral, introduced by Guido Fubini in 1907. One may switch the order of integration if th ...
).


Three-dimensional, time-dependent case

A Leibniz integral rule for a two dimensional surface moving in three dimensional space is \frac \iint_ \mathbf (\mathbf, t) \cdot d \mathbf = \iint_ \left(\mathbf_t (\mathbf, t) + \left nabla \cdot \mathbf (\mathbf, t) \right\mathbf \right) \cdot d \mathbf - \oint_ \left \mathbf \times \mathbf ( \mathbf, t) \right\cdot d \mathbf, where: * is a vector field at the spatial position at time , * is a surface bounded by the closed curve , * is a vector element of the surface , * is a vector element of the curve , * is the velocity of movement of the region , * is the vector
divergence In vector calculus, divergence is a vector operator that operates on a vector field, producing a scalar field giving the quantity of the vector field's source at each point. More technically, the divergence represents the volume density of t ...
, * is the
vector cross product In mathematics, the cross product or vector product (occasionally directed area product, to emphasize its geometric significance) is a binary operation on two vectors in a three-dimensional oriented Euclidean vector space (named here E), and is d ...
, *The double integrals are surface integrals over the surface , and the
line integral In mathematics, a line integral is an integral where the function to be integrated is evaluated along a curve. The terms ''path integral'', ''curve integral'', and ''curvilinear integral'' are also used; ''contour integral'' is used as well, al ...
is over the bounding curve .


Higher dimensions

The Leibniz integral rule can be extended to multidimensional integrals. In two and three dimensions, this rule is better known from the field of fluid dynamics as the
Reynolds transport theorem In differential calculus, the Reynolds transport theorem (also known as the Leibniz–Reynolds transport theorem), or simply the Reynolds theorem, named after Osborne Reynolds (1842–1912), is a three-dimensional generalization of the Leibniz in ...
: \frac \int_ F(\mathbf x, t) \,dV = \int_ \frac F(\mathbf x, t)\,dV + \int_ F(\mathbf x, t) \mathbf v_b \cdot d\mathbf, where F(\mathbf x, t) is a scalar function, and denote a time-varying connected region of R3 and its boundary, respectively, \mathbf v_b is the Eulerian velocity of the boundary (see
Lagrangian and Eulerian coordinates __NOTOC__ In classical field theories, the Lagrangian specification of the flow field is a way of looking at fluid motion where the observer follows an individual fluid parcel as it moves through space and time. Plotting the position of an indi ...
) and is the unit normal component of the
surface A surface, as the term is most generally used, is the outermost or uppermost layer of a physical object or space. It is the portion or region of the object that can first be perceived by an observer using the senses of sight and touch, and is ...
element. The general statement of the Leibniz integral rule requires concepts from differential geometry, specifically
differential forms In mathematics, differential forms provide a unified approach to define integrands over curves, surfaces, solids, and higher-dimensional manifolds. The modern notion of differential forms was pioneered by Élie Cartan. It has many applications, ...
, exterior derivatives,
wedge product A wedge is a triangular shaped tool, and is a portable inclined plane, and one of the six simple machines. It can be used to separate two objects or portions of an object, lift up an object, or hold an object in place. It functions by convert ...
s and
interior product In mathematics, the interior product (also known as interior derivative, interior multiplication, inner multiplication, inner derivative, insertion operator, or inner derivation) is a degree −1 (anti)derivation on the exterior algebra of ...
s. With those tools, the Leibniz integral rule in ''n'' dimensions is \frac\int_\omega=\int_ i_(d_x\omega)+\int_ i_ \omega + \int_ \dot, where is a time-varying domain of integration, ''ω'' is a ''p''-form, \mathbf v=\frac is the vector field of the velocity, i_ denotes the
interior product In mathematics, the interior product (also known as interior derivative, interior multiplication, inner multiplication, inner derivative, insertion operator, or inner derivation) is a degree −1 (anti)derivation on the exterior algebra of ...
with \mathbf v, ''d''''x''''ω'' is the exterior derivative of ''ω'' with respect to the space variables only and \dot is the time derivative of ''ω''. However, all of these identities can be derived from a most general statement about Lie derivatives: \left.\frac\_\int_ \omega = \int_ \mathcal_\Psi \omega, Here, the ambient manifold on which the differential form \omega lives includes both space and time. *\Omega is the region of integration (a submanifold) at a given instant (it does not depend on t, since its parametrization as a submanifold defines its position in time), *\mathcal is the
Lie derivative In differential geometry, the Lie derivative ( ), named after Sophus Lie by Władysław Ślebodziński, evaluates the change of a tensor field (including scalar functions, vector fields and one-forms), along the flow defined by another vector fi ...
, *\Psi is the spacetime vector field obtained from adding the unitary vector field in the direction of time to the purely spatial vector field \mathbf v from the previous formulas (i.e, \Psi is the spacetime velocity of \Omega), *\psi_t is a diffeomorphism from the one-parameter group generated by the flow of \Psi, and *\text_(\Omega) is the image of \Omega under such diffeomorphism. Something remarkable about this form, is that it can account for the case when \Omega changes its shape and size over time, since such deformations are fully determined by \Psi.


Measure theory statement

Let X be an open subset of \mathbf, and \Omega be a measure space. Suppose f\colon X \times \Omega \to \mathbf satisfies the following conditions: #f(x,\omega) is a Lebesgue-integrable function of \omega for each x \in X. #For almost all \omega \in \Omega , the partial derivative f_x exists for all x \in X. #There is an integrable function \theta \colon \Omega \to \mathbf such that , f_x(x,\omega), \leq \theta ( \omega) for all x \in X and almost every \omega \in \Omega. Then, for all x \in X, \frac \int_\Omega f(x, \omega) \, d\omega = \int_ f_x (x, \omega) \, d\omega. The proof relies on the
dominated convergence theorem In measure theory, Lebesgue's dominated convergence theorem provides sufficient conditions under which almost everywhere convergence of a sequence of functions implies convergence in the ''L''1 norm. Its power and utility are two of the primary t ...
and the mean value theorem (details below).


Proofs


Proof of basic form

We first prove the case of constant limits of integration ''a'' and ''b''. We use
Fubini's theorem In mathematical analysis Fubini's theorem is a result that gives conditions under which it is possible to compute a double integral by using an iterated integral, introduced by Guido Fubini in 1907. One may switch the order of integration if th ...
to change the order of integration. For every and , such that and both and are within , we have: \int_x^ \int_a^b f_x(x,t) \,dt \,dx = \int_a^b \int_x^ f_x(x,t) \,dx \,dt = \int_a^b \left(f(x+h,t) - f(x,t)\right) \,dt = \int_a^b f(x+h,t) \,dt - \int_a^b f(x,t) \,dt Note that the integrals at hand are well defined since f_x(x,t) is continuous at the closed rectangle _0, x_1\times ,b and thus also uniformly continuous there; thus its integrals by either ''dt'' or ''dx'' are continuous in the other variable and also integrable by it (essentially this is because for uniformly continuous functions, one may pass the limit through the integration sign, as elaborated below). Therefore: \frac = \frac\int_x^ \int_a^b f_x(x,t) \,dt \,dx = \frac Where we have defined: F(u) \equiv \int_^ \int_a^b f_x(x,t) \,dt \,dx (we may replace ''x''0 here by any other point between ''x''0 and ''x'') ''F'' is differentiable with derivative \int_a^b f_x(x,t) \,dt , so we can take the limit where approaches zero. For the left hand side this limit is: \frac\int_a^b f(x,t) \, dt For the right hand side, we get: F'(x) = \int_a^b f_x(x,t) \, dt And we thus prove the desired result: \frac\int_a^b f(x,t) \, dt = \int_a^b f_x(x,t) \, dt


Another proof using the bounded convergence theorem

If the integrals at hand are
Lebesgue integral In mathematics, the integral of a non-negative function of a single variable can be regarded, in the simplest case, as the area between the graph of that function and the -axis. The Lebesgue integral, named after French mathematician Henri Lebe ...
s, we may use the bounded convergence theorem (valid for these integrals, but not for
Riemann integral In the branch of mathematics known as real analysis, the Riemann integral, created by Bernhard Riemann, was the first rigorous definition of the integral of a function on an interval. It was presented to the faculty at the University of GÃ ...
s) in order to show that the limit can be passed through the integral sign. Note that this proof is weaker in the sense that it only shows that ''fx''(''x'',''t'') is Lebesgue integrable, but not that it is Riemann integrable. In the former (stronger) proof, if ''f''(''x'',''t'') is Riemann integrable, then so is ''fx''(''x'',''t'') (and thus is obviously also Lebesgue integrable). Let By the definition of the derivative, Substitute equation () into equation (). The difference of two integrals equals the integral of the difference, and 1/''h'' is a constant, so \begin u'(x) &= \lim_ \frac \\ &= \lim_ \frac \\ &= \lim_ \int_a^b \frac \,dt. \end We now show that the limit can be passed through the integral sign. We claim that the passage of the limit under the integral sign is valid by the bounded convergence theorem (a corollary of the
dominated convergence theorem In measure theory, Lebesgue's dominated convergence theorem provides sufficient conditions under which almost everywhere convergence of a sequence of functions implies convergence in the ''L''1 norm. Its power and utility are two of the primary t ...
). For each ''δ'' > 0, consider the
difference quotient In single-variable calculus, the difference quotient is usually the name for the expression : \frac which when taken to the limit as ''h'' approaches 0 gives the derivative of the function ''f''. The name of the expression stems from the fact ...
f_\delta(x, t) = \frac. For ''t'' fixed, the mean value theorem implies there exists ''z'' in the interval 'x'', ''x'' + ''δ''such that f_\delta(x, t) = f_x(z, t). Continuity of ''f''''x''(''x'', ''t'') and compactness of the domain together imply that ''f''''x''(''x'', ''t'') is bounded. The above application of the mean value theorem therefore gives a uniform (independent of t) bound on f_\delta(x, t). The difference quotients converge pointwise to the partial derivative ''f''''x'' by the assumption that the partial derivative exists. The above argument shows that for every sequence → 0, the sequence \ is uniformly bounded and converges pointwise to ''f''''x''. The bounded convergence theorem states that if a sequence of functions on a set of finite measure is uniformly bounded and converges pointwise, then passage of the limit under the integral is valid. In particular, the limit and integral may be exchanged for every sequence → 0. Therefore, the limit as ''δ'' → 0 may be passed through the integral sign.


Variable limits form

For a
continuous Continuity or continuous may refer to: Mathematics * Continuity (mathematics), the opposing concept to discreteness; common examples include ** Continuous probability distribution or random variable in probability and statistics ** Continuous ...
real valued function ''g'' of one real variable, and real valued
differentiable In mathematics, a differentiable function of one real variable is a function whose derivative exists at each point in its domain. In other words, the graph of a differentiable function has a non-vertical tangent line at each interior point in its ...
functions f_1 and f_2 of one real variable, \frac \left( \int_^ g(t) \,dt \right )= g\left(f_2(x)\right) - g\left(f_1(x)\right) . This follows from the
chain rule In calculus, the chain rule is a formula that expresses the derivative of the composition of two differentiable functions and in terms of the derivatives of and . More precisely, if h=f\circ g is the function such that h(x)=f(g(x)) for every , ...
and the
First Fundamental Theorem of Calculus The fundamental theorem of calculus is a theorem that links the concept of differentiating a function (calculating its slopes, or rate of change at each time) with the concept of integrating a function (calculating the area under its graph, or ...
. Define G(x) = \int_^ g(t) \, dt, and \Gamma(x) = \int_^ g(t) \, dt. (The lower limit just has to be some number in the domain of g ) Then, G(x) can be written as a
composition Composition or Compositions may refer to: Arts and literature *Composition (dance), practice and teaching of choreography *Composition (language), in literature and rhetoric, producing a work in spoken tradition and written discourse, to include v ...
: G(x) = (\Gamma \circ f_2)(x) - (\Gamma \circ f_1)(x) . The
Chain Rule In calculus, the chain rule is a formula that expresses the derivative of the composition of two differentiable functions and in terms of the derivatives of and . More precisely, if h=f\circ g is the function such that h(x)=f(g(x)) for every , ...
then implies that G'(x) = \Gamma'\left(f_2(x)\right) f_2'(x) - \Gamma'\left(f_1(x)\right) f_1'(x). By the
First Fundamental Theorem of Calculus The fundamental theorem of calculus is a theorem that links the concept of differentiating a function (calculating its slopes, or rate of change at each time) with the concept of integrating a function (calculating the area under its graph, or ...
, \Gamma'(x) = g(x) . Therefore, substituting this result above, we get the desired equation: G'(x) = g\left(f_2(x)\right) - g\left(f_1(x)\right) . Note: This form can be particularly useful if the expression to be differentiated is of the form: \int_^ h(x)g(t) \,dt Because h(x) does not depend on the limits of integration, it may be moved out from under the integral sign, and the above form may be used with the
Product rule In calculus, the product rule (or Leibniz rule or Leibniz product rule) is a formula used to find the derivatives of products of two or more functions. For two functions, it may be stated in Lagrange's notation as (u \cdot v)' = u ' \cdot v ...
, i.e., \frac \left( \int_^ h(x)g(t) \,dt \right ) = \frac \left(h(x) \int_^ g(t) \,dt \right ) = h'(x)\int_^ g(t) \,dt + h(x) \frac \left(\int_^ g(t) \,dt \right )


General form with variable limits

Set \varphi(\alpha) = \int_a^b f(x,\alpha)\,dx, where ''a'' and ''b'' are functions of ''α'' that exhibit increments Δ''a'' and Δ''b'', respectively, when ''α'' is increased by Δ''α''. Then, \begin \Delta\varphi &= \varphi(\alpha + \Delta\alpha) - \varphi(\alpha) \\ pt&= \int_^f(x, \alpha + \Delta\alpha)\,dx - \int_a^b f(x, \alpha)\,dx \\ pt&= \int_^af(x, \alpha + \Delta\alpha)\,dx + \int_a^bf(x, \alpha + \Delta\alpha)\,dx + \int_b^ f(x, \alpha+\Delta\alpha)\,dx - \int_a^b f(x, \alpha)\,dx \\ pt&= -\int_a^ f(x, \alpha + \Delta\alpha)\,dx + \int_a^b (x, \alpha + \Delta\alpha) - f(x,\alpha),dx + \int_b^ f(x, \alpha + \Delta\alpha)\,dx. \end A form of the mean value theorem, \int_a^b f(x)\,dx = (b - a)f(\xi), where ''a'' < ''ξ'' < ''b'', may be applied to the first and last integrals of the formula for Δ''φ'' above, resulting in \Delta\varphi = -\Delta a f(\xi_1, \alpha + \Delta\alpha) + \int_a^b (x, \alpha + \Delta\alpha) - f(x,\alpha),dx + \Delta b f(\xi_2, \alpha + \Delta\alpha). Divide by Δ''α'' and let Δ''α'' → 0. Notice ''ξ''1 → ''a'' and ''ξ''2 → ''b''. We may pass the limit through the integral sign: \lim_\int_a^b \frac\,dx = \int_a^b \fracf(x, \alpha)\,dx, again by the bounded convergence theorem. This yields the general form of the Leibniz integral rule, \frac = \int_a^b \fracf(x, \alpha)\,dx + f(b, \alpha) \frac - f(a, \alpha)\frac.


Alternative proof of the general form with variable limits, using the chain rule

The general form of Leibniz's Integral Rule with variable limits can be derived as a consequence of the basic form of Leibniz's Integral Rule, the multivariable chain rule, and the
First Fundamental Theorem of Calculus The fundamental theorem of calculus is a theorem that links the concept of differentiating a function (calculating its slopes, or rate of change at each time) with the concept of integrating a function (calculating the area under its graph, or ...
. Suppose f is defined in a rectangle in the x-t plane, for x \in _1, x_2 and t \in
_1, t_2 Onekama ( ) is a village in Manistee County in the U.S. state of Michigan. The population was 411 at the 2010 census. The village is located on the shores of Portage Lake and is surrounded by Onekama Township. The town's name is derived from "On ...
. Also, assume f and the partial derivative \frac are both continuous functions on this rectangle. Suppose a, b are
differentiable In mathematics, a differentiable function of one real variable is a function whose derivative exists at each point in its domain. In other words, the graph of a differentiable function has a non-vertical tangent line at each interior point in its ...
real valued functions defined on _1, x_2/math>, with values in
_1, t_2 Onekama ( ) is a village in Manistee County in the U.S. state of Michigan. The population was 411 at the 2010 census. The village is located on the shores of Portage Lake and is surrounded by Onekama Township. The town's name is derived from "On ...
(i.e. for every x \in _1, x_2 a(x) , b(x) \in
_1, t_2 Onekama ( ) is a village in Manistee County in the U.S. state of Michigan. The population was 411 at the 2010 census. The village is located on the shores of Portage Lake and is surrounded by Onekama Township. The town's name is derived from "On ...
). Now, set F(x,y) = \int_^ f(x,t)\,dt , \qquad \text ~ x \in _1, x_2~\text~ y \in
_1, t_2 Onekama ( ) is a village in Manistee County in the U.S. state of Michigan. The population was 411 at the 2010 census. The village is located on the shores of Portage Lake and is surrounded by Onekama Township. The town's name is derived from "On ...
and G(x) = \int_^ f(x,t)\,dt , \quad \text ~ x \in _1, x_2 Then, by properties of Definite Integrals, we can write \begin G(x) &= \int_^ f(x,t)\,dt - \int_^ f(x,t)\,dt \\ pt&= F(x, b(x)) - F(x, a(x)) \end Since the functions F, a, b are all differentiable (see the remark at the end of the proof), by the Multivariable Chain Rule, it follows that G is differentiable, and its derivative is given by the formula: G'(x) = \left(\frac (x, b(x)) + \frac (x, b(x) ) b'(x) \right) - \left(\frac (x, a(x)) + \frac (x, a(x)) a'(x) \right) Now, note that for every x \in _1, x_2, and for every y \in
_1, t_2 Onekama ( ) is a village in Manistee County in the U.S. state of Michigan. The population was 411 at the 2010 census. The village is located on the shores of Portage Lake and is surrounded by Onekama Township. The town's name is derived from "On ...
, we have that \frac(x, y) = \int_^y \frac(x,t) \, dt , because when taking the partial derivative with respect to x of F , we are keeping y fixed in the expression \int_^ f(x,t)\,dt ; thus the basic form of Leibniz's Integral Rule with constant limits of integration applies. Next, by the
First Fundamental Theorem of Calculus The fundamental theorem of calculus is a theorem that links the concept of differentiating a function (calculating its slopes, or rate of change at each time) with the concept of integrating a function (calculating the area under its graph, or ...
, we have that \frac(x, y) = f(x,y) ; because when taking the partial derivative with respect to y of F , the first variable x is fixed, so the fundamental theorem can indeed be applied. Substituting these results into the equation for G'(x) above gives: \begin G'(x) &= \left(\int_^ \frac(x,t) \, dt + f(x, b(x)) b'(x) \right) - \left(\int_^ \dfrac(x,t) \, dt + f(x, a(x)) a'(x) \right) \\ pt&= f(x,b(x)) b'(x) - f(x,a(x)) a'(x) + \int_^ \frac(x,t) \, dt, \end as desired. There is a technical point in the proof above which is worth noting: applying the Chain Rule to G requires that F already be
differentiable In mathematics, a differentiable function of one real variable is a function whose derivative exists at each point in its domain. In other words, the graph of a differentiable function has a non-vertical tangent line at each interior point in its ...
. This is where we use our assumptions about f . As mentioned above, the partial derivatives of F are given by the formulas \frac(x, y) = \int_^y \frac(x,t) \, dt and \frac(x, y) = f(x,y) . Since \dfrac is continuous, its integral is also a continuous function, and since f is also continuous, these two results show that both the partial derivatives of F are continuous. Since continuity of partial derivatives implies differentiability of the function, F is indeed differentiable.


Three-dimensional, time-dependent form

At time ''t'' the surface Σ in
Figure 1 Figure 1 is a Toronto, Ontario-based online social networking service for healthcare professionals to post and comment on medical images. Figure 1 was founded in Toronto by Dr. Joshua Landy, Richard Penner and Gregory Levey. The platform launched ...
contains a set of points arranged about a centroid \mathbf(t). The function \mathbf(\mathbf, t) can be written as \mathbf(\mathbf(t) + \mathbf - \mathbf(t), t) = \mathbf(\mathbf(t) + \mathbf, t), with \mathbf independent of time. Variables are shifted to a new frame of reference attached to the moving surface, with origin at \mathbf(t). For a rigidly translating surface, the limits of integration are then independent of time, so: \frac \left (\iint_ d \mathbf_\cdot \mathbf(\mathbf, t) \right) = \iint_\Sigma d \mathbf_ \cdot \frac \mathbf(\mathbf(t) + \mathbf, t), where the limits of integration confining the integral to the region Σ no longer are time dependent so differentiation passes through the integration to act on the integrand only: \frac \mathbf( \mathbf(t) + \mathbf, t) = \mathbf_t(\mathbf(t) + \mathbf, t) + \mathbf(\mathbf(t) + \mathbf, t) = \mathbf_t(\mathbf, t) + \mathbf \cdot \nabla \mathbf(\mathbf, t), with the velocity of motion of the surface defined by \mathbf = \frac \mathbf (t). This equation expresses the
material derivative In continuum mechanics, the material derivative describes the time rate of change of some physical quantity (like heat or momentum) of a material element that is subjected to a space-and-time-dependent macroscopic velocity field. The material der ...
of the field, that is, the derivative with respect to a coordinate system attached to the moving surface. Having found the derivative, variables can be switched back to the original frame of reference. We notice that (see article on curl) \nabla \times \left(\mathbf \times \mathbf\right) = (\nabla \cdot \mathbf + \mathbf \cdot \nabla) \mathbf- (\nabla \cdot \mathbf + \mathbf \cdot \nabla) \mathbf, and that Stokes theorem equates the surface integral of the curl over Σ with a line integral over : \frac \left(\iint_ \mathbf (\mathbf, t) \cdot d \mathbf\right) = \iint_ \big(\mathbf_t (\mathbf, t) + \left(\mathbf \right)\mathbf + \left(\nabla \cdot \mathbf \right) \mathbf - (\nabla \cdot \mathbf)\mathbf\big)\cdot d\mathbf - \oint_\left(\mathbf \times \mathbf\right)\cdot d\mathbf. The sign of the line integral is based on the right-hand rule for the choice of direction of line element ''d''s. To establish this sign, for example, suppose the field F points in the positive ''z''-direction, and the surface Σ is a portion of the ''xy''-plane with perimeter ∂Σ. We adopt the normal to Σ to be in the positive ''z''-direction. Positive traversal of ∂Σ is then counterclockwise (right-hand rule with thumb along ''z''-axis). Then the integral on the left-hand side determines a ''positive'' flux of F through Σ. Suppose Σ translates in the positive ''x''-direction at velocity v. An element of the boundary of Σ parallel to the ''y''-axis, say ''d''s, sweeps out an area v''t'' × ''d''s in time ''t''. If we integrate around the boundary ∂Σ in a counterclockwise sense, v''t'' × ''d''s points in the negative ''z''-direction on the left side of ∂Σ (where ''d''s points downward), and in the positive ''z''-direction on the right side of ∂Σ (where ''d''s points upward), which makes sense because Σ is moving to the right, adding area on the right and losing it on the left. On that basis, the flux of F is increasing on the right of ∂Σ and decreasing on the left. However, the dot product . Consequently, the sign of the line integral is taken as negative. If v is a constant, \frac \iint_ \mathbf (\mathbf, t) \cdot d \mathbf = \iint_ \big(\mathbf_t (\mathbf, t) + \left(\nabla \cdot \mathbf \right) \mathbf\big) \cdot d \mathbf - \oint_\left(\mathbf \times \mathbf\right) \cdot \,d\mathbf, which is the quoted result. This proof does not consider the possibility of the surface deforming as it moves.


Alternative derivation

Lemma. One has: \frac \left (\int_a^b f(x) \,dx \right ) = f(b), \qquad \frac \left (\int_a^b f(x) \,dx \right )= -f(a). Proof. From the proof of the fundamental theorem of calculus, \begin \frac \left (\int_a^b f(x) \,dx \right ) &= \lim_ \frac \left \int_a^ f(x)\,dx - \int_a^b f(x)\,dx \right\\ pt &= \lim_ \frac \int_b^ f(x)\,dx \\ pt &= \lim_ \frac \left f(b) \Delta b + O\left(\Delta b^2\right) \right \\ pt &= f(b), \end and \begin \frac \left (\int_a^b f(x) \,dx \right )&= \lim_ \frac \left \int_^b f(x)\,dx - \int_a^b f(x)\,dx \right\\ pt &= \lim_ \frac \int_^a f(x)\,dx \\ pt &= \lim_ \frac \left -f(a) \Delta a + O\left(\Delta a^2\right) \right\ pt &= -f(a). \end Suppose ''a'' and ''b'' are constant, and that ''f''(''x'') involves a parameter ''α'' which is constant in the integration but may vary to form different integrals. Assume that ''f''(''x'', ''α'') is a continuous function of ''x'' and ''α'' in the compact set , and that the partial derivative ''f''''α''(''x'', ''α'') exists and is continuous. If one defines: \varphi(\alpha) = \int_a^b f(x,\alpha)\,dx, then \varphi may be differentiated with respect to ''α'' by differentiating under the integral sign, i.e., \frac=\int_a^b\fracf(x,\alpha)\,dx. By the
Heine–Cantor theorem In mathematics, the Heine–Cantor theorem, named after Eduard Heine and Georg Cantor, states that if f \colon M \to N is a continuous function between two metric spaces M and N, and M is compact, then f is uniformly continuous. An important speci ...
it is uniformly continuous in that set. In other words, for any ''ε'' > 0 there exists Δ''α'' such that for all values of ''x'' in 'a'', ''b'' , f(x,\alpha+\Delta \alpha)-f(x,\alpha), <\varepsilon. On the other hand, \begin \Delta\varphi &=\varphi(\alpha+\Delta \alpha)-\varphi(\alpha) \\ pt&=\int_a^b f(x,\alpha+\Delta\alpha)\,dx - \int_a^b f(x,\alpha)\, dx \\ pt&=\int_a^b \left (f(x,\alpha+\Delta\alpha)-f(x,\alpha) \right )\,dx \\ pt&\leq \varepsilon (b-a). \end Hence ''φ''(''α'') is a continuous function. Similarly if \frac f(x,\alpha) exists and is continuous, then for all ''ε'' > 0 there exists Δ''α'' such that: \forall x \in
, b The comma is a punctuation mark that appears in several variants in different languages. It has the same shape as an apostrophe or single closing quotation mark () in many typefaces, but it differs from them in being placed on the baseline o ...
\quad \left, \frac - \frac\<\varepsilon. Therefore, \frac=\int_a^b\frac\,dx = \int_a^b \frac\,dx + R, where , R, < \int_a^b \varepsilon\, dx = \varepsilon(b-a). Now, ''ε'' → 0 as Δ''α'' → 0, so \lim_\frac= \frac = \int_a^b \frac f(x,\alpha)\,dx. This is the formula we set out to prove. Now, suppose \int_a^b f(x,\alpha)\,dx=\varphi(\alpha), where ''a'' and ''b'' are functions of ''α'' which take increments Δ''a'' and Δ''b'', respectively, when ''α'' is increased by Δ''α''. Then, \begin \Delta\varphi &=\varphi(\alpha+\Delta\alpha)-\varphi(\alpha) \\ pt&=\int_^f(x,\alpha+\Delta\alpha)\,dx -\int_a^b f(x,\alpha)\,dx \\ pt&=\int_^af(x,\alpha+\Delta\alpha)\,dx+\int_a^bf(x,\alpha+\Delta\alpha)\,dx+\int_b^f(x,\alpha+\Delta\alpha)\,dx -\int_a^b f(x,\alpha)\,dx \\ pt&=-\int_a^ f(x,\alpha+\Delta\alpha)\,dx+\int_a^b (x,\alpha+\Delta\alpha)-f(x,\alpha),dx+\int_b^ f(x,\alpha+\Delta\alpha)\,dx. \end A form of the mean value theorem, \int_a^b f(x)\,dx=(b-a)f(\xi), where ''a'' < ''ξ'' < ''b'', can be applied to the first and last integrals of the formula for Δ''φ'' above, resulting in \Delta\varphi=-\Delta a\,f(\xi_1,\alpha+\Delta\alpha)+\int_a^b (x,\alpha+\Delta\alpha)-f(x,\alpha),dx+\Delta b\,f(\xi_2,\alpha+\Delta\alpha). Dividing by Δ''α'', letting Δ''α'' → 0, noticing ''ξ''1 → ''a'' and ''ξ''2 → ''b'' and using the above derivation for \frac = \int_a^b\frac f(x,\alpha)\,dx yields \frac = \int_a^b\frac f(x,\alpha)\,dx+f(b,\alpha)\frac-f(a,\alpha)\frac. This is the general form of the Leibniz integral rule.


Examples


Example 1: Fixed limits

Consider the function \varphi(\alpha)=\int_0^1\frac\,dx. The function under the integral sign is not continuous at the point (''x'', ''α'') = (0, 0), and the function ''φ''(''α'') has a discontinuity at ''α'' = 0 because ''φ''(''α'') approaches ±''π''/2 as ''α'' → 0±. If we differentiate ''φ''(''α'') with respect to ''α'' under the integral sign, we get \frac \varphi(\alpha)=\int_0^1\frac\left(\frac\right)\,dx=\int_0^1\frac dx=\left.-\frac\_0^1=-\frac, which is, of course, true for all values of ''α'' except ''α'' = 0. This may be integrated (with respect to ''α'') to find \varphi(\alpha) = \begin 0, & \alpha = 0, \\ -\arctan()+\frac, & \alpha \neq 0. \end


Example 2: Variable limits

An example with variable limits: \begin \frac \int_^ \cosh t^2\,dt &= \cosh\left(\cos^2 x\right) \frac(\cos x) - \cosh\left(\sin^2 x\right) \frac (\sin x) + \int_^ \frac (\cosh t^2) \, dt \\ pt &= \cosh(\cos^2 x) (-\sin x) - \cosh(\sin^2 x) (\cos x) + 0 \\ pt &= - \cosh(\cos^2 x) \sin x - \cosh(\sin^2 x) \cos x. \end


Applications


Evaluating definite integrals

The formula \frac \left (\int_^f(x,t) \, dt \right) = f\big(x,b(x)\big)\cdot \frac b(x) - f\big(x,a(x)\big)\cdot \frac a(x) + \int_^\frac f(x,t) \, dt can be of use when evaluating certain definite integrals. When used in this context, the Leibniz integral rule for differentiating under the integral sign is also known as Feynman's trick for integration.


Example 3

Consider \varphi(\alpha)=\int_0^\pi \ln \left (1-2\alpha\cos(x)+\alpha^2 \right )\,dx, \qquad , \alpha, > 1. Now, \begin \frac \varphi(\alpha) &=\int_0^\pi \frac dx \\ pt &=\frac\int_0^\pi \left(1-\frac \right) dx \\ pt &=\left. \frac-\frac\left\ \_0^\pi. \end As x varies from 0 to \pi, we have \begin \frac \tan\left(\frac\right) \geq 0, & , \alpha, < 1, \\ \frac \tan \left( \frac\right) \leq 0, & , \alpha, > 1. \end Hence, \left. \arctan\left(\frac\tan\left(\frac\right)\right)\_0^\pi= \begin \frac, & , \alpha, < 1, \\ -\frac, & , \alpha, > 1. \end Therefore, \frac \varphi(\alpha)= \begin 0, & , \alpha, < 1, \\ \frac, & , \alpha, > 1. \end Integrating both sides with respect to \alpha, we get: \varphi (\alpha) = \begin C_1, & , \alpha, < 1, \\ 2\pi \ln , \alpha, + C_2, & , \alpha, > 1. \end C_ = 0 follows from evaluating \varphi (0): \varphi(0) =\int_0^\pi \ln(1)\,dx =\int_0^\pi 0\,dx=0. To determine C_2 in the same manner, we should need to substitute in a value of \alpha greater than 1 in \varphi (\alpha). This is somewhat inconvenient. Instead, we substitute \alpha = \frac, where , \beta, < 1. Then, \begin \varphi(\alpha) &=\int_0^\pi\left(\ln \left (1-2\beta \cos(x)+\beta^2 \right )-2\ln, \beta, \right) dx \\ pt &= \int_0^\pi \ln \left (1-2\beta \cos(x)+\beta^2 \right )\,dx -\int_0^\pi 2\ln, \beta, dx \\ pt &=0-2\pi\ln, \beta, \\ pt &=2\pi\ln, \alpha, . \end Therefore, C_ = 0 The definition of \varphi (\alpha) is now complete: \varphi (\alpha) = \begin 0, & , \alpha, < 1, \\ 2\pi \ln , \alpha, , & , \alpha, > 1. \end The foregoing discussion, of course, does not apply when \alpha = \pm 1, since the conditions for differentiability are not met.


Example 4

\mathbf I = \int_0^ \frac\,dx,\qquad a,b > 0. First we calculate: \begin \mathbf &= \int_0^ \frac dx \\ pt&= \int_0^ \frac dx \\ pt&= \int_0^ \frac dx \\ pt&= \frac \int_0^ \frac\,d(\tan x) \\ pt&= \left.\frac\arctan \left(\sqrt\tan x\right) \_0^ \\ pt&= \frac. \end The limits of integration being independent of a, we have: \frac=-\int_0^ \frac\,dx On the other hand: \frac= \frac \left(\frac\right) =-\frac. Equating these two relations then yields \int_0^ \frac\,dx=\frac. In a similar fashion, pursuing \frac yields \int_0^\frac\,dx = \frac. Adding the two results then produces \mathbf I = \int_0^\frac\,dx=\frac\left(\frac+\frac\right), which computes \mathbf I as desired. This derivation may be generalized. Note that if we define \mathbf I_n = \int_0^ \frac\,dx, it can easily be shown that (1-n)\mathbf I_n = \frac + \frac Given \mathbf_1, this integral reduction formula can be used to compute all of the values of \mathbf_n for n > 1. Integrals like \mathbf and \mathbf may also be handled using the Weierstrass substitution.


Example 5

Here, we consider the integral \mathbf I(\alpha)=\int_0^ \frac\,dx, \qquad 0 < \alpha < \pi. Differentiating under the integral with respect to \alpha, we have \begin \frac \mathbf(\alpha) &= \int_0^ \frac \left(\frac\right) \, dx \\ pt&=-\int_0^\frac\,dx \\ &=-\int_0^\frac \, dx \\ pt&=-\frac \int_0^ \frac \frac \, dx \\ pt&=-\frac \int_0^ \frac \, dx \\ pt&=-\frac \int_0^ \frac \, d\left(\tan \frac\right)\\ pt&=-2\cot \frac\int_0^ \frac\,d\left(\tan \frac\right)\\ pt&=-2\arctan \left(\tan \frac \tan \frac \right) \bigg, _0^\\ pt&=-\alpha. \end Therefore: \mathbf(\alpha) = C - \frac. But \mathbf \left(\frac \right) = 0 by definition so C = \frac and \mathbf I(\alpha) = \frac-\frac.


Example 6

Here, we consider the integral \int_0^ e^ \cos(\sin\theta) \, d\theta. We introduce a new variable ''φ'' and rewrite the integral as f(\varphi) = \int_0^ e^ \cos(\varphi\sin\theta)\,d\theta. When ''φ'' = 1 this equals the original integral. However, this more general integral may be differentiated with respect to \varphi: \begin \frac &= \int_0^ \frac\left(e^ \cos(\varphi\sin\theta)\right)\,d\theta \\ pt &= \int_0^ e^ \left( \cos\theta\cos(\varphi\sin\theta)-\sin\theta\sin(\varphi\sin\theta) \right)\,d\theta. \end Now, fix ''φ'', and consider the vector field on \mathbf^2 defined by \mathbf(x,y) = (F_1(x,y), F_2(x,y)) := (e^ \sin (\varphi y), e^ \cos (\varphi y)) . Further, choose the positive oriented parametrization of the
unit circle In mathematics, a unit circle is a circle of unit radius—that is, a radius of 1. Frequently, especially in trigonometry, the unit circle is the circle of radius 1 centered at the origin (0, 0) in the Cartesian coordinate system in the Eucli ...
S^1 given by \mathbf \colon \begin &_\int_0^_e^_\left(_\cos\theta\cos(\varphi\sin\theta)-\sin\theta\sin(\varphi\sin\theta)_\right)\,d\theta_\\_pt=__&_\int_0^_(e^_\sin_(\varphi_\sin_\theta),_e^_\cos_(\varphi_\sin_\theta))_\cdot_(-\sin_\theta,_\cos_\theta)_\,_d\theta\\_pt=__&_\int_0^_\mathbf(\mathbf(\theta))_\cdot_\mathbf'(\theta)_\,_d\theta\\_pt=__&_\oint__\mathbf(\mathbf)_\cdot_d\mathbf =_\oint__F_1_\,_dx_+_F_2_\,_dy, \end the_line_integral_of_\mathbf_over_S^1._By_
\begin &_\int_0^_e^_\left(_\cos\theta\cos(\varphi\sin\theta)-\sin\theta\sin(\varphi\sin\theta)_\right)\,d\theta_\\_pt=__&_\int_0^_(e^_\sin_(\varphi_\sin_\theta),_e^_\cos_(\varphi_\sin_\theta))_\cdot_(-\sin_\theta,_\cos_\theta)_\,_d\theta\\_pt=__&_\int_0^_\mathbf(\mathbf(\theta))_\cdot_\mathbf'(\theta)_\,_d\theta\\_pt=__&_\oint__\mathbf(\mathbf)_\cdot_d\mathbf =_\oint__F_1_\,_dx_+_F_2_\,_dy, \end the_line_integral_of_\mathbf_over_S^1._By_Green's_theorem">Green's_Theorem,_this_equals_the_double_integral \iint_D_\frac_-_\frac_\,_dA, where_D_is_the_closed_Unit_disk.html" "title="Green's_theorem.html" ;"title=", 2\pi) \to \mathbf^2 , \mathbf(\theta) := (\cos \theta, \sin \theta) , so that \mathbf'(t) = (-\sin \theta, \cos \theta) . Then the final integral above is precisely \begin & \int_0^ e^ \left( \cos\theta\cos(\varphi\sin\theta)-\sin\theta\sin(\varphi\sin\theta) \right)\,d\theta \\ pt= & \int_0^ (e^ \sin (\varphi \sin \theta), e^ \cos (\varphi \sin \theta)) \cdot (-\sin \theta, \cos \theta) \, d\theta\\ pt= & \int_0^ \mathbf(\mathbf(\theta)) \cdot \mathbf'(\theta) \, d\theta\\ pt= & \oint_ \mathbf(\mathbf) \cdot d\mathbf = \oint_ F_1 \, dx + F_2 \, dy, \end the line integral of \mathbf over S^1. By Green's theorem">Green's Theorem, this equals the double integral \iint_D \frac - \frac \, dA, where D is the closed Unit disk">unit disc In mathematics, the open unit disk (or disc) around ''P'' (where ''P'' is a given point in the plane), is the set of points whose distance from ''P'' is less than 1: :D_1(P) = \.\, The closed unit disk around ''P'' is the set of points whose d ...
. Its integrand is identically 0, so df/d\varphi is likewise identically zero. This implies that ''f''(''φ'') is constant. The constant may be determined by evaluating f at \varphi = 0: f(0) = \int_0^ 1\,d\theta = 2\pi. Therefore, the original integral also equals 2\pi.


Other problems to solve

There are innumerable other integrals that can be solved using the technique of differentiation under the integral sign. For example, in each of the following cases, the original integral may be replaced by a similar integral having a new parameter \alpha: \begin \int_0^\infty \frac\,dx &\to \int_0^\infty e^ \frac dx,\\ pt\int_0^ \frac\,dx &\to\int_0^ \frac dx,\\ pt\int_0^\infty \frac\,dx &\to\int_0^\infty \frac dx \\ pt\int_0^1 \frac\,dx &\to \int_0^1 \frac dx. \end The first integral, the Dirichlet integral, is absolutely convergent for positive ''α'' but only conditionally convergent when \alpha = 0. Therefore, differentiation under the integral sign is easy to justify when \alpha > 0, but proving that the resulting formula remains valid when \alpha = 0 requires some careful work.


Infinite series

The measure-theoretic version of differentiation under the integral sign also applies to summation (finite or infinite) by interpreting summation as
counting measure In mathematics, specifically measure theory, the counting measure is an intuitive way to put a measure on any set – the "size" of a subset is taken to be the number of elements in the subset if the subset has finitely many elements, and infinity ...
. An example of an application is the fact that power series are differentiable in their radius of convergence.


In popular culture

Differentiation under the integral sign is mentioned in the late
physicist A physicist is a scientist who specializes in the field of physics, which encompasses the interactions of matter and energy at all length and time scales in the physical universe. Physicists generally are interested in the root or ultimate cau ...
Richard Feynman Richard Phillips Feynman (; May 11, 1918 – February 15, 1988) was an American theoretical physicist, known for his work in the path integral formulation of quantum mechanics, the theory of quantum electrodynamics, the physics of the superfl ...
's best-selling memoir ''
Surely You're Joking, Mr. Feynman! ''"Surely You're Joking, Mr. Feynman!": Adventures of a Curious Character'' is an edited collection of reminiscences by the Nobel Prize–winning physicist Richard Feynman. The book, released in 1985, covers a variety of instances in Feynman's l ...
'' in the chapter "A Different Box of Tools". He describes learning it, while in
high school A secondary school describes an institution that provides secondary education and also usually includes the building where this takes place. Some secondary schools provide both '' lower secondary education'' (ages 11 to 14) and ''upper seconda ...
, from an old text, ''Advanced Calculus'' (1926), by Frederick S. Woods (who was a professor of mathematics in the
Massachusetts Institute of Technology The Massachusetts Institute of Technology (MIT) is a private land-grant research university in Cambridge, Massachusetts. Established in 1861, MIT has played a key role in the development of modern technology and science, and is one of the ...
). The technique was not often taught when Feynman later received his formal education in
calculus Calculus, originally called infinitesimal calculus or "the calculus of infinitesimals", is the mathematical study of continuous change, in the same way that geometry is the study of shape, and algebra is the study of generalizations of arithm ...
, but using this technique, Feynman was able to solve otherwise difficult integration problems upon his arrival at graduate school at
Princeton University Princeton University is a private research university in Princeton, New Jersey. Founded in 1746 in Elizabeth as the College of New Jersey, Princeton is the fourth-oldest institution of higher education in the United States and one of the ...
:
One thing I never did learn was
contour integration In the mathematical field of complex analysis, contour integration is a method of evaluating certain integrals along paths in the complex plane. Contour integration is closely related to the calculus of residues, a method of complex analysis. ...
. I had learned to do integrals by various methods shown in a book that my high school physics teacher Mr. Bader had given me. One day he told me to stay after class. "Feynman," he said, "you talk too much and you make too much noise. I know why. You're bored. So I'm going to give you a book. You go up there in the back, in the corner, and study this book, and when you know everything that's in this book, you can talk again." So every physics class, I paid no attention to what was going on with Pascal's Law, or whatever they were doing. I was up in the back with this book
"Advanced Calculus"
by Woods. Bader knew I had studie
"Calculus for the Practical Man"
a little bit, so he gave me the real works—it was for a junior or senior course in college. It had Fourier series,
Bessel function Bessel functions, first defined by the mathematician Daniel Bernoulli and then generalized by Friedrich Bessel, are canonical solutions of Bessel's differential equation x^2 \frac + x \frac + \left(x^2 - \alpha^2 \right)y = 0 for an arbitrar ...
s,
determinant In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if a ...
s, elliptic functions—all kinds of wonderful stuff that I didn't know anything about. That book also showed how to differentiate parameters under the integral sign—it's a certain operation. It turns out that's not taught very much in the universities; they don't emphasize it. But I caught on how to use that method, and I used that one damn tool again and again. So because I was self-taught using that book, I had peculiar methods of doing integrals. The result was, when guys at MIT or
Princeton Princeton University is a private research university in Princeton, New Jersey. Founded in 1746 in Elizabeth as the College of New Jersey, Princeton is the fourth-oldest institution of higher education in the United States and one of the ni ...
had trouble doing a certain integral, it was because they couldn't do it with the standard methods they had learned in school. If it was contour integration, they would have found it; if it was a simple series expansion, they would have found it. Then I come along and try differentiating under the integral sign, and often it worked. So I got a great reputation for doing integrals, only because my box of tools was different from everybody else's, and they had tried all their tools on it before giving the problem to me.


See also

*
Chain rule In calculus, the chain rule is a formula that expresses the derivative of the composition of two differentiable functions and in terms of the derivatives of and . More precisely, if h=f\circ g is the function such that h(x)=f(g(x)) for every , ...
* Differentiation of integrals *
Leibniz rule (generalized product rule) In calculus, the general Leibniz rule, named after Gottfried Wilhelm Leibniz, generalizes the product rule (which is also known as "Leibniz's rule"). It states that if f and g are n-times differentiable functions, then the product fg is also n-t ...
*
Reynolds transport theorem In differential calculus, the Reynolds transport theorem (also known as the Leibniz–Reynolds transport theorem), or simply the Reynolds theorem, named after Osborne Reynolds (1842–1912), is a three-dimensional generalization of the Leibniz in ...
, a generalization of Leibniz rule


References


Further reading

* *


External links

*{{cite web , url=https://math.hawaii.edu/~rharron/teaching/MAT203/LeibnizRule.pdf , title=The Leibniz Rule , first=Rob , last=Harron , work=MAT-203 Gottfried Wilhelm Leibniz Multivariable calculus Integral calculus Differential calculus Articles containing proofs