Product Rule

Get Product Rule essential facts below. View Videos or join the Product Rule discussion. Add Product Rule to your PopFlock.com topic list for future reference or share this resource on social media.
## Discovery

## Examples

## Proofs

### Proof by factoring (from first principles)

### Brief proof

### Quarter squares

### Chain rule

### Non-standard analysis

### Smooth infinitesimal analysis

## Generalizations

### Product of more than two factors

### Higher derivatives

### Higher partial derivatives

### Banach space

### Derivations in abstract algebra

### In vector calculus

## Applications

## See also

## References

This article uses material from the Wikipedia page available here. It is released under the Creative Commons Attribution-Share-Alike License 3.0.

Product Rule

In calculus, the **product rule** (or **Leibniz rule**^{[1]} or **Leibniz product rule**) is a formula used to find the derivatives of products of two or more functions. For two functions, it may be stated in Lagrange's notation as

or in Leibniz's notation as

The rule may be extended or generalized to products of three or more functions, to a rule for higher-order derivatives of a product, and to other contexts.

Discovery of this rule is credited to Gottfried Leibniz, who demonstrated it using differentials.^{[2]} (However, J. M. Child, a translator of Leibniz's papers,^{[3]} argues that it is due to Isaac Barrow.) Here is Leibniz's argument: Let *u*(*x*) and *v*(*x*) be two differentiable functions of *x*. Then the differential of *uv* is

Since the term *du*·*dv* is "negligible" (compared to *du* and *dv*), Leibniz concluded that

and this is indeed the differential form of the product rule. If we divide through by the differential *dx*, we obtain

which can also be written in Lagrange's notation as

- Suppose we want to differentiate
*f*(*x*) =*x*^{2}sin(*x*). By using the product rule, one gets the derivative*f*(*x*) = 2*x*sin(*x*) +*x*^{2}cos(*x*) (since the derivative of*x*^{2}is 2*x*and the derivative of the sine function is the cosine function). - One special case of the product rule is the constant multiple rule, which states: if
*c*is a number and*f*(*x*) is a differentiable function, then*cf*(*x*) is also differentiable, and its derivative is (*cf*)(*x*) =*cf*(*x*). This follows from the product rule since the derivative of any constant is zero. This, combined with the sum rule for derivatives, shows that differentiation is linear. - The rule for integration by parts is derived from the product rule, as is (a weak version of) the quotient rule. (It is a "weak" version in that it does not prove that the quotient is differentiable, but only says what its derivative is
*if*it is differentiable.)

Let *h*(*x*) = *f*(*x*)*g*(*x*) and suppose that f and g are each differentiable at x. We want to prove that h is differentiable at x and that its derivative, *h*(*x*), is given by *f*(*x*)*g*(*x*) + *f*(*x*)*g*(*x*). To do this, (which is zero, and thus does not change the value) is added to the numerator to permit its factoring, and then properties of limits are used.

The fact that

is deduced from a theorem that states that differentiable functions are continuous.

By definition, if are differentiable at then we can write

such that also written . Then:

The "other terms" consist of items such as and It is not difficult to show that they are all Dividing by and taking the limit for small gives the result.

There is a proof using quarter square multiplication which relies on the chain rule and on the properties of the quarter square function (shown here as *q*, i.e., with ):

Differentiating both sides:

The product rule can be considered a special case of the chain rule for several variables.

Let *u* and *v* be continuous functions in *x*, and let *dx*, *du* and *dv* be infinitesimals within the framework of non-standard analysis, specifically the hyperreal numbers. Using st to denote the standard part function that associates to a finite hyperreal number the real infinitely close to it, this gives

This was essentially Leibniz's proof exploiting the transcendental law of homogeneity (in place of the standard part above).

In the context of Lawvere's approach to infinitesimals, let *dx* be a nilsquare infinitesimal. Then *du* = *u*′ *dx* and *dv* = *v* ′ *dx*, so that

since

The product rule can be generalized to products of more than two factors. For example, for three factors we have

For a collection of functions , we have

The logarithmic derivative provides a simpler expression of the last form, as well as a direct proof that does not involve any recursion. The *logarithmic derivative* of a function f, denoted here Logder(*f*), is the derivative of the logarithm of the function. It follows that

Using that the logarithm of a product is the sum of the logarithms of the factors, the sum rule for derivatives gives immediately

The last above expression of the derivative of a product is obtained by multiplying both members of this equation by the product of the

It can also be generalized to the general Leibniz rule for the *n*th derivative of a product of two factors, by symbolically expanding according to the binomial theorem:

Applied at a specific point *x*, the above formula gives:

Furthermore, for the *n*th derivative of an arbitrary number of factors:

For partial derivatives, we have^{[4]}

where the index S runs through all 2^{n} subsets of {1, ..., *n*}, and || is the cardinality of S. For example, when *n* = 3,

Suppose *X*, *Y*, and *Z* are Banach spaces (which includes Euclidean space) and *B* : *X* × *Y* -> *Z* is a continuous bilinear operator. Then *B* is differentiable, and its derivative at the point (*x*,*y*) in *X* × *Y* is the linear map *D*_{(x,y)}*B* : *X* × *Y* -> *Z* given by

In abstract algebra, the product rule is used to *define* what is called a derivation, not vice versa.

The product rule extends to scalar multiplication, dot products, and cross products of vector functions, as follows.^{[5]}

For scalar multiplication:

For dot products:

For cross products:

There are also analogues for other analogs of the derivative: if *f* and *g* are scalar fields then there is a product rule with the gradient:

Among the applications of the product rule is a proof that

when *n* is a positive integer (this rule is true even if *n* is not positive or is not an integer, but the proof of that must rely on other methods). The proof is by mathematical induction on the exponent *n*. If *n* = 0 then *x*^{n} is constant and *nx*^{n − 1} = 0. The rule holds in that case because the derivative of a constant function is 0. If the rule holds for any particular exponent *n*, then for the next value, *n* + 1, we have

Therefore, if the proposition is true for *n*, it is true also for *n* + 1, and therefore for all natural *n*.

**^**https://encyclopediaofmath.orghttps://www.popflock.com/learn?s=Leibniz_rule**^**Michelle Cirillo (August 2007). "Humanizing Calculus".*The Mathematics Teacher*.**101**(1): 23-27.**^**Leibniz, G. W. (2005) [1920],*The Early Mathematical Manuscripts of Leibniz*(PDF), translated by J.M. Child, Dover, p. 28, footnote 58, ISBN 978-0-486-44596-0**^**Micheal Hardy (January 2006). "Combinatorics of Partial Derivatives" (PDF).*The Electronic Journal of Combinatorics*.**13**.**^**Stewart, James (2016),*Calculus*(8 ed.), Cengage, Section 13.2.

This article uses material from the Wikipedia page available here. It is released under the Creative Commons Attribution-Share-Alike License 3.0.

Popular Products

Music Scenes

Popular Artists