Fundamental theorem of calculus

Fundamental theorem of calculus

The fundamental theorem of calculus specifies the relationship between the two central operations of calculus, differentiation and integration.

The first part of the theorem, sometimes called the first fundamental theorem of calculus, shows that an indefinite integration [More exactly, the theorem deals with definite integration with variable upper limit and arbitrarily selected lower limit. This particular kind of definite integration allows us to compute one of the infinitely many antiderivatives of a function (except for those which do not have a zero). Hence, it is almost equivalent to indefinite integration, defined by most authors as an operation which yields any one of the possible antiderivatives of a function, including those without a zero.] can be reversed by a differentiation.

The second part, sometimes called the second fundamental theorem of calculus, allows one to compute the definite integral of a function by using any one of its infinitely many antiderivatives. This part of the theorem has invaluable practical applications, because it markedly simplifies the computation of definite integrals.

The first published statement and proof of a restricted version of the fundamental theorem was by James Gregory (1638-1675) [See, e.g., Marlow Anderson, Victor J. Katz, Robin J. Wilson, "Sherlock Holmes in Babylon and Other Tales of Mathematical History", Mathematical Association of America, 2004, [ p. 114] .] . Isaac Barrow proved the first completely general version of the theorem, while Barrow's student Isaac Newton (1643–1727) completed the development of the surrounding mathematical theory. Gottfried Leibniz (1646–1716) systematized the knowledge into a calculus for infinitesimal quantities.

The Fundamental Theorem of Calculus is sometimes also called Leibniz's Fundamental Theorem of Calculus or the Torricelli–Barrow theorem.

Physical intuition

Intuitively, the theorem simply states that the sum of infinitesimal changes in a quantity over time (or some other quantity) add up to the net change in the quantity.

To comprehend this statement, we will start with an example. Suppose a particle travels in a straight line with its position, "x", given by "x"("t") where "t" is time and "x"("t") means that "x" is a function of "t". The derivative of this function is equal to the infinitesimal change in quantity, d"x", per infinitesimal change in time, d"t" (of course, the derivative itself is dependent on time). Let us define this change in distance per change in time as the velocity "v" of the particle. In Leibniz's notation:

:frac{dx}{dt} = v(t).

Rearranging this equation, it follows that:

:dx = v(t),dt.

By the logic above, a change in "x" (or Δ"x") is the sum of the infinitesimal changes d"x". It is also equal to the sum of the infinitesimal products of the derivative and time. This infinite summation is integration; hence, the integration operation allows the recovery of the original function from its derivative. As one can reasonably infer, this operation works in reverse as we can differentiate the result of our integral to recover the original derivative.

Geometric intuition

Suppose we are given a smooth continuous function nowrap|1="y" = ƒ("x"), and have plotted its graph as a curve. Then for each value of "x", intuitively it makes sense that there is a corresponding area function "A"("x"), representing the area beneath the curve between 0 and "x". We may not know a "formula" for the function "A"("x"), but intuitively we understand that it is simply the area under the curve.

Now suppose we wanted to compute the area under the curve between "x" and "x"+"h". We could compute this area by finding the area between 0 and "x"+"h", then deducting the area between 0 and "x". In other words, the area of this “sliver” would be nowrap|"A"("x"+"h") − "A"("x").

There is another way to "estimate" the area of this same sliver. Multiply "h" by ƒ("x") to find the area of a rectangle that is approximately the same size as this sliver. In fact, it makes intuitive sense that for very small values of "h", the approximation will become very good.

At this point we know that "A"("x"+"h") − "A"("x") is approximately equal to ƒ("x")·"h", and we intuitively understand that this approximation becomes better as "h" grows smaller. In other words, nowrap|ƒ("x")·"h" ≈ "A"("x"+"h") − "A"("x"), with this approximation becoming an equality as "h" approaches 0 in the limit.

Divide both sides of this equation by "h". Then we have :f(x) approx frac{A(x+h)-A(x)}{h}.

As "h" approaches 0, we recognized that the right hand side of this equation is simply the derivative "A"’("x") of the area function "A"("x"). The left-hand side of the equation simply remains ƒ("x"), since no "h" is present.

We have shown informally that nowrap|1=ƒ("x") = "A"’("x"). In other words, the derivative of the area function "A"("x") is the original function ƒ("x"). Or, to put it another way, the area function is simply the antiderivative of the original function.

What we have shown is that, intuitively, computing the derivative of a function and “finding the area” under its curve are "opposite" operations. This is the crux of the Fundamental Theorem of Calculus. The bulk of the theorem's proof is devoted to showing that the area function "A"("x") exists in the first place.

Formal statements

There are two parts to the Fundamental Theorem of Calculus. Loosely put, the first part deals with the derivative of an antiderivative, while the second part deals with the relationship between antiderivatives and definite integrals.

First part

This part is sometimes referred to as the First Fundamental Theorem of Calculus.

Let "f" be a continuous real-valued function defined on a closed interval ["a", "b"] . Let "F" be the function defined, for all "x" in ["a", "b"] , by:F(x) = int_a^x f(t), dt,.Then, "F" is continuous on ["a", "b"] , differentiable on the open interval ("a", "b"), and:F'(x) = f(x),for all "x" in ("a", "b").

Second part

This part is sometimes referred to as the Second Fundamental Theorem of Calculus.

Let "f" be a continuous real-valued function defined on a closed interval ["a", "b"] . Let "F" be an antiderivative of "f", that is one of the infinitely many functions such that, for all "x" in ["a", "b"] ,:f(x) = F'(x),.Then:int_a^b f(x),dx = F(b) - F(a),.


Let "f" be a continuous real-valued function defined on a closed interval ["a", "b"] . Let "F" be a function such that, for all "x" in ["a", "b"] ,:f(x) = F'(x),.Then, for all "x" in ["a", "b"] ,:F(x) = int_a^x f(t),dt + F(a)and:f(x) = frac{d}{dx} int_a^x f(t),dt.


As an example, suppose you need to calculate

:int_2^5 x^2, dx. Here, f(x) = x^2 and we can use F(x) = {x^3over 3} as the antiderivative. Therefore:

:int_2^5 x^2, dx = F(5) - F(2) = {125 over 3} - {8 over 3} = {117 over 3} = 39.

Or, more generally, suppose you need to calculate

:{d over dx} int_0^x t^3, dt.

Here, f(t) = t^3 and we can use F(t) = {t^4 over 4} as the antiderivative. Therefore:

:{d over dx} int_0^x t^3, dt = {d over dx} F(x) - {d over dx} F(0) = {d over dx} {x^4 over 4} = x^3.

But this result could have been found much more easily as

:{d over dx} int_0^x t^3, dt = f(x) {dx over dx} - f(0) {d0 over dx} = x^3.

Proof of the First Part

For a given "f"("t"), define the function "F"("x") as:F(x) = int_{a}^{x} f(t) ,dt,.

For any two numbers "x"1 and "x"1 + Δ"x" in ["a", "b"] , we have:F(x_1) = int_{a}^{x_1} f(t) ,dtand:F(x_1 + Delta x) = int_{a}^{x_1 + Delta x} f(t) ,dt,.

Subtracting the two equations gives:F(x_1 + Delta x) - F(x_1) = int_{a}^{x_1 + Delta x} f(t) ,dt - int_{a}^{x_1} f(t) ,dt. qquad (1)

It can be shown that:int_{a}^{x_1} f(t) ,dt + int_{x_1}^{x_1 + Delta x} f(t) ,dt = int_{a}^{x_1 + Delta x} f(t) ,dt. :(The sum of the areas of two adjacent regions is equal to the area of both regions combined.)Manipulating this equation gives:int_{a}^{x_1 + Delta x} f(t) ,dt - int_{a}^{x_1} f(t) ,dt = int_{x_1}^{x_1 + Delta x} f(t) ,dt.

Substituting the above into (1) results in:F(x_1 + Delta x) - F(x_1) = int_{x_1}^{x_1 + Delta x} f(t) ,dt. qquad (2)

According to the Intermediate value theorem for integration, there exists a "c" in ["x"1, "x"1 + Δ"x"] such that:int_{x_1}^{x_1 + Delta x} f(t) ,dt = f(c) Delta x ,.

Substituting the above into (2) we get:F(x_1 + Delta x) - F(x_1) = f(c) Delta x ,.

Dividing both sides by Δ"x" gives:frac{F(x_1 + Delta x) - F(x_1)}{Delta x} = f(c). :Notice that the expression on the left side of the equation is Newton's difference quotient for "F" at "x"1.

Take the limit as Δ"x" → 0 on both sides of the equation.:lim_{Delta x o 0} frac{F(x_1 + Delta x) - F(x_1)}{Delta x} = lim_{Delta x o 0} f(c).

The expression on the left side of the equation is the definition of the derivative of "F" at "x"1.:F'(x_1) = lim_{Delta x o 0} f(c). qquad (3)

To find the other limit, we will use the squeeze theorem. The number "c" is in the interval ["x"1, "x"1 + Δ"x"] , so "x"1 ≤ "c" ≤ "x"1 + Δ"x".

Also, lim_{Delta x o 0} x_1 = x_1 and lim_{Delta x o 0} x_1 + Delta x = x_1,.

Therefore, according to the squeeze theorem,:lim_{Delta x o 0} c = x_1,.

Substituting into (3), we get:F'(x_1) = lim_{c o x_1} f(c),.

The function "f" is continuous at "c", so the limit can be taken inside the function. Therefore, we get:F'(x_1) = f(x_1) ,.which completes the proof.

(Leithold et al, 1996)

Proof of the Second Part

This is a limit proof by Riemann sums.

Let "f" be continuous on the interval ["a", "b"] , and let "F" be an antiderivative of "f". Begin with the quantity:F(b) - F(a),.

Let there be numbers

:"x"1, ..., "x""n"

such that

:a = x_0 < x_1 < x_2 < ldots < x_{n-1} < x_n = b,.

It follows that:F(b) - F(a) = F(x_n) - F(x_0) ,.

Now, we add each "F"("x""i") along with its additive inverse, so that the resulting quantity is equal::egin{matrix} F(b) - F(a) & = & F(x_n),+, [-F(x_{n-1}),+,F(x_{n-1})] ,+,ldots,+, [-F(x_1) + F(x_1)] ,-,F(x_0) , \& = & [F(x_n),-,F(x_{n-1})] ,+, [F(x_{n-1}),+,ldots,-,F(x_1)] ,+, [F(x_1),-,F(x_0)] ,. end{matrix}

The above quantity can be written as the following sum::F(b) - F(a) = sum_{i=1}^n , [F(x_i) - F(x_{i-1})] ,. qquad (1)

Next we will employ the mean value theorem. Stated briefly,

Let "F" be continuous on the closed interval ["a", "b"] and differentiable on the open interval ("a", "b"). Then there exists some "c" in ("a", "b") such that:F'(c) = frac{F(b) - F(a)}{b - a},.

It follows that:F'(c)(b - a) = F(b) - F(a). ,

The function "F" is differentiable on the interval ["a", "b"] ; therefore, it is also differentiable and continuous on each interval "x""i"-1. Therefore, according to the mean value theorem (above),:F(x_i) - F(x_{i-1}) = F'(c_i)(x_i - x_{i-1}) ,.

Substituting the above into (1), we get:F(b) - F(a) = sum_{i=1}^n , [F'(c_i)(x_i - x_{i-1})] ,.

The assumption implies F'(c_i) = f(c_i). Also, x_i - x_{i-1} can be expressed as Delta x of partition i.

:F(b) - F(a) = sum_{i=1}^n , [f(c_i)(Delta x_i)] ,. qquad (2)

Notice that we are describing the area of a rectangle, with the width times the height, and we are adding the areas together. Each rectangle, by virtue of the Mean Value Theorem, describes an approximation of the curve section it is drawn over. Also notice that Delta x_i does not need to be the same for any value of i, or in other words that the width of the rectangles can differ. What we have to do is approximate the curve with n rectangles. Now, as the size of the partitions get smaller and n increases, resulting in more partitions to cover the space, we will get closer and closer to the actual area of the curve.

By taking the limit of the expression as the norm of the partitions approaches zero, we arrive at the Riemann integral. That is, we take the limit as the largest of the partitions approaches zero in size, so that all other partitions are smaller and the number of partitions approaches infinity.

So, we take the limit on both sides of (2). This gives us:lim_{| Delta | o 0} F(b) - F(a) = lim_{| Delta | o 0} sum_{i=1}^n , [f(c_i)(Delta x_i)] ,.

Neither "F"("b") nor "F"("a") is dependent on ||Δ||, so the limit on the left side remains "F"("b") - "F"("a").:F(b) - F(a) = lim_{| Delta | o 0} sum_{i=1}^n , [f(c_i)(Delta x_i)] ,.

The expression on the right side of the equation defines an integral over "f" from "a" to "b". Therefore, we obtain:F(b) - F(a) = int_{a}^{b} f(x),dx,,which completes the proof.


We don't need to assume continuity of "f" on the whole interval. Part I of the theorem then says: if "f" is any Lebesgue integrable function on ["a", "b"] and "x"0 is a number in ["a", "b"] such that "f" is continuous at "x"0, then

:F(x) = int_a^x f(t), dt

is differentiable for "x" = "x"0 with "F'"("x"0) = "f"("x"0). We can relax the conditions on "f" still further and suppose that it is merely locally integrable. In that case, we can conclude that the function "F" is differentiable almost everywhere and "F'"("x") = "f"("x") almost everywhere. This is sometimes known as Lebesgue's differentiation theorem.

Part II of the theorem is true for any Lebesgue integrable function "f" which has an antiderivative "F" (not all integrable functions do, though).

The version of Taylor's theorem which expresses the error term as an integral can be seen as a generalization of the Fundamental Theorem.

There is a version of the theorem for complex functions: suppose "U" is an open set in C and "f": "U" → C is a function which has a holomorphic antiderivative "F" on "U". Then for every curve γ: ["a", "b"] → "U", the curve integral can be computed as

:int_{gamma} f(z) ,dz = F(gamma(b)) - F(gamma(a)),.

The fundamental theorem can be generalized to curve and surface integrals in higher dimensions and on manifolds.

One of the most powerful statements in this direction is Stokes' theorem: Let "M" be an oriented piecewise smooth manifold of dimension "n" and let omega be an "n"−1 form that is a compactly supported differential form on "M" of class C1. If ∂"M" denotes the boundary of "M" with its induced orientation, then

:int_M mathrm{d}omega = oint_{partial M} omega,.

Here mathrm{d}!, is the exterior derivative, which is defined using the manifold structure only.

The theorem is often used in situations where "M" is an embedded oriented submanifold of some bigger manifold on which the form omega is defined.

ee also

* Differintegral



* Larson, Ron, Bruce H. Edwards, David E. Heyd. "Calculus of a single variable". 7th ed. Boston: Houghton Mifflin Company, 2002.
* Leithold, L. (1996). "The calculus 7 of a single variable". 6th ed. New York: HarperCollins College Publishers.
* Malet, A, "Studies on James Gregorie (1638-1675)" (PhD Thesis, Princeton, 1989).
* Stewart, J. (2003). Fundamental Theorem of Calculus. In Integrals. In "Calculus: early transcendentals". Belmont, California: Thomson/Brooks/Cole.
* Turnbull, H. W. (ed.), "The James Gregory Tercentenary Memorial Volume" (London, 1939)

External links

* [ James Gregory's Euclidean Proof of the Fundamental Theorem of Calculus] at [ Convergence]
* [ Isaac Barrow's proof of the Fundamental Theorem of Calculus]

Wikimedia Foundation. 2010.

Look at other dictionaries:

  • fundamental theorem of calculus — Basic principle of calculus. It relates the derivative to the integral and provides the principal method for evaluating definite integrals (see differential calculus; integral calculus). In brief, it states that any function that is continuous… …   Universalium

  • Fundamental theorem — In mathematics, there are a number of fundamental theorems for different fields. The names are mostly traditional; so that for example the fundamental theorem of arithmetic is basic to what would now be called number theory. Theorems may be… …   Wikipedia

  • Fundamental theorem of arithmetic — In number theory, the fundamental theorem of arithmetic (or unique prime factorization theorem) states that every natural number greater than 1 can be written as a unique product of prime numbers. For instance, : 6936 = 2^3 imes 3 imes 17^2 , ,! …   Wikipedia

  • Calculus — This article is about the branch of mathematics. For other uses, see Calculus (disambiguation). Topics in Calculus Fundamental theorem Limits of functions Continuity Mean value theorem Differential calculus  Derivative Change of variables …   Wikipedia

  • calculus — /kal kyeuh leuhs/, n., pl. calculi / luy /, calculuses. 1. Math. a method of calculation, esp. one of several highly systematic methods of treating problems by a special system of algebraic notations, as differential or integral calculus. 2.… …   Universalium

  • theorem — theorematic /thee euhr euh mat ik, thear euh /, adj. theorematically, adv. /thee euhr euhm, thear euhm/, n. 1. Math. a theoretical proposition, statement, or formula embodying something to be proved from other propositions or formulas. 2. a rule… …   Universalium

  • fundamental — fundamentality, fundamentalness, n. fundamentally, adv. /fun deuh men tl/, adj. 1. serving as, or being an essential part of, a foundation or basis; basic; underlying: fundamental principles; the fundamental structure. 2. of, pertaining to, or… …   Universalium

  • Calculus of variations — is a field of mathematics that deals with extremizing functionals, as opposed to ordinary calculus which deals with functions. A functional is usually a mapping from a set of functions to the real numbers. Functionals are often formed as definite …   Wikipedia

  • Calculus with polynomials — Topics in Calculus Fundamental theorem Limits of functions Continuity Mean value theorem Differential calculus  Derivative Change of variables Implicit differentiation Taylor s theorem Related rates …   Wikipedia

  • Theorem — The Pythagorean theorem has at least 370 known proofs[1] In mathematics, a theorem is a statement that has been proven on the basis of previously established statements, such as other theorems, and previously accepted statements …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”

We are using cookies for the best presentation of our site. Continuing to use this site, you agree with this.