Local Extrema of Functions
Definition of Local Maximum and Local Minimum
Let a function y = f (x) be defined in a δ-neighborhood of a point x0, where δ > 0. The function f (x) is said to have a local (or relative) maximum at the point x0, if for all points x ≠ x0 belonging to the neighborhood (x0 − δ, x0 + δ) the following inequality holds:
If the strict inequality holds for all points x ≠ x0 in some neighborhood of x0:
then the point x0 is a strict local maximum point.
Similarly, we define a local (or relative) minimum of the function \(f\left( x \right).\) In this case, the following inequality is valid for all points \(x \ne {x_0}\) of the \(\delta\)-neighborhood \(\left( {{x_0} - \delta ,{x_0} + \delta } \right)\) of the point \({x_0}:\)
Accordingly, a strict local minimum is described by the inequality
The concepts of local maximum and local minimum are united under the general term local extremum. The word "local" is often ommitted for brevity, so it is said simply about maxima and minima of functions.
Figure \(1\) schematically shows the different extrema points. The point \(A\left( {{x_1}} \right)\) is a strict local minimum point, since there exists a \(\delta\)-neighborhood \(\left( {{x_1} - \delta ,{x_1} + \delta } \right),\) in which the following inequality holds:
Similarly, the point \(B\left( {{x_2}} \right)\) is a strict local maximum point. At this point, we have the inequality
(Of course, the number \(\delta\) at each point may be different.)
The subsequent points are classified as follows:
- \(C\left( {{x_3}} \right)\) is a strict minimum point;
- \(D\left( {{x_4}} \right)\) is a non-strict maximum point;
- \(E\left( {{x_5}} \right)\) is a non-strict maximum or minimum point;
- \(F\left( {{x_6}} \right)\) is a non-strict maximum point;
- \(G\left( {{x_7}} \right)\) is a non-strict minimum point;
- \(H\left( {{x_8}} \right)\) is a non-strict maximum or minimum point;
- \(I\left( {{x_9}} \right)\) is a non-strict maximum point;
- \(J\left( {{x_{10}}} \right)\) − there is no extremum.
Necessary Condition for an Extremum
We introduce some more concepts.
The points at which the derivative of the function \(f\left( x \right)\) is equal to zero are called the stationary points.
The points at which the derivative of the function \(f\left( x \right)\) is equal to zero or does not exist are called the critical points of the function. Consequently, the stationary points are a subset of the set of critical points.
A necessary condition for an extremum is formulated as follows:
If the point \({x_0}\) is an extremum point of the function \(f\left( x \right),\) then the derivative at this point either is zero or does not exist. In other words, the extrema of a function are contained among its critical points.
The proof of the necessary condition follows from Fermat's theorem.
Note that the necessary condition does not guarantee the existence of an extremum. A classic illustration here is the cubic function \(f\left( x \right) = {x^3}.\) Despite the fact that the derivative of the function at the point \(x = 0\) is zero: \(f'\left( {x = 0} \right) = 0,\) this point is not an extremum.
Local extrema of differentiable functions exist when the sufficient conditions are satisfied. These conditions are based on the use of the first-, second-, or higher-order derivative. Respectively, \(3\) sufficient conditions for local extrema are considered. Now we turn to their formulation and proof.
First Derivative Test
Let the function \(f\left( x \right)\) be differentiable in a neighborhood of the point \({x_0},\) except perhaps at the point \({x_0}\) itself, in which, however, the function is continuous. Then:
- If the derivative \(f'\left( x \right)\) changes sign from minus to plus when passing through the point \({x_0}\) (from left to right), then \({x_0}\) is a strict minimum point (Figure \(2\)). In other words, in this case there exists a number \(\delta \gt 0\) such that
\[\forall \;x \in \left( {{x_0} - \delta ,{x_0}} \right) \Rightarrow f'\left( x \right) \lt 0,\]\[\forall \;x \in \left( {{x_0}, {x_0} + \delta} \right) \Rightarrow f'\left( x \right) \gt 0.\]
- If the derivative \(f'\left( x \right),\) on the contrary, changes sign from plus to minus when passing through the point \({x_0},\) then \({x_0}\) is a strict maximum point (Figure \(3\)). In other words, there exists a number \(\delta \gt 0\) such that
\[\forall \;x \in \left( {{x_0} - \delta ,{x_0}} \right) \Rightarrow f'\left( x \right) \gt 0,\]\[\forall \;x \in \left( {{x_0}, {x_0} + \delta} \right) \Rightarrow f'\left( x \right) \lt 0.\]
Proof.
We confine ourselves to the case of the minimum. Suppose that the derivative \(f'\left( x \right)\) changes sign from minus to plus when passing through the point \({x_0}.\) To the left from the point \({x_0},\) the following condition is satisfied:
By Lagrange's theorem, the difference of the values of the function at the points \(x\) and \({x_0}\) is written as
where the point \(c\) belongs to the interval \(\left( {{x_0} - \delta ,{x_0}} \right),\) in which the derivative is negative, i.e. \(f'\left( c \right) \lt 0.\) Since \(x - {x_0} \lt 0\) to the left of the point \({x_0},\) then
Likewise, it is established that
(to the right of the point \({x_0}\)).
Based on the definition, we conclude that \({x_0}\) is a strict minimum point of the function \(f\left( x \right).\)
Similarly, we can prove the first derivative test for a strict maximum.
Note that the first derivative test does not require the function to be differentiable at the point \({x_0}.\) If the derivative at this point is infinite or does not exist (i.e. the point \({x_0}\) is critical, but not stationary), the first derivative test can still be used to investigate the local extrema of the function.
Second Derivative Test
Let the first derivative of a function \(f\left( x \right)\) at the point\({x_0}\) be equal to zero: \(f\left( {{x_0}} \right) = 0,\) that is \({x_0}\) is a stationary point of \(f\left( x \right).\) Suppose also that there exists the second derivative \(f^{\prime\prime}\left( {{x_0}} \right)\) at this point. Then
- If \(f^{\prime\prime}\left( {{x_0}} \right) \gt 0,\) then \({x_0}\) is a strict minimum point of the function \(f\left( x \right)\);
- If \(f^{\prime\prime}\left( {{x_0}} \right) \lt 0,\) then \({x_0}\) is a strict maximum point of the function \(f\left( x \right).\)
Proof.
In the case of a strict minimum \(f^{\prime\prime}\left( {{x_0}} \right) \gt 0.\) Then the first derivative is an increasing function at the point \({x_0}.\) Consequently, there exists a number \(\delta \gt 0\) such that
Since \(f^{\prime\prime}\left( {{x_0}} \right) = 0\) (because \({x_0}\) is a stationary point), therefore the first derivative is negative in the \(\delta\)-neighborhood to the left of the point \({x_0}\), and is positive to the right, i.e. the derivative changes sign from minus to plus when passing through the point \({x_0}.\) By the first derivative test, this means that \({x_0}\) is a strict minimum point.
The case of the maximum can be considered in a similar way.
The second derivative test is convenient to use when calculation of the first derivatives in the neighborhood of a stationary point is difficult. On the other hand, the second test may be used only for stationary points (where the first derivative is zero) − in contrast to the first derivative test, which is applicable to any critical points.
Third Derivative Test
Let the function \(f\left( x \right)\) have derivatives at the point \({x_0}\) up to the \(n\)th order inclusively. Then if
the point \({x_0}\) for even \(n\) is
- a strict minimum point if \({f^{\left( n \right)}}\left( {{x_0}} \right) \gt 0,\) and
- a strict maximum point if \({f^{\left( n \right)}}\left( {{x_0}} \right) \gt 0.\)
For odd \(n,\) the extremum at \({x_0}\) does not exist.
It is clear that for \(n = 2\), we obtain as a special case the second derivative test for local extrema considered above. To avoid such a transition, the third derivative test implies that \(n \gt 2.\)
Proof.
Expand the function \(f\left( x \right)\) at the point \({x_0}\) in a Taylor series:
Since, by assumption, all of the first derivatives up to the \(\left( {n - 1} \right)\)th order are equal to zero, we obtain:
where the remainder term \(\omicron\left( {{{\left( {x - {x_0}} \right)}^n}} \right)\) has a higher order of smallness than \(n.\) As a result, the sign of the difference \(f\left( x \right) - f\left( {{x_0}} \right)\) in the \(\delta\)-neighborhood of the point \({x_0}\) will be determined by the sign of the \(n\)th term in the Taylor series:
or
If \(n\) is an even number (\(n = 2k\)), then
Consequently, in this case
If \({f^{\left( n \right)}}\left( {{x_0}} \right) \gt 0\) in the \(\delta\)-neighborhood of the point \({x_0}\), then the following inequality holds:
By definition, this means that \({x_0}\) is a strict minimum point of the function \(f\left( x \right).\)
Similarly, if \({f^{\left( n \right)}}\left( {{x_0}} \right) \lt 0\) in the \(\delta\)-neighborhood of the point \({x_0}\), we have the inequality
that corresponds to a strict maximum point.
If \(n\) is an odd number \(\left(n = 2k + 1\right),\) the degree of \({\left( {x - {x_0}} \right)^{2k + 1}}\) will change sign when passing through the point \({x_0}.\) Then it follows from the formula
that the difference \({f\left( x \right) - f\left( {{x_0}} \right)}\) also changes sign when passing through \({x_0}.\) In this case, the extremum at the point \({x_0}\) does not exist.