5.2. Finding the Minimum of a Function of Several Variables#
Last revised on November 20, 2025.
References:
[Sullivan, 2021] Section 3.4.2, Multivariable Optimization.
[Sauer, 2022] Chapter 13, Optimization, Sub-sections 13.2.2, Steepest Descent, and 13.1.3, Nelder-Mead.
[Chenney and Kincaid, 2013] Section 13.2, Multivariate Case.
[Kincaid and Chenney, 1990] Sections 11.2 (Descent Methods) and 11.5 (Nelder-Meade Algorithm).
5.2.1. Introduction#
This section will eventually describe several methods for computing the minimum (and its location) of a function \(f(x, y, \dots)\) of several variables:
Applying Newton’s method to finding critical points by solving \(\nabla f = \mathbf 0\).
Steepest Descent (a.k.a. Gradient Search) where the gradient is used at each iteration to find the direction in which to search for a new approximate location where \(f\) has a lower value.
The method of Nelder and Meade, which avoids the use of derivatives.
For now, just the first two are introduced.
5.2.2. Newton’s Method#
A (local) minimum of a function \(\mathbf f(\mathbf x)\) of \(n\) variables occurs at a zero of its gradient: \(\nabla \mathbf f(\mathbf x) = \mathbf{0}\), so one stratgy is to solve the latter equation by Newton’s method.
For a function of one variable, this leads to the iterative form
To extend to higher dimensions, we need to see what replaces that second derivative.
Firstly, eliminate the division and solve for the increment \(\delta x^{k} := x^{k+1} = x^{k}\):
Next, for the system form, each component of the gradient must vanish at a critical point, giving the \(n\) equations
The tangent plane approximation of a function\(f:\mathbb{R}^n \to \mathbb{R}\) near a point \(\textbf x^{(k)}\) is
so applying to each component of the above system (5.1) gives
The term with the second partial derivatives is the \(i\)-th component of a matrix multiplication, with the (symmetric) matrix-valued function being the Hessian of \(f\), with components \( \displaystyle H_{ij}(\mathbf x) = \frac{\partial^2 f}{\partial x_i \partial x_j}(\mathbf x) \)
With this notation, the above collection of approximations is
With this, the linear approximation as used in Newton’s method comes from setting this to zero for \(\mathbf x = \mathbf {x^{(k+1)}}\):
In other words,
5.2.3. Steepest Descent (a.k.a. Gradient Search)#
An alternative that at least avoids the need for second derivatives and which can build on the more robust minimization methods seen in Section Finding the Minimum of a Function of One Variable Without Using Derivatives, is the strategy of Steepest Descent.
The two key ideas are that:
A function \(f:\mathbb{R}^n \to \mathbb{R}\) decreases most rapidy from a point \(\mathbf{x}\) in the direction opposite its gradient: \(-\nabla f(\mathbf x)\).
We can search for the lowest value in that “downhill direction” by using a method for minimizing a function of one variable.
The basic strategy is that, starting at an approximation \(\mathbf{x}^{(k)}\), one looks along the “downhill ray” \(\mathbf x = \mathbf{x}^{(k)} - t \nabla f(\mathbf{x}^{(k)} )\), \(t \geq 0\), to find the value \(t_k\) of \(t\) which minimizes the auxilliary function
giving the improved approximation
This minimization can be done by one of the methods in the section on Finding the Minimum of a Function of One Variable, such as the Golden Section Search.
Some details need to be addressed, like the fact that in principle one has to search the infinite domain \(t \in [0, \infty)\) rather than a closed interval like \(t \in [0,T]\).
Thus some experimentation is needed.
5.2.4. Exercises#
Exercise 5.6
Implement the method of Steepest Descent (using your choice of one-dimensional minimzation algorithm), and test it initially on an example like
which has an obvious, unique minimum.
Exercise 5.7
Test the above on some more challenging examples like
with minimum at \(x=y=1\).
Start with \(K = 1\); the try with a larger value like \(K = 100\), which makes it harder — a graph might reveal why.
Exercise 5.8
It can also be interesting to explore a function that has multiple local minima, like