
(3) The determinant of each leading principal minor is positive. A principal minor of
order j is a minor obtained by deleting from an n
×
n square matrix, n – j pairs of rows
and columns that intersect on the main diagonal (see the definition of a minor in
Section 2.2.6). In other words, the index numbers of the rows deleted are the same as
the index numbers of the columns deleted. A leading principal minor of order j is a
principal minor of order j that contains the first j rows an d j columns of the n × n
square matrix. A 3 × 3 matrix has three principal minors of order 2:
a
11
a
12
a
21
a
22
;
a
11
a
13
a
31
a
33
;
a
22
a
23
a
32
a
33
:
Of these, the first principal minor is the leading principal minor of order 2 since it
contains the first two rows and columns of the matrix. An n × n square matrix has n
leading principal minors. Can you list the three leading principal minors of a 3 × 3
matrix?
To ascertain the positive definiteness of a matrix, one needs to show that either the
first property in the list above is true or that both properties (2) and (3) hold true
simultaneously.
A critical point x* is a local maximum if Hx
ðÞis negative definite. A matrix A is
said to be negative definite if A is symmetric and x
T
Ax
5
0 for all non-zero vectors x.
A negative definite matrix has negative eigenvalues, diagonal elements that are all
less than zero, and leading principal minors that all have a negative determinant. In
contrast, at a saddle point, the Hessian matrix is neither positive definite nor
negative definite.
A variety of search methods for unconstrained multidimensional optimization
problems have been developed. The general procedure of most multivariable opti-
mization methods is to (1) determine a search direction and (2) find the minimum
value of the function along the line that passes through the current point and runs
parallel to the search direction. Derivative-based methods for multivariable mini-
mization are called indirect search methods because the analytical form (or numerical
form) of the partial derivatives of the function must be determined and evaluated at
the initial guessed values before the numerical step can be applied. In this section we
present two indirect methods of optimization: the steepest descent method
(Section 8.3.1) and Newton’s method for multidimensional search (Section 8.3.2).
The latter algorithm requires calculation of the first and second partial derivatives of
the objective function at each step, while the former requires evaluation of only the
first partial derivatives of the function at the approximation x
i
to provide a new
search direction at each step. In Section 8.3.3, the simplex method based on the
Nelder–Mead algorithm is discussed. This is a direct search method that evaluates
only the function to proceed towards the minimum.
8.3.1 Steepest descent or gradient method
In the steepest descent method, the negative of the gradient vector evaluated at x
provides the search direction at x. The negative of the gradient vector points in the
directionofthelargestrateofdecreaseoffðxÞ. If the first guessed point is x
ð0Þ
,thenext
point x
ð1Þ
lies on the line x
ð0Þ
αrf x
ð0Þ
,whereα is the step size taken in the search
direction from x
ð0Þ
. A minimization search is carried out to locate a point on this line
where the function has the least value. This type of one-dimensional search is called a
line search and involves the optimization of a single variable α. The next point is given by
502
Nonlinear model regression and optimization