
5.2 Regularized Decomposition 209
for some value of
Δ
ν
(Exercise 4), which then suggests the general form of a trust-
region method (see, e.g., Conn, Gould, and Toint [2000]). The norm as well as the
centering point can also be varied in this approach. Linderoth and Wright [2003] use
the ∞ -norm (maximum component deviation) to obtain a trust region algorithm for
stochastic programs that also allows for significant parallelization and can achieve
substantial computational efficiency.
Exercises
1. Check that, with the same starting point, both the L -shaped and the multicut
methods require five iterations in Example 1.
2. The regularized decomposition only makes sense with a reasonable starting
point. To illustrate this, consider the same example taking as starting point a
highly negative value, e.g., a
1
= −20 . At Iteration 1, the cuts
θ
1
≥−
x−1
2
and
θ
2
≥−
3
4
x are created. Observe that, for many subsequent iterations, no new
cuts are generated as the sequence of trial points a
ν
move from −20 to −
75
4
,
then −
70
4
, −
65
4
, .. .each time by a change of
5
4
, until reaching 0 , where new
cuts will be generated. Thus a long sequence of approximate serious steps is
taken.
3. As we mentioned in the introduction of this section, the regularized decom-
position algorithm works with a more general regularizing term of the form
α
2
x −a
ν
2
.
(a) Observe that the proof of convergence relies on strict convexity of the
objective function (Lemma 5), thus
α
> 0 is needed. It also relies on
∇
α
2
x
ν
−a
ν
2
→ 0asx
ν
→ a
ν
, which is simply obtained by taking a
finite
α
. The algorithm can thus be tuned for any positive
α
and
α
can
vary within the algorithm.
(b) Taking the same starting point and data as in Exercise 2, show that by se-
lecting different values of
α
, any point in ] −20,20] can be obtained as a
solution of the regularized master at the second iteration (where 20 is the
upper bound on x and the first iteration only consists of adding cuts on
θ
1
and
θ
2
).
(c) Again taking the same starting point and data as in Exercise 2, how would
you take
α
to reduce the number of iterations? Discuss some alternatives.
(d) Let
α
= 1 for Iterations 1 and 2. As of Iteration 2, consider the following
rule for changing
α
dynamically. For each null step,
α
is doubled. At each
exact step,
α
is halved. Show why this would improve the performance
of the regularized decomposition in the case of Exercise 2. Consider the
starting point x
1
= −0.5 as in Example 1 and observe that the same path
as before is followed.
4. Show the equivalence of (2.1)and(2.11).