Title: ENGINEERING OPTIMIZATION
1ENGINEERING OPTIMIZATION Methods and Applications
A. Ravindran, K. M. Ragsdell, G. V. Reklaitis
Book Review
2Chapter 9 Direction Generation Methods Based on
Linearization
Part 1 Ferhat Dikbiyik Part 2Mohammad F. Habib
Review Session July 30, 2010
3The Linearization-based algorithms in Ch. 8
- LP solution techniques to specify the sequence of
intermediate solution points.
The linearized subproblem at this point is updated
The exact location of next iterate is determined
by LP
The linearized subproblem cannot be expected to
give a very good estimate of either boundaries of
the feasible solution region or the contours of
the objective function
4Good Direction Search
- Rather than relying on the admittedly inaccurate
linearization to define the precise location of a
point, it is more realistic to utilize the linear
approximations only to determine a locally good
direction for search.
5Outline
- 9.1 Method of Feasible Directions
- 9.2 Simplex Extensions for Linearly Constrained
Problems - 9.3 Generalized Reduced Gradient Method
- 9.4 Design Application
69.1 Method of Feasible Directions
G. Zoutendijk Mehtods of Feasible Directions,
Elsevier, Amsterdam, 1960
7Preliminaries
Source Dr. Muhammad Al-Slamah, Industrial
Engineering, KFUPM
8Preliminaries
Source Dr. Muhammad Al-Slamah, Industrial
Engineering, KFUPM
99.1 Method of Feasible Directions
- Suppose that is a starting point that
satisfies all constraints. - and suppose that a certain subset of these
constraints are binding at .
109.1 Method of Feasible Directions
- Suppose is a feasible point
- Define x as
- The first order Taylor approximation of f(x) is
given by - In order for , we have to have
- A direction satisfying this relationship is
called a descent direction
119.1 Method of Feasible Directions
- This relationship dictates the angle between d
and to be greater than 90 and
less than 270 .
Source Dr. Muhammad Al-Slamah, Industrial
Engineering, KFUPM
129.1 Method of Feasible Directions
- The first order Taylor approximation for
constraints - And with assumption
- (because its binding)
- In order for x to be a feasible,
hence - Any direction d satisfying this relationship
called a feasible direction
139.1 Method of Feasible Directions
- This relationship dictates the angle between d
and has to be between 0 and than
90 .
Source Dr. Muhammad Al-Slamah, Industrial
Engineering, KFUPM
149.1 Method of Feasible Directions
- In order for x to solve the inequality
constrained problem, the direction d has to be
both a descent and feasible solution.
Source Dr. Muhammad Al-Slamah, Industrial
Engineering, KFUPM
159.1 Method of Feasible Directions
- Zoutendijks basic idea is at each stage of
iteration to determine a vector d that will be
both a feasible direction and a descent
direction. This is accomplished numerically by
finding a normalized direction vector d and a
scalar parameter ? gt 0 such that
and ? is as large as possible.
169.1 Method of Feasible Directions
Source Dr. Muhammad Al-Slamah, Industrial
Engineering, KFUPM
179.1.1 Basic Algorithm
- The active constraint set is defined as
- for some small
189.1.1 Basic Algorithm
- Step 1. Solve the linear programming problem
-
- Label the solution and
199.1.1 Basic Algorithm
- Step 2. If the iteration
terminates, since no further improvement is
possible. Otherwise, determine - If no exists, set
209.1.1 Basic Algorithm
- Step 3. Find such that
- Set and continue.
21Example 9.1
, since g_1 is the only binding constraint.
22Example 9.1
We must search along the ray
to find the point at which boundary of feasible
region is intersected
23Example 9.1
Since
is positive for all a 0 and is not violated as
a is increased. To determine the point at which
will be intersected, we solve
Finally, we search on a over range
to determine the optimum of
24Example 9.1
259.1.2 Active Constraint Sets and Jamming
269.1.2 Active Constraint Sets and Jamming
- The active constraint set used in the basic form
of feasible direction algorithm, namely, - cannot only slow down the process of iterations
but also lead to convergence to points that are
not Kuhn-Tucker points. - This type of false convergence is known as
jamming
279.1.2.1 e-Perturbation Method
- At iteration point and with given
, define and carry out step 1 of the basic
algorithm. - Modify step 2 with the following If
, - set and continue. However, if
, set and proceed
with line search of the basic method. If
, then a Kuhn-Tucker point has been found.
With this modification, it is efficient to set e
rather loosely initially so as to include the
constraints in a larger neighborhood of the point
. Then, as the iterations proceed, the size
of the neighborhood will be reduced only when it
is found to be necessary.
289.1.2.2 Topkis-Veinott Variant
- This approach simply dispense with the active
constraint concept altogether and redefine the
direction-finding subproblem as follows
If the constraint loose at , then the
selection of d is less affected by constraint j,
because the positive constraint value will
counterbalance the effect of the gradient term.
This ensures that no sudden changes are
introduced in the search direction.
299.2 Simplex Extensions for Linearly Constrained
Problems
- At a given point, the number of directions that
are both descent and feasible directions is
generally infinite. - In the case of linear programs, the generation of
search directions was simplified by changing one
variable at a time feasibility was ensured by
checking sign restrictions, and descent was
ensured by selecting a variable with negative
relative-cost coefficient.
309.2.1 Convex Simplex Method
N components
M rows
- Given a feasible point the x variable
is partitioned into two sets - the basic variables , which are all positive
- the nonbasic variables , which are all zero
an M vector
an N-M vector
319.2.1 Convex Simplex Method
329.2.1 Convex Simplex Method
The relative-cost coefficients
The nonbasic variable to enter is selected by
finding such that
The basic variable to leave the basis is
selected using the minimum-ratio rule. That is,
we find r such that
elements of matrix
339.2.1 Convex Simplex Method
- The new feasible solution
- and all other variables zero. At this point, the
variables and are relabeled. Since an
exchange will have occurred, - will be redefined. The matrix is
recomputed and another cycle of iterations is
begun.
349.2.1 Convex Simplex Method
- The application of same algorithm to linearized
form of a non-linear objective function - The relative-cost factor
35Example 9.4
36Example 9.4
37Example 9.4
The relative-cost factor
38Example 9.4
The nonbasic variable to enter will be
, since
The basic variable to leave will be
, since
39Example 9.4
- The new point is thus
- A line search between and is now
required to locate minimum of . Note that
remains at 0, while changes as given by
40Example 9.4
41Convex Simplex Algorithm
42Convex Simplex Algorithm
439.2.2 Reduced Gradient Method
- The nonbasic variable direction vector
- This definition ensures that when
for all i, the Kuhn-Tucker conditions are
satisfied.
449.2.2 Reduced Gradient Method
- In the first case, the limiting a value will be
given by - If all , then set
- In the second case,
If all , then set
45Reduced Gradient Algorithm
46(No Transcript)