Title: Linear
1Linear Nonlinear Programming --Basic Properties
of Solutions and Algorithms
2Outline
- First-order Necessary Condition
- Examples of Unconstrained Problems
- Second-order Conditions
- Convex and Concave Functions
- Minimization and Maximization of Convex Functions
- Global Convergence of Descent Algorithms
- Speed of Convergence
3Introduction
- Considering optimization problem of the form
-
- where f is a real-valued function and ?, the
feasible set, is a subset of En.
4Weierstras Theorem
- If f is continuous and ? is compact, a solution
point of minimization problem exists.
5Two Kinds of Solution Points
- Definition of a relative minimum point
- A point is said to be a relative minimum point
of f over ? if there is an such that
f(x) f(x) for all within
a distance of x. If f(x) gt f(x) for all
, x?x, within a distance of x,
then x is said to be a strict relative minimum
point of f over . - Definition of a global minimum point
- A point is said to be a global minimum point of
f over ? if f(x) f(x) for all . If
f(x) gt f(x) for all , x?x, then x
is said to be a strict global minimum point of f
over .
6Two Kinds of Solution Points (contd)
- We can achieve relative minimum by using
differential calculus or a convergent stepwise
procedure. - Global conditions and global solutions can, as a
rule, only be found if the problem possesses
certain convexity properties that essentially
guarantee that any relative minimum is a global
minimum.
7Feasible Directions
- To derive necessary conditions satisfied by a
relative minimum point x, the basic idea is to
consider movement away from the point in some
given direction. - A vector d is a feasible direction at x if there
is an such that for all a,
.
8Feasible Directions (contd)
- Proposition 1 ( first-order necessary conditions)
- Let ? be a subset of En and let be a
function on ?. If x is a relative minimum point
of f over ?, then for any that is a
feasible direction at x, we have - Proof
-
9Feasible Directions (contd)
- Corollary ( unconstrained case )
- Let ? be a subset of En and let be a
function on ?. If x is a relative minimum point
of f over ? and x is an interior point of ?,
then . - Since in this case d can be any direction from
x, and hence - for all . This
implies .
10Example 1 about Feasible Directions
- Example 1 ( unconstrained )
- minimize
- There are no constrains, so ? En
- These have the unique solution x11, x22, so it
is a global minimum of f .
11Example 2 about Feasible Directions
- Example 2 (a constrained case)
- minimize
- subject to x1 0, x2 0.
- since we know that there is a global minimum at
x11/2, x20, then
12Example 2 (contd)
x2
3/2
1
Feasible region
1/2
d (d1 , d2)
x1
0
1
(1/2, 0)
3/2
2
13Example 3 of Unconstrained Problems
- The problem faced by an electric utility when
selecting its power-generating facilities. - Its power-generating requirements are summarized
by a curve, h(x), as shown in Fig.6.2(a), which
shows the total hours in a year that a power
level of at least x is required for each x. - For convenience the curve is normalized so that
the upper limit is unity. - The power company may meet these requirements by
ins-talling generating equipment, such as (1)
nuclear or (2) coal-fired, or by purchasing power
from a central energy. -
14Example 3 (contd)
- Associated with type i ( i 1,2 ) of generating
equipment is a yearly unit capital cost bi and a
unit operating cost ci. - The unit price of power purchased from the grid
is c3. - The requirements are satisfied as shown in
Fig.6.2(b), where x1 and x2 denote the capacities
of the nuclear and coal-fired plants,
respectively.
15Example 3 (contd)
hours required h(x)
coal
nuclear
purchase
1
x1
x2
power (megawatts)
(b)
Fig.6.2 Power requirement curve
16Example 3 (contd)
- The total cost is
- And the company wishes to minimize the set
defined by - x1 0 , x2 0 ,
x1x2 1
17Example 3 (contd)
- Assume that the solution is interior to the
constraints, by setting the partial derivatives
equal to zero, we obtain the two equations - which represent the necessary conditions
- In addition,
- If x10, then equality (1) relax to 0
- If x20, then equality (2) relax to 0
18Second-order Conditions
- Proposition 1 ( second-order necessary conditions
) - Let ? be a subset of En and let be a
function on ?. - if x is a relative minimum point of f over ?,
then for any which is a feasible
direction at x we have - i)
- ii) if , then
- Proof
- The first condition is just propotion1, and the
second condition applies only if
.
19Proposition 1 (contd)
20Example 1 about Proposition 1
- For the same problem as Example 2 of Section 6.1,
we have d (d1, d2) - Thus condition (ii) of Proposition 1 applies only
if d2 0. In that case we have -
- so condition (ii) is satisfied.
21Proposition 2
- Proposition 2 (unconstrained case)
- Let x be an interior point of the set ?, and
suppose x is a relative minimum point over ? of
the function . Then - i)
- ii) for all d, .
- It means that F(x), simplified notation of
, is positive semi-definite.
22Example 2 about Proposition 2
- Consider the problem
- If we assume the solution is in the interior of
the feasible set, that is, if x1 gt 0, x2 gt 0,
then the first-order necessary conditions are -
23Example 2 (contd)
- Boundary solution is x1 x2 0
- Another solution at x1 6, x2 9
- If we fixed x1 at x1 6, then the relative
minimum with respect to x2 at x2 9. - Conversely, with x2 fixed at x2 9, the
objective attains a relative minimum w.r.t. x1 at
x1 6. - Despite this fact, the point x1 6, x2 9 is
not a relative minimum point, because the Hessian
matrix F(x)x1 6, x2 9 is not a positive
semi-definite since its determinant is negative.
24Sufficient Conditions for a Relative Minimum
- We give here the conditions that apply only to
unconstrained problems, or to problems where the
minimum point is interior to the feasible
solution. - Since the corresponding conditions for problems
where the minimum is achieved on a boundary point
of the feasible set are a good deal more
difficult and of marginal practical or
theoretical value. - A more general result, applicable to problems
with functional constrains, is given in
Chapter10.
25Proposition 3
- Proposition 3 (2-order sufficient
conditions-unconstrained case) - Let be a function defined on a region
in which the point x is an interior point.
Suppose in addition that - i)
- is positive definite
- Then x is a strict relative minimum point of f.
- Proof
- since is positive definite , there is
an a gt 0 such that - for all d, . Thus by
Taylors Theorem
26Convex Functions
- Definition
- A function f defined on a convex set ? is said to
be convex. - If, for every and every a, 0 a
1, there holds -
-
- If, for every a, 0 lta lt 1, and x1?x2, there
holds - then f is said to be strictly convex.
27Concave Functions
- Definition
- A function g defined on a convex set ? is said to
be concave if the function f -g is convex. The
function g is strictly concave if -g is strict
convex.
28Graphs of Strict Convex Function
29Graphs of Convex Function
f
x1
x2
ax1(1-a)x2
30Graphs of Concave Function
31Graph of Neither Convex or Concave
32Combinations of Convex Functions
- Proposition 1
- Let f1 and f2 be convex function on the convex
set ?. Then the f1 f2 is convex on ?. - Proposition 2
- Let f be a convex function over the convex set ?.
Then a f is convex for any a 0.
33Combinations (contd)
- Through the above two propositions it follows
that a positive combination a1 f1a1 f2am fm of
is again convex.
34Convex Inequality Constrains
- Proposition 3
- Let f be a convex function on a convex set ?. The
set is a convex for every real number
c.
35Proof
- Let , then
- and for 0 ltalt 1 ,
-
-
- Thus
-
36Properties of Differentiable Convex Functions
- Proposition 4
- Let , then f is convex over a convex set ? if
and only if - for all
37Recall
- The original definition essentially states that
linear interpolation between two points
overestimates the function, - while here stating that linear approximation
based on the local derivative underestimates the
function.
38Recall (contd)
- f is a convex function between two points
39Recall (contd)
f(y)
y
x
40Two Continuously Differentiable Functions
- Proposition 5
- Let , then f is convex over a convex set ?
containing an interior point if only if the
Hessian matrix F of f is positive semi-definite
through ?.
41Proof
- By Taylors theorem we have
- for some a, 0 a 1.
- if the Hessian is everywhere positive
semi-definite, we have -
-
42Minimization and Maximization of Convex Functions
- Theorem 1
- Let f be a convex function defined on the convex
set ?, then the set where f achieves its minimum
is convex, and any relative minimum of f is a
global minimum. - Proof (contradiction)
-
43Minimization and Maximization of Convex Functions
(contd)
- Theorem 2
- Let be a convex on the convex set ?. If
there is a point such that, for all - then x is a global minimum point of f over ?.
- Proof
-
44Minimization and Maximization of Convex Functions
(contd)
- Theorem 3
- Let f be a convex function defined on the
bounded, closed convex set ?. If f has a maximum
over ?, then it is achieved at an extreme point
of ?.
45Global Convergence of Descent Algorithms
- A good portion of the remainder of this book is
devoted to presentation and analysis of various
algorithms designed to solve nonlinear
programming problems. However, they have the
common heritage of all being iterative descent
algorithms. - Iterative
- The algorithm generated a series of points, each
point being calculated on the basis of the points
preceding it. - Descent
- As each new point is generated by the algorithm
the corresponding value of some function
(evaluated at the most recent point) decreases in
values.
46Global Convergence of Descent Algorithms (contd)
- Globally convergent
- If for arbitrary starting points the algorithm is
guaranteed to generate a sequence of points
converging to a solution, then the algorithm is
said to be globally convergent.
47Algorithm and Algorithmic Map
- We formally define an algorithm A as a mapping
taking points in a space X into other points in
X, then the generated sequence xk defined by - With this intuitive idea of an algorithm in mind,
we now generalize the concept somewhat so as to
provide greater flexibility in our analysis. - Definition
- An algorithm A is a mapping defined on a space X
that assigns to every point a subset of X.
48Mappings
- Given the algorithm yields A(xk ) which is a
subset of X. From this subset an arbitrary
element xk1 is selected. In this way, given an
initial point x0, the algorithm generates
sequences through the iteration - The most important aspect of the definition is
that the mapping A is a point-to-set mapping of
X.
49Example 1
- Suppose for x on the real line we define
- so that A(x) is an interval of the real line.
Starting at x0 100, each of the sequences below
might be generated from iterative application of
this algorithm. - 100, 50, 25, 12, -6, -2, 1, 1/2,
- 100, -40, 20, -5, -2, 1, 1/4, 1/8,
- 100, 10, 1/16, 1/100, -1/1000,
50Descent
- Definition
- Let be a given solution set and let A be an
algorithm on X. A continuous real-valued
functions Z on X is said to be a descent function
for and A if it satisfies -
51Closed Mapping
- Definition
- A point-to-set mapping A from X to Y is said to
be closed at if the assumptions - The point-to-set map A is said to be closed on X
if it is closed at each point of X. -
52Closed Mapping in Different Space
- Many complex algorithms are regards as the
composition of two or more simple point-to-point
mappings. - Definition
- Let be point-to-set mappings. The
composite mapping C BA is defined as the
point-to-set mapping - The definition is illustrated in Fig. 6.6
53Fig. 6.6
Y
X
A
A(x)
y
x
C
B
C(x)
B(x)
Z
54Corollaries of Closed Mapping
55Global Convergence Theorem
- The Global Convergence Theorem is used to
establish convergence for the following
situation. - There is a solution set . Points are
generated according to the algorithm , and
each new point always strictly decreases a
descent function Z unless the solution set is
reached. - For example, in nonlinear programming, the
solution set may be the set of minimum points
(perhaps only one point), and the descent
function may be the objective function itself.
56Global Convergence Theorem (contd)
- Let A be an algorithm on X, and suppose that,
given x0 the sequence is generated
satisfying -
- let a solution set be given, and suppose
- Then the limit of any convergence subsequence of
xk is a solution
57Global Convergence Theorem (contd)
- Corollary
- If under the conditions of the Global Convergence
Theorem consists of a single point , then
the sequence xk converges to .
58Examples to Illustrate Conditions
- Examples 4
- On the real line consider the point-to-point
algorithm - and solution set ,descent function
- The condition (iii) does not holds since A is
not closed at x 1, so, the limit of any
convergent subsequence of xk is not a solution
.
59Examples (contd)
- Example 5
- On the real line X consider the solution set
, the descent function Z(x) e-x, and the
algorithm A(x) x 1. - All the conditions of the convergence theorem
holds except (i) since the sequence generated
from any starting point x0 diverges to infinity. - It means S is not a compact set.
606.7 Order of convergence
- Definition
- Let the sequence converge to the limit
r . The order of convergence of rk is defined
as the supremum of the nonnegative numbers p
satisfying - Larger values of the order p imply faster
convergence.
61Order of convergence (contd)
- If the sequence has order p and the limit
- exists, then asymptotically we have
62Examples
- Example 1
- The sequence with rk ak where 0 lt a lt 1
- Solution
- It converges to zero with order unity, since
63Examples (contd)
- Example 2
- The sequence with for 0 lt a lt 1
- Solution
- It converges to zero with order two, since
64Linear convergence
- Most algorithm discussed in this book have an
order of convergence equal to unity. - Definition
- If the sequence rk converges to r in such way
that - the sequence is said to converge linearly to
r with convergence ratio
65Linear convergence (contd)
- The linearly convergence is sometimes referred to
as geometric convergence. - Since one with convergence ratio can be said
to have a tail that converges at least as fast as
the geometric sequence for some constant c. - The smaller the ratio the faster the rate.
- The ultimate case where is referred to
as superlinear convergence.
66Examples
- Example 3
- The sequence with rk 1/k
- Solution
- It converges to zero. The convergence is of order
one but it is not linear, since - is not less than one.
67Examples (contd)
- Example 4
- The sequence with rk (1/k)k
- Solution
- The sequence is of order unity, since
- for p gt 1.
- However,
- and hence this is superlinear convergence.
68Average rates
- All the definitions given above can be referred
to as step-wise concepts of convergence, since
they define bounds on the progress made by going
a single step from k to k1. - Another approach is to define concepts related to
the average progress per step over a large number
of steps.
69Average rates (contd)
- Definition
- Let the sequence rk converge to r . The
average order of convergence is the infimum of
the numbers p gt 1 such that - The order is infinity if the equality holds for
no p gt 1
70Examples
- Example 5
- For the sequence , 0 lt a lt 1
- Solution
- for p 2, we have ,
- for p gt 2, we have
- Thus the average order is two
71Examples (contd)
- Example 6
- For the sequence rk ak, with 0 lt a lt 1
-
- Solution
- for p gt 1 ,we have
- Thus the average order is unity.
72Convergence ratio by average method
- The most important case is that of unity order.
- We define the average convergence ratio as
- Thus for the geometric sequence rk cak, for 0
lt a lt 1 the average convergence ratio is a .