Class 1
-
What is an Algorithm?
-
Illustration of a trust region algorithm for unconstrained
minimization.
-
Introduction to Nonlinear Programming
-
Mathematical Formulation: example, level sets
-
Definitions: local and global optima
-
Definitions: convexity
-
Fundamentals of Unconstrained Optimization
(Recognizing Solutions)
-
Definitions: gradient, Hessian, Taylor's Theorem, order notation (big
and little O)
-
directional derivative, curvature, linear model, direction of steepest
descent
-
first and second order necessary optimality conditions
-
second order sufficient optimality conditions
Class 2, Cortona, Italy/02
Class 2
-
Overview of Algorithms
-
Importance of convex functions and sets for global optima
-
line search and trust region methods
-
line search: steepest descent, Newton's method (scale invariance)
-
quasi-Newton methods, trust region methods (comparisons)
-
scaling
-
rates of convergence
-
Line Search Methods,
of type
where pk is the search direction and
alphak is the step length
-
Step Length
-
Wolfe conditions (geometric interpretations)
sufficient decrease - step is not too large
curvature condition - step is not too small
-
existence of step lengths
-
Convergence of Line Search Methods
(cos theta geometric interpretation)
-
Rate of Convergence
-
linear convergence rate for steepest descent -
dependence on condition number of Hessian
-
Dennis-More condition - quasi-Newton method Hessian
approximate needs (asymptotic) convergence in search direction for
superlinear convergence
-
with proof! - quadratic convergence of Newton's method near
optimum.
Class 3, Cortona, Italy/02
Class 3
-
Trust Region Methods
-
basic outline,
-
Trust Region Subproblem
-
secular equation, optimality conditions, scaling, Cauchy point
-
Conjugate Direction Methods
-
Solving Ax=b for A pos. def. Expanding subspace theorem
-
Conjugate Gradient Methods
-
Solving Ax=b for A pos. def. Practical form. Rates of convergence
(clustering eigenvalues). Preconditioning.
-
Gauss-Newton method for nonlinear least squares: properties of GN
direction, convergence rate
Class 4, Cortona, Italy/02
Class 4
-
Nonlinear Equations
-
Newton's Method, derivation using linearization, WITH
PROOF, convergence rate, Inexact Newton method, merit functions
-
Continuation (Homotopy) methods, (general outline only)
-
Theory of Constrained Optimization,
-
Definitions: constrained program, feasible set Omega, tangent cone of Omega at
x, polar cone (dual cone).
- Geometric necessary conditions of optimality (extension of Fermat's
theorem) for min f(x) x in Omega
- Geometric characterization of optimality for the convex case,
min f(x) x in Omega , where f convex function and Omega convex set
-
Definitions: local vs global minimum, smoothing problems by adding
constraints, active constraints,
Class 5, Cortona, Italy/02
Class 5
-
Theory of Constrained Optimization, cont...
(with basic convex analysis)
-
Basic Hyperplane Separation Theorem (with proof)
-
Lemma: K is a closed convex cone if and only K=K++
(for derivation of Farkas' Lemma)
-
Linearizing Cone, Cone of Gradients, Weakest Constraint Qualification,
Karush-Kuhn-Tucker conditions
-
Lemma: K is a ccc iff K=K++ and proof using separation theorem
-
derivation of Farkas' Lemma (using above lemma); extensions/failure of Farkas'
lemma
Class 6, Cortona, Italy/02
Class 6
-
Theory of Constrained Optimization, cont...
-
Lagrange Multiplier Theorem (with proof) of
Generalized Convex Program
-
dual of (generalized) convex programs (derivation using game theory)
Class 7, Cortona, Italy/02
Class 7
-
Theory of Constrained Optimization, cont...
-
Examples using KKT conditions
-
Examples of duals of convex programs
-
sensitivity interpretation of Lagrange multipliers for convex programs
Class 8, Cortona, Italy/02
Class 8
-
Optimality and Duality Continued...
-
Examples of LP and convex QP duals in finite/infinite dimensions
-
Comparison of examples in Hilbert and Banach spaces, i.e. Lagrange
multiplier acts using an inner-product when the image of the constraint
is an inner-product space; the action is still a bilinear form when the
Lagrange multiplier is in the dual of a normed space.
(The dual of C[0,1] is BV[0,1]. The optimal value in Example 3 is attained by
a discontinuous function in (22), but the value can be
approximated as closely as desired by a continuous function.
This does not contradict the fact that C[0,1] is a Banach space.)
-
Examples of duality gaps in convex programs
-
First and Second order optimality conditions for general NLPs
-
Basics of Linear Programming
Class 9, Cortona, Italy/02
Class 9
-
Penalty Function Methods, SQP methods
Class 10, Cortona, Italy/02
Class 10
-
Primal-Dual Interior-Point Methods for Linear (and Semidefinite) Programming (via log-barrier problem)