EECS260 Optimization (Fall semester 2011)
Instructor
Miguel Á. Carreira-Perpiñán
Associate professor
Electrical Engineering and Computer Science
School of Engineering
University of California, Merced
mcarreira-perpinan-[at]-ucmerced.edu; 209-2284545
Office: 284, Science & Engineering Building
Office hours: Tuesdays 2:45-3:45pm (SE284).
Lectures: Tuesdays/Thursdays 1:30-2:45pm (COB265).
Lab class: Wednesdays 12:30-3:20pm (Linux Lab, SE138).
Course web page: http://faculty.ucmerced.edu/mcarreira-perpinan/EECS260
Course description
Optimization problems arise in multiple areas of science, engineering and business. This course introduces theory and numerical methods for continuous multivariate optimization (constrained and unconstrained), including: line-search and trust-region strategies; conjugate-gradient, Newton, quasi-Newton and large-scale methods; linear programming; quadratic programming; penalty and augmented Lagrangian methods; sequential quadratic programming; and interior-point methods. The primary programming tool for this course is Matlab.
Prerequisites: MATH 23, MATH 24, MATH 141 or equivalent (undergraduate courses in linear algebra and multivariate calculus). Basic concepts will be briefly reviewed during the course.
Textbook
Required textbook (get the errata and additional errata):
Other recommended books:
If you want to refresh your knowledge of linear algebra and multivariate calculus, the following are helpful (any edition or similar book is fine):
- Seymour Lipschutz: Schaum's Outline of Linear Algebra. McGraw-Hill.
- Murray Spiegel: Schaum's Outline of Vector Analysis. McGraw-Hill.
- Frank Ayres: Schaum's Outline of Matrices. McGraw-Hill.
- Richard Bronson: Schaum's Outline of Matrix Operations. McGraw-Hill.
- Erwin Kreyszig: Advanced Engineering Mathematics. Wiley.
- Gilbert Strang: Introduction to Linear Algebra. Wellesley-Cambridge Press.
- Gilbert Strang: Linear Algebra and Its Applications. Brooks Cole.
Syllabus and required textbook reading
Syllabus
Before each class, you should have read the corresponding part of the textbook and the notes. I will teach the material in the order below (which is more or less the order in the book).
- Introduction (ch. 1: all) and math review (appendix A: skim it)
- Unconstrained optimization:
- Fundamentals of unconstrained optimization (ch. 2: all)
- Line search methods (ch. 3: pp. 30-49; skip proofs in sec. 3.3)
- Trust-region methods (ch. 4: pp. 66-71)
- Conjugate gradient methods (ch. 5: all, but skim pp. 113-114, 119-120, 125-133)
- Quasi-Newton methods (ch. 6: pp. 135-146; skip 147-152; skim rest)
- Large-scale unconstrained optimization (ch. 7: pp. 164-170, 176-181, 185-190; skip 171-175, 182-184)
- Calculating derivatives (ch. 8: pp. 193-206; skim 199-201, 203; skip rest)
- Derivative-free optimization (ch. 9: all except pp. 226-229, 237)
- Least-squares problems (ch. 10: all but skim pp. 251-252, 259-268)
- Nonlinear equations (ch. 11: all but skim 277-279, 287-295; skip 283-284)
- Constrained optimization:
- Theory of constrained optimization (ch. 12: all except as follows: from section 12.2 see defs. 12.3, 12.4 and skip the rest; from section 12.5 skip the proofs of second-order conditions; skip sections 12.4, 12.6, 12.7)
- Linear programming: the simplex method (ch. 13: all except as follows: skip sections 13.4 to 13.7)
- Linear programming: interior-point methods (ch. 14: all except as follows: skip lemmas 14.1, 14.2 and the proof of th. 14.3; skim pp. 410, 412-413; skip section 14.3)
- Fundamentals of algorithms for nonlinear constrained optimization (ch. 15: pp. 421-437)
- Quadratic programming (ch. 16: all except as follows: skip sections 16.2-3, skim section 16.5 and skim pp. 482-485)
- Penalty and augmented Lagrangian methods (ch. 17: all except as follows: skip def. 17.1, skim section 17.4)
- Sequential quadratic programming (ch. 18: sections 18.1-2; skip rest)
- Interior-point methods for nonlinear programming (ch. 19: sections 19.1, 19.2 and 19.6; skip rest)
Handouts and assignments
Course grading
The course grading will be based on three projects and a final exam, as follows (but note that too low a grade in the exams cannot be compensated by a high grade in the projects or vice versa):
- Project 1 (15%): implementing a method for unconstrained optimization problems
- Project 2 (25%): implementing a method for constrained optimization problems
- Project 3 (10%): discussing research papers related to optimization
- Midterm exam (25%): in-class, closed-book, consisting of problems and conceptual questions
- Final exam (25%): as the midterm
While I encourage you to discuss your work with other students, the projects and the exam must be the result of your own work without collaboration.
I will also give homework exercises (mainly from the textbook) of two types, pencil-and-paper and Matlab programming. I will not ask you to solve them, i.e., they will not count towards your grade. I will give the solutions and solve some of the exercises in class. However, I strongly recommend that you try to solve all the exercises on your own.
Grade curves: midterm, final.
Optimization links
- Demos:
- Matlab Optimization Toolbox. You may find useful to compare the textbook and the user's guide. You can see some demos by running Matlab and typing demo toolbox optimization in the command window. But note that you will be writing your own code, rather than using the Toolbox.
- NEOS Guide: information about most areas of optimization
- Optimization Online: a repository of eprints about optimization
- Mathematical Programming Glossary
- Matrix identities (handy formulas for matrix derivatives, inverses, etc.):
Matlab tutorials
If you have never used Matlab, there are many online tutorials, for example:
Other links
Miguel A. Carreira-Perpinan
Last modified: Sun Jan 22 19:42:56 PST 2012
UC Merced |
EECS |
MACP's Home Page