Summary of big ideas

advertisement
Mathematical Modeling and
Optimization:
Summary of “Big Ideas”
A schematic view of
modeling/optimization process
Real-world
problem
assumptions,
abstraction,data,
simplifications
makes sense?
change the model,
assumptions?
Solution to
real-world
problem
Mathematical
model
optimization
algorithm
interpretation
Solution to
model
Mathematical models in
Optimization
• The general form of an optimization model:
min or max f(x1,…,xn)
(objective function)
subject to gi(x1,…,xn) ≥ 0
(functional constraints)
x1,…,xn  S
(set constraints)
• x1,…,xn are called decision variables
• In words,
the goal is to find x1,…,xn that
– satisfy the constraints;
– achieve min (max) objective function value.
Types of
Optimization Models
Stochastic
(probabilistic
information on data)
Discrete, Integer
(S = Zn)
Linear
(f and g are linear)
Deterministic
(data are certain)
Continuous
(S = Rn)
Nonlinear
(f and g are nonlinear)
What is Discrete Optimization?
Discrete Optimization
is a field of applied mathematics,
combining techniques from
• combinatorics and graph theory,
• linear programming,
• theory of algorithms,
to solve optimization problems
over discrete structures.
Solution Methods for Discrete
Optimization Problems
•
•
•
•
Integer Programming
Network Algorithms
Dynamic Programming
Approximation Algorithms
Integer Programming
• Programming = Planning in this context
• In a survey of Fortune 500 firms, 85% of those
responding said that they had used linear or integer
programming.
• Why is it so popular?
– Many different real-life situations can be modeled as IPs.
– There are efficient algorithms to solve IPs.
Topics in this class about
Integer Programming
• Modeling real-life situations as integer
programs
• Applications of integer programming
• Solution methods (algorithms) for integer
programs
• (maybe) Using software (called AMPL) to
solve integer programs
IP modeling techniques
Using binary variables
Restrictions on number of options
Contingent decisions
Variables (functions) with k possible values
Either-Or Constraints
Big M method
Balance constraints
Fixed Charges
Making choices with non-binary variables
Piecewise linear functions
IP applications
Facility Location Problem
Knapsack Problem
Multi-period production planning
Inventory management
Fair representation in electoral systems
Consultant hiring
Bin Packing Problem
Pairing Problem
Traveling Salesman Problem
Difficulties of real-life modeling
• The problems that you will encounter in the real
world are always a lot messier than the ``clean''
problems we looked at in this class; there are always
side constraints that complicate getting even a
feasible solution.
• Most real world problems have multiple objectives,
and it is hard to choose from among them.
• In the real world you must gain experience with how
to adapt the idealized models of academia to each
new problem you are asked to solve.
Utilizing the relationship between
problems
Important modeling skill:
– Suppose you know
how to model Problems A1,…,Ap;
– You need to solve Problem B;
– Notice the similarities between Problems Ai and B;
– Build a model for Problem B, using the model for
Problem Ai as a prototype.
• Examples:
– The version of the facility location problem as a special
case of the knapsack problem
– Transshipment and Assignment problems as special cases
of Maximum Flow Problem
Complexity of Solving Discrete
Optimization Problems
Two classes:
 Class 1 problems have polynomial-time algorithms for
solving the problems optimally.
Examples: Min. Spanning Tree Problem,
Maximum Flow Problem
 For Class 2 problems (NP-hard problems)
• No polynomial-time algorithm is known;
• And more likely there is no one.
Example: Traveling Salesman Problem
• Most discrete optimization problems are in the second class.
Three main directions to solve
NP-hard discrete optimization problems:
• Integer programming techniques
• Approximation algorithms
• Heuristics
 Important Observation: Any solution method suggests a
tradeoff between time and accuracy.
On time-accuracy tradeoff schedule:
Brute force
Integer
programming
Approximation
algorithms
Heuristics
Most
accuracy
Least
accuracy
Worst time
Best time
Solving Integer Programs (IP)
vs solving Linear Programs (LP)
LP algorithms
– Simplex Method
– Interior-point methods
IP algorithms use the above-mentioned LP
algorithms as subroutines.
The algorithms for solving LPs are much more timeefficient than the algorithms for IPs.
 Important modeling consideration: Whenever
possible avoid integer variables in your model.
LP-relaxation-based solution methods
for Integer Programs
 Branch-and-Bound Technique
 Cutting Plane Algorithms
Basic Concepts of Branch-and-Bound
The basic concept underlying the branch-and-bound technique
is to divide and conquer.
Since the original “large” problem is hard to solve directly,
it is divided into smaller and smaller subproblems
until these subproblems can be conquered.
The dividing (branching) is done by partitioning
the entire set of feasible solutions into smaller and smaller
subsets.
The conquering (fathoming) is done partially by
(i) giving a bound for the best solution in the subset;
(ii) discarding the subset if the bound indicates that
it can’t contain an optimal solution.
Summary of branch-and-bound
Steps for each iteration:
1) Branching: Among the unfathomed subproblems, select the
one that was created most recently. (Break ties according to
which has larger LP value.)
Choose a variable xi which has a noninteger value xi* in the
LP solution of the subproblem. Create two new subproblems
by adding the respective constraints xi   xi* and xi ≥  xi* .
2) Bounding: Solve the new subproblems, record their LP
solutions. Based on the LP values, update the incumbent,
and the lower and upper bounds for OPT(IP) if necessary.
3) Fathoming: For each new subproblem, apply the three
fathoming tests. Discard the subproblems that are fathomed.
4) Optimality test: If there are no unfathomed subproblems left
then return the current incumbent as optimal solution
(if there is no incumbent then IP is infeasible.)
Otherwise, perform another iteration.
Importance of tight lower and upper
bounds in branch-and-bound
 Having tight lower and upper bounds on the IP
optimal value might significantly reduce the number
of branch-and-bound iterations.
 For maximization problem,
• A lower bound can be found
by applying a fast heuristic algorithm to the problem.
• An upper bound can be found by solving a relaxation
of the problem (e.g., LP-relaxation).
 If the current lower and upper bounds are close
enough, we can stop the branch-and-bound
algorithm and return the current incumbent solution
(it can’t be too far from the optimum).
General Idea of Cutting Plane Technique



Add new constraints (cutting planes) to the problem such that
(i) the set of feasible integer solutions remains the same, i.e.,
we still have the same integer program.
(ii) the new constraints cut off some of the fractional solutions
making the feasible region of the LP-relaxation smaller.
Smaller feasible region might result in a better LP value (i.e.,
closer to the IP value), thus making the search for the optimal
IP solution more efficient.
Each integer program might have many different
formulations.
Important modeling skill:
Give as tight formulation as possible.
How? Find cutting planes that make the formulation of the
original IP tighter.
Methods of getting Cutting Planes
1) Exploit the special structure of the problem
to get cutting planes
(e.g., bin packing problem, pairing problem)
 Often can be hard to get
 Topic of intensive research
2) More general methods are also available
 Can be used automatically for many problems
(e.g., knapsack-type constraints)
Branch-and-cut algorithms
• Integer programs are rarely solved based solely on cutting
plane method.
• More often cutting planes and branch-and-bound are
combined to provide a powerful algorithmic approach for
solving integer programs.
• Cutting planes are added to
the subproblems created in branch-and-bound
to achieve tighter bounds
and thus to accelerate the solution process.
• This kind of methods are known as
branch-and-cut algorithms.
Network Models
• Minimum Spanning Tree Problem
• Traveling Salesman Problem
• Min. Spanning Tree is in Class 1 (has polynomialtime algorithms for solving the problem optimally).
• Traveling Salesman Problem is in Class 2 (NP-hard
problem).
Methods for solving NP-hard problems
Three main directions to solve
NP-hard discrete optimization problems:
• Integer programming techniques
• Heuristics
• Approximation algorithms
• We gave examples of all three methods for TSP.
• 2-approximation algorithm for TSP was given and
analyzed in details.
Some Characteristics of
Approximation Algorithms
•
•
•
•
Time-efficient (sometimes not as efficient as heuristics)
Don’t guarantee optimal solution
Guarantee good solution within some factor of the optimum
Rigorous mathematical analysis to prove the approximation
guarantee
• Often use algorithms for related problems as subroutines
The 2-approximation algorithm for TSP used the algorithm of
finding a minimum spanning tree as subroutine.
Performance of TSP algorithms in practice
• A more sophisticated algorithm (which again uses the MST
algorithm as a subroutine) guarantees a solution within
factor of 1.5 of the optimum (Christofides).
• For many discrete optimization problems, there are
benchmarks of instances on which algorithms are tested.
• For TSP, such a benchmark is TSPLIB.
• On TSPLIB instances, the Christofides’ algorithm outputs
solutions which are on average 1.09 times the optimum.
For comparison, the nearest neighbor algorithm outputs
solutions which are on average 1.26 times the optimum.
• A good approximation factor often leads to good
performance in practice.
Dynamic Programming
• Dynamic programming is a widely-used mathematical
technique for solving problems that can be divided into
stages and where decisions are required in each stage.
• The goal of dynamic programming is to find a combination
of decisions that optimizes a certain amount associated with
a system.
General characteristics of Dynamic
Programming
• The problem structure is divided into stages
• Each stage has a number of states associated with it
• Making decisions at one stage transforms one state of the
current stage into a state in the next stage.
• Given the current state, the optimal decision for each of the
remaining states does not depend on the previous states or
decisions. This is known as the principle of optimality for
dynamic programming.
• The principle of optimality allows to solve the problem
stage by stage recursively.
Problems that can be solved by
Dynamic Programming
• Shortest path problems
• Multi-period production planning
• Resource allocation problems
Advantages of Dynamic Programming
• More time-efficient compared to integer
programming (but large-scale real-life problems
might require lots of states)
• Can solve a variety of problems (however integer
programming is a more universal technique)
• Can handle stochastic data (for example, stochastic
demand in multi-period production planning)
Download