Contributed talks in ECCO:

advertisement
European Chapter on
Combinatorial Optimization
(ECCO) XXII
May 17-19, 2009
Jerusalem
Abstracts booklet
The conference is supported by:
Sunday
9:15-10:15 Plenary talk – Alain Hertz "Solution methods for the inventory routing
problem"
We consider an inventory routing problem in discrete time where a supplier has to
serve a set of customers over a time horizon. At each discrete time, a fixed quantity is
produced at the supplier and a fixed quantity is consumed at each customer. A
capacity constraint for the inventory is given for the supplier and each customer, and
the service cannot cause any stock-out situation. Two different replenishment policies
are considered: when a customer is served, one can either deliver any quantity that
does not violate the capacity constraint, or impose that the inventory level at the
customer should reach its maximal value. The transportation cost is proportional to
the distance traveled, whereas the inventory holding cost is proportional to the level of
the inventory at the customers and at the supplier. The objective is the minimization
of the sum of the inventory and transportation costs. We describe the current best
exact and heuristic algorithms for the solution of this complex problem.
10:45-12:00 Parallel sessions:
Track 1: Integer and mixed integer formulations 1
1. Julia Rieck and Jürgen Zimmermann, " A new mixed integer linear model for a
rich vehicle routing problem with docking constraints"
In this paper we address a rich vehicle routing problem that arises in real-life
applications. Among other aspects we consider
•
•
•
•
•
•
heterogenous vehicles,
simultaneous delivery and pick-up at customer locations,
a total working time for vehicles from departure to arrival back at the depot,
time windows for customers during which delivery and pick-up can occur,
a time window for the depot, and
multiple use of vehicles throughout the planning horizon.
To guarantee a coordinated material flow at the depot, we include the timed allocation
of vehicles to loading bays at which the (un-)loading activities can occur. The
resulting rich vehicle routing problem can be formulated as a two-index vehicle-flow
formulation where we use binary variables to indicate if a vehicle traverses an arc in
the optimal solution and to sequence the (un-)loading activities at the depot.
In our performance analysis, we use CPLEX 11.0 to solve instances that are
derived from the extended Solomon test set. The selective implementation of
preprocessing techniques and cutting planes improves the solver performance
significantly. In this context, we strengthen the domains of auxiliary variables as well
as the big-M-constraints of our model. We identified clique, implied bound, and flow
cover cuts as particularly suitable for our purposes. Additionally, we take into
consideration the rounded capacity cuts for the capacitated vehicle routing problem.
As average computational time for instances with 25 customers we obtain 310.06
seconds and for instances with 30 customers 1305.10 seconds. A comparison to the
results obtained for a three-index vehicle-flow formulation shows that the model is
suitable for practical applications.
2. Bartosz Sawik, "Bi-objective dynamic portfolio optimization by mixed integer
programming"
This paper presents the dynamic portfolio optimization problem formulated as a biobjective mixed integer linear program. The computational efficiency of MILP model
is very important for applications to real-life financial and other decisions where the
constructed portfolios have to meet numerous side constraints. The portfolio selection
problem considered is based on a multi-period model of investment, in which the
investor buys and sells securities in consecutive periods. An extension of the
Markowitz portfolio optimization model is considered, in which the variance has been
replaced with the Value-at-Risk (VaR). The VaR is a quantile of the return
distribution function. The advantage of using this measure in portfolio optimization is
that this value of risk is independent of any distribution hypothesis.
The portfolio selection problem is usually considered as a bi-objective optimization
problem where a reasonable trade-off between expected rate of return and risk is
sought. The objective is to maximize future returns by picking the best amount of
stocks. The only way to improve future returns is to increase the risk level that the
decision maker is willing to accept. In the classical Markowitz approach, future
returns are random variables controlled by such parameters as the portfolio efficiency,
which is measured by the expectation, while risk is calculated by the standard
deviation. As a result the classical problem is formulated as a quadratic program with
continuous variables and some side constraints.
The objective of the problem considered in this paper is to allocate wealth on different
securities to maximize the portfolio expected return and the threshold of the
probability that the return is not less than a required level. The auxiliary objective is
minimization of risk probability of portfolio loss.
The input data for computations consist of 3500 historic daily quotations divided into
14 investment periods. The results of some computational experiments with the mixed
integer programming approach modeled on a real data from the Warsaw Stock
Exchange are reported.
3. Alberto Ceselli, Giovanni Righini and Gregorio Tirado Dominguez,
"Mathematical programming algorithms for the double TSP with multiple stacks"
The Double TSP with Multiple Stacks requires to find a minimum cost pair of
Hamiltonian tours, visiting two given sets of pickup and delivery customers, such that
the two sequences of visits can be executed by a single vehicle that can trasnport
goods organizing them in stacks of identical capacity. Stacks are managed according
to a LIFO policy and items cannot be rearranged. For this problem only heuristic
algorithms have been published so far.
We present a decomposition approach that leads to Lagrangian relaxation and column
generation algorithms. We also present some preliminary computational results.
Track 2: Approximation and Online algorithms:
1. Zvi Lotker, Boaz Patt-Shamir and Dror Rawitz, "Rent, lease or buy: randomized
strategies for multislope ski rental"
In the Multislope Ski Rental problem, the user needs a certain resource for some
unknown period of time. To use the resource, the user must subscribe to one of
several options, each of which consists of a one-time setup cost (“buying price”), and
cost proportional to the duration of the usage (“rental rate”). The larger the price, the
smaller the rent. The actual usage time is determined by an adversary, and the goal of
an algorithm is to minimize the cost by choosing the best option at any point in time.
Multislope Ski Rental is a natural generalization of the classical Ski Rental problem
(where the only options are pure rent and pure buy), which is one of the fundamental
problems of online computation. The Multislope Ski Rental problem is an abstraction
of many problems, where online choices cannot be modeled by just two alternatives,
e.g., power management in systems which can be shut down in parts.
In this work we study randomized online strategies for Multislope Ski Rental.
Our results include an algorithm that produces the best possible randomized online
strategy for any additive instance, where the cost of switching from one alternative to
another is the difference in their buying prices; and an e-competitive randomized
strategy for any (non-additive) instance. We also provide a randomized strategy with
a matching lower bound for the case of two slopes, where both slopes have positive
rents.
2. Asaf Levin and Uri Yovel, "Algorithms for (p,k)-uniform unweighted set cover
problem"
We are given n base elements and a finite collection of subsets of them. The size of
any subset varies between p to k (p<k). In addition, we assume that the input contains
all possible subsets of size p. Our problem is to find a subcollection of minimumcardinality which covers all the elements. This problem is known to be NP-hard. We
provide two approximation-algorithms for it, one for the generic case, and an
improved one for the special case of (p,k) = (2,4).
The algorithm for the generic case is a greedy one, based on packing phases: at each
phase we pick a collection of disjoint subsets covering i new elements, starting from
i=k down to i=p+1. At a final step we cover the remaining base elements by the
subsets of size p. We derive the exact approximation ratio of this algorithm for all
values of k and p, which is less than Hk, where Hk is the k’th harmonic number.
However, the algorithm exhibits the known improvement methods over the greedy
one for the unweighted k-set cover problem (in which subset sizes are only restricted
not to exceed k), and hence it serves as a benchmark for our improved algorithm.
The improved algorithm for the special case of (p,k) = (2,4) is based on local-search:
it starts with a feasible cover, and then repeatedly tries to replace sets of size 3 and 4
so as to maximize an objective function which prefers big sets over small ones. For
this case, our generic algorithm achieves an approximation ratio of 1.5, and the localsearch algorithm achieves a better ratio, which is bounded by 1.458333...
3. Leah Epstein and Gyorgy Dosa, "Preemptive online scheduling with reordering"
We consider online preemptive scheduling of jobs, arriving one by one, on m parallel
machines. A buffer of fixed size K>0, which assists in partial reordering of the input,
is available, to be used for the storage of at most K unscheduled jobs. We consider
several variants of the problem. For general inputs and identical machines, we show
that a buffer of size Theta(m) reduces the overall competitive ratio from e/(e-1) to 4/3.
Surprisingly, the competitive ratio as a function of m is not monotone, unlike the case
where K=0.
12:10-13:00 Parallel sessions:
Track 1: Integer and mixed integer formulations 2:
1. Jan Pelikán and Jakub Fischer, "Optimal doubling models"
The article deals with the problem of doubling the part of the system in order to
increase the reliability of this system. The task is to find out which part should be
doubled if it is not possible to double all of them for economic reasons. The goal is to
maximize the system’s reliability characterized by the probability of the faultless
functioning or to minimize the losses caused by the system’s failure. The article
proposes a procedure based on linear programming methods that find the optimal
decision, which of the parts should be doubled.
2. Jorge Riera-Ledesma and Juan José Salazar González, "A branch-and-cut-andprice for the multiple vehicle traveling purchaser problem"
The Multiple Vehicle Traveling Purchaser Problem aims designing a set of optimal
school bus routes with the purpose of carrying pupils to their center of learning. The
problem is defined as follows: given a set of users going to a specific location, let's
say school, there exist a set of potential bus stops. Each bus stop is reachable by a
subset of users. A fleet of homogeneous vehicles is available with this purpose. The
aim is assigning the set of users to a subset of potential stops, finding least cost trips
that starts and ends at the school, and choosing a vehicle to serve each trip so that
each stop with user assigned is served by a trip, and the total number of users assigned
to the stops of a trip does not exceed the capacity of the vehicle.
We propose an integer linear formulation, and several families of valid inequalities
are derived to strengthen the linear relaxation. A branch-and-cut-and-price procedure
has been developed, and a extensive computational experience is presented.
Track 2: Approximation algorithms for bin packing problems:
1. Rolf Harren and Rob van Stee, "An absolute 2-approximation algorithm for twodimensional bin packing"
We consider the problem of packing rectangles into an unlimited supply of equallysized, rectangular bins. All rectangles have to be packed non-overlapping and
orthogonal, i.e., axis-parallel. The goal is to minimize the number of bins used.
In general, multi-dimensional packing is strongly related to scheduling, as one
dimension can be associated with the processing time of the tasks (items) and the
other dimensions make up geometrically ordered resources. In two-dimensional
packing, e.g., tasks have to be processed on consecutive machines. The bin packing
setting that we consider here applies when the machines work in shifts, e.g., only
during the daytime. Thus the schedule has to be fixed into separate blocks.
Most of the previous work on two-dimensional bin packing has focused on the
asymptotic approximation ratio, i.e., allowing a constant number of spare bins. In this
setting, algorithms by Caprara and by Bansal et al. have approximation ratio better
than 2. On the other hand, the additive constants are (very) large and thus these
algorithms give bad approximation ratios for instances with small optimal value.
In terms of absolute approximability, i.e., without wasting additional bins, the
previously best-known algorithms have a ratio of 3. If rotations by 90 degrees are
permitted, the problem becomes somewhat easier as we do not have to combine
differently oriented thin items into a bin. We previously showed an absolute 2approximation for this case.
Here we present an algorithm for the problem without rotations with an absolute
worst-case ratio of 2, i.e., our algorithm uses at most twice the optimal number of
bins. It was shown that it is strongly NP-complete to decide wether a set of squares
can be packed into a given square. Therefore our result is best possible unless P = NP
and we thus settle the question of absolute approximability of two-dimensional bin
packing.
Our algorithm consists of three main parts depending on the optimal value of the
given instance. As we do not know this optimal value in advance, we apply all three
algorithms and allow them to fail if the requirement on the optimal value is not
satisfied. The algorithms with asymptotic approximation ratio less than 2 give an
absolute 2-approximation for instances with large optimal value. We design further
algorithms to solve instances that fit into one bin and that fit into a constant number of
bins, respectively. All three algorithms together make up an absolute 2approximation.
2. Leah Epstein and Asaf Levin, "Bin packing with general cost structures"
Following the work of Anily et al., we consider a variant of bin packing, called bin
packing with general cost structures and design an asymptotic fully polynomial time
approximation scheme (AFPTAS) for this problem. In the classic bin packing
problem, a set of one-dimensional items is to be assigned to subsets of total size at
most 1, that is, to be packed into unit sized bins. However, in our problem, the cost of
a bin is not 1 as in classic bin packing, but it is a non-decreasing and concave function
of the number of items packed in it, where the cost of an empty bin is zero. The
construction of the AFPTAS requires novel techniques for dealing with small items,
which are developed in this work. In addition, we develop a fast approximation
algorithm which acts identically for all non-decreasing and concave functions, and has
an asymptotic approximation ratio of 1.5 for all functions simultaneously.
14:00-15:15 Parallel sessions:
Track 1: Algorithms:
1. Igor Averbakh, "Minmax regret bottleneck problems with solution-induced
interval uncertainty structure"
We consider minmax regret bottleneck subset-type combinatorial optimization
problems, where feasible solutions are some subsets of a finite ground set. The
weights of elements of the ground set are uncertain; for each element, an uncertainty
interval that contains its weight is given. In contrast with previously studied interval
data minmax regret models, where the set of scenarios (possible realizations of the
vector of weights) does not depend on the chosen feasible solution, we consider the
problem with solution-induced interval uncertainty structure. That is, for each element
of the ground set, a nominal weight from the corresponding uncertainty interval is
fixed, and it is assumed that only the weights of the elements included in the chosen
feasible solution can deviate from their respective nominal values. We present a
number of algorithmic results for bottleneck minmax regret problems of this type, in
particular, a quadratic algorithm for the problem on a uniform matroid.
2. Endre Boros, Ondrej Cepek, Alex Kogan and Petr Kucera, "A subclass of Horn
CNFs optimally compressible in polynomial time"
The problem of Horn Minimization (HM) can be stated as follows: given a Horn CNF
representing a Boolean function f, find a CNF representation of f which consists of a
minimum possible number of clauses. This is a classical combinatorial optimization
problem with many practical applications. For instance, the problem of knowledge
compression for speeding up queries to propositional Horn expert systems is
equivalent to HM.
HM is a computationally difficult problem: it is known to be NP-hard even if the
input is restricted to cubic Horn CNFs. On the other hand, there are two subclasses of
Horn CNFs for which HM is known to be solvable in polynomial time: acyclic and
quasi-acyclic Horn CNFs. In this talk we introduce a new class of Horn CNFs which
properly contains both of the known classes and describe a polynomial time HM
algorithm for this new class.
3. Alexander Zadorojniy, Guy Even and Adam Shwartz, "A strongly polynomial
algorithm for controlled queues"
We consider the problem of computing optimal policies of finite-state, finite-action
Markov Decision Processes (MDPs). A reduction to a continuum of constrained
MDPs (CMDPs) is presented such that the optimal policies for these CMDPs
constitute a path in a graph defined over the deterministic policies. This path
contains, in particular, an optimal policy of the original MDP. We present an
algorithm based on this new approach that finds this path and thus an optimal policy.
In the general case this path might be exponentially long in number of states and
actions. We prove that the length of this path is polynomial if the MDP satisfies a
coupling property. Thus we obtain a strongly polynomial algorithm for MDPs that
satisfy the coupling property. We prove that discrete time versions of controlled
M/M/1 queues induce MDPs that satisfy the coupling property. The only previously
known polynomial algorithm for controlled M/M/1 queues in the expected average
cost model is based on linear programming (and is not known to be strongly
polynomial). Our algorithm works both for the discounted and expected average cost
models, and the running time does not depend on the discount factor.
Track 2: Scheduling 1:
1.
Bezalel Gavish, "Batch sizing and scheduling in serial multistage production
system"
The authors propose a procedure to model and determine the optimal batch sizes and
their schedule in a serial multistage production process with constant demand rate for
the final product. The model takes into consideration the batch setup time in each
stage, the processing times of a batch as a function of batch size and production stage.
It considers the holding and shortage costs of the final product, and the setup and
processing costs at each stage of the production process. The objective is to maximize
the net profit per unit of time derived from selling the final products. We report on
extensive computational experiments that tested the impact of different factors on the
objective function and the optimal batch size. A feature unique to this paper is the
concentration on multistage serial systems consisting of hundreds of stages in the
production process, in such systems a batch can spend from one to eight months in the
production process. Batch sizing and scheduling has a great impact on the net
revenues derived from the production facility. Due to the tight tolerances of the
production process which are computer controlled, setup and processing times are
deterministic.
2. Yaron Leyvand, Dvir Shabtay, George Steiner and Liron Yedidsion, "Just-intime scheduling with controllable processing times on single and parallel
machines"
We study bicriteria scheduling problems with controllable processing times on single
and parallel machines. Our objectives are to maximize the weighted number of jobs
that are completed exactly at their due date and to minimize the total resource
allocation cost. We consider four different models for treating the two criteria. We
prove that three of these problems are NP-hard even on a single machine, but
somewhat surprisingly, the problem of maximizing an integrated objective function
can be solved in polynomial time even for the general case of a fixed number of
unrelated parallel machines. For the three NP-hard versions of the problem, with a
fixed number of machines and a discrete resource type, we provide a pseudopolynomial time optimization algorithm, which is converted to a fully polynomial
time approximation scheme to find a Pareto optimal solution.
3. Gur Mosheiov and Assaf Sarig, "Scheduling a maintenance activity to minimize
total weighted completion-time"
We study a single machine scheduling problem. The processor needs to go through a
maintenance activity, which has to be completed prior to a given deadline. The
objective function is minimum total weighted completion time. The problem is proved
to be NP-hard, and an introduction of a pseudo-polynomial dynamic programming
algorithm indicates that it is NP-hard in the ordinary sense. We also present an
efficient heuristic which is shown numerically to perform well.
15:25-16:15 Parallel sessions:
Track 1: Supply chain:
1. Tadeusz Sawik, "A two-objective mixed integer programming approach for
supply chain scheduling"
The integrated and hierarchical approaches based on mixed integer programming are
proposed for a bi-objective coordinated scheduling in a customer driven supply chain.
The supply chain consists of multiple suppliers (manufacturers) of parts and a single
producer of finished products. Given a set of customer orders, the problem objective
is to determine a coordinated schedule for the manufacture of parts by each supplier,
for the delivery of parts from each supplier to the producer, and for the assignment of
orders to planning periods at the producer, such that a high revenue or a high
customer service level is achieved and the total cost of holding supply chain inventory
is minimized.
The proposed two approaches are described below.
1. Integrated (simultaneous) approach.
The coordinated schedules for
customer orders, for manufacturing of parts and for supply of parts are
determined simultaneously to achieve a minimum number of tardy orders or a
minimum lost revenue due to tardiness or full rejection of orders, at a
minimum cost of holding total supply chain inventory of parts and finished
products.
2. Hierachical (sequential) approach.
First, the order acceptance/due date
setting decisions are made to select maximal subset of orders that can be
completed by requested due dates, and for the remaining orders delayed due
dates are determined to minimize the lost revenue owing to tardiness or
rejection of the orders. The due dates must satisfy capacity constraints and are
considered to be deadlines at the order scheduling level. Next, order deadline
scheduling is performed to meet all committed due dates and to minimize cost
of holding finished product inventory of orders completed before the
deadlines. Finally, scheduling of manufacturing and supply of parts,
coordinated with the schedule for orders is accomplished to meet demand for
parts at a minimum cost of part supply and inventory holding.
Numerical examples modeled after a real-world scheduling in a customer driven
supply chain of high-tech products are presented and some computational results are
reported to compare the two approaches. To determine subsets of non-dominated
solutions for the integrated approach, in the computational experiments the weightedsum program is compared with the Tchebycheff program.
2. Grzegorz Pawlak, "Multi-layer agent scheduling and control model in the car
factory"
The complex planning and production tasks in the car production factory will generate
the necessity to apply complicated scheduling and control systems. The goal of the
work is to have the production in balance, without breaks and without having too
many parts in a stock. The production process is controlled and measured by taking
into account many coefficients. The inventory level, the buffers occupation,
production flow, deadlines, sequence quality, failures etc. Building the global
optimization system is difficult and usually such a system is not considering all
constraints and dependencies. The alternative is to build the model with the
distributed local agents – schedulers and control units fast adapting to the changing
circumstances. The disturbance of the process caused by the demand changes or break
downs in the lines should lead to the appropriate reaction of the scheduling and
control system. In the fast changing environment the intelligent and self adopted
scheduling and control system may be the advantage in the production optimization
process.
In the situations where is a need to set the production speed or the synchronization of
production line in multi-line production system the agent driven control could be the
reasonable alternative. Additionally, the usage and intelligent control of buffers and
switches can improve the production efficiency. In every car there are thousands of
parts to be linked and that is why such production system, can be regarded as one of
most complicated and other segment companies can follow that application example.
So the model is not restricted only to the car factories.
The presented model is taking into account the multi – layer control agent system.
There are agents to play specific roles. There is the control, monitoring and
visualization, scheduling agents. Each of them has defined the execution domain and
the activity area. The configuration and the management of such agent systems have
been taken into account. The classical scheduling models and algorithms have been
incorporated into the bigger view of the production
Summarizing, there is a need to have a self adopting production control system, which
will be able to keep production with appropriate speed available for the demand
driven production plan and according to the current state of production facilities and
resources. In response to this problem, we would like to create the multi- line, self
adopting production control system, which will consists of high-level control
software, communication protocols and production line controllers behavior
scenarios, prepared for different modes and situations.
Track 2: Networks:
1. Oren Ben-Zwi, Danny Hermelin, Daniel Lokshtanov and Ilan Newman, "An
exact almost optimal algorithm for target set selection in social networks"
The Target Set Selection problem proposed by Kempe, Kleinberg, and Tardos, gives a
nice clean combinatorial formulation for many problems arising in economy,
sociology, and medicine. Its input is a graph with vertex thresholds, the social
network, and the goal is to find a subset of vertices, the target set, that ``activates" a
prespecified number of vertices in the graph. Activation of a vertex is defined via a
so-called activation process as follows: Initially, all vertices in the target set become
active. Then at each step i of the process, each vertex gets activated if the number of
its active neighbors at iteration i-1 exceeds its threshold. The activation process is
``monotone" in the sense that once a vertex is activated, it remains active for the
entire process.
Unsurprisingly perhaps, Target Set Selection is NPC. More surprising is the fact that
both of its maximization and minimization variants turn out to be extremely
hard to approximate, even for very restrictive special cases. The only known case for
which the problem is known to have some sort of acceptable worst-case solution is the
case where the given social network is a tree and the problem becomes polynomialtime solvable. In this paper, we attempt at extending this sparse landscape of tractable
instances by considering the treewidth parameter of graphs. This parameter roughly
measures the degree of tree-likeness of a given graph, e.g. the treewidth of a tree is 1,
and has previously been used to tackle many classical NPhard problems in the
literature.
Our contribution is twofold: First, we present an algorithm for Target Set Selection
running in nO(w) time, for graphs with n vertices and treewidth bounded by w. The
algorithm utilizes various combinational properties of the problem; drifting somewhat
from standard dynamic-programming algorithms for small treewidth graphs.
Also, it can be adopted to much more general settings, including the case of directed
graphs, weighted edges, and weighted vertices. On the other hand, we also show that
it is highly unlikely to find an no(\sqrt{w}) time algorithm for Target Set Selection, as
this would imply a sub-exponential algorithm for all problems in SNPclass. Together
with our upper bound result, this shows that the treewidth parameter determines the
complexity of Target Set Selection to a large extent, and should be taken into
consideration when tackling this problem in any scenario.
2. André Amaral, Alberto Caprara, Juan José Salazar González and Adam
Lechford, "Lower bounds for the minimum linear arrangement of a graph"
Minimum Linear Arrangement is a classical basic combinatorial optimization
problem from the 1960s that turns out to be extremely challenging in practice. In
particular, for most of its benchmark instances, even the order of magnitude of the
optimal solution value is unknown, as testified by the surveys on the problem that
contain tables in which the best known solution value has many more digits than the
best known lower bound value. Since, for most of the benchmark instances, finding a
provably optimal solution appears to be completely out of reach at the moment, in this
paper we propose a linear-programming based approach to compute lower bounds on
the optimum. This allows us, for the first time, to show that the best known solutions
are indeed not far from optimal.
16:45-17:45 Plenary talk – Maurice Queyranne "Structural and algorithmic
properties for parametric minimum cuts"
We consider the minimum s-t-cut problem in a network with parametrized arc
capacities. Classes of this parametric problem have been shown to enjoy the nice
structural property that minimum cuts are nested and, following the seminal work of
Gallo, Grigoriadis and Tarjan (1989), the nice algorithmic property that all minimum
cuts can be computed in the same asymptotic time as a single minimum cut.
We present a general framework for parametric minimum cuts that extends and
unifies such results.
We define two conditions on parametrized arc capacities that are necessary and
sufficient for (strictly) decreasing differences of the parametric cut function.
Known results in parametric submodular optimization then imply the structural
property.
We show how to construct appropriate flow updates in linear time under the above
conditions, implying that the algorithmic property also holds under these conditions.
We then consider other classes of parametric minimum cut problems, without
decreasing differences, for which we establish the structural and/or the algorithmic
property, as well as other cases where nested minimum cuts arise.
This is joint work with Frieda Granot and S. Thomas McCormick (Sauder School of
Business at UBC) and Fabio Tardella (Universitá La Sapienza, Rome)
17:55-19:10 Parallel sessions:
Track 1: Scheduling 2:
1. Jan Pelikán, " Hybrid flow shop with adjustment"
The subject of this paper is a flow-shop based on a case study aimed at the
optimisation of ordering production jobs in mechanical engineering, in order to
minimise the overall processing time, the makespan.
There is a given set of production jobs to be processed by the machines installed on
the shop floor. A job is a product batch for a number of units of given product
assigned to processing on the machine. Each job is assigned to a certain machine,
which has to be adjusted by an adjuster. This worker adjusts all of the machines
installed on the floor but at a given time, he can adjust only one machine, i.e. he
cannot adjust more machines simultanously. Each adjusted machine is supposed to be
capable of immediate starting to process the job for which it has been adjusted. No job
processing is allowed to be broken (i.e., intermittent processing is not admissible), and
each machine is able to process only one job at a time. The processing time
minimisation problem is aimed at the reduction of the waiting times of both the
machines for the adjuster and of the adjuster for a machine to be adjusted. After
completing a job on the machine, the machine has to wait if the adjuster is finishes
adjusting another machine. If all machines are processing jobs, the adjuster has to
wait. A solution is represented by the order of the jobs in which the adjuster adjusts
the respective machines. This order also determines the order of the machines to be
adjusted by the adjuster. At the same time, the job ordering generates the order of the
jobs on the machines when several jobs are assigned to the same machine. In the
literature, hybrid flow-shop is defined as a problem of processing jobs which consists
of two or more stages, with one or more processors at each stage. Each of the jobs to
be processed consists of two or more tasks and each task is processed within its own
stage. The jobs are non-preemptable and each subsequent stage is only started after
the processing of the previous stage is completed. A hybrid flow-shop as described
here further assumes that the following job at the first stage can be processed
immediately upon completion of the preceding job at the first stage. Therefore, the
problem of this case study can be formulated as a two-stage hybrid flow-shop, in
which the first stage is represented by the work of an adjuster, who is viewed as the
only one processor at the first stage. The second stage is represented by the machines
on the shop floor, viewed as parallel processors; however, jobs are uniquely assigned
to these processors. The objective function is defined as the makespan value, i.e., the
overall processing time spent on all jobs at both stages. Contrary of the hybrid flowshop defined in literature, there are two differences for two consecutive jobs here:
a) If both jobs are assigned to the same machine, i.e., the same second-stage
processor, the second job's first stage can only be started upon completion of the first
job's second stage, when the machine is available;
b) If each of the two consecutive jobs is assigned to a different machine as a secondstage processor, the second job's first stage can be started immediately upon
completion of the first job's first stage.A mathematical model is proposed, a heuristic
method is formulated, and the NP hardness of the problem, called a "hybrid flow-shop
with adjustment," is proved.
2. Alexander Lazarev, "An approximation method for estimating optimal value of
minimizing scheduling problems"
3. Alexander Kvaratskhelia and Alexander Lazarev, "An algorithm for total
weighted completion time minimization in preemptive equal job length with release
dates on a single machine"
We consider the minimizing total weighted completion time in preemptive equal job
length scheduling problem on a single machine with release dates. We propose a
polynomial time algorithm that solves the problem. Before this paper, the problem is
known to be open
(http://www.lix.polytechnique.fr/~durr/OpenProblems/1_rj_pmtn_pjp_sumWjCj/).
We refer to the total weighted completion time problem (without release dates,
preemption and arbitrary job length) that can be solved in polynomial time using
Smith’s rule.
Our algorithm is also based on the Smith’s rule, which is applied to solve a subproblem with purpose to find an optimal processing order within a time interval
between two adjacent release dates. An optimal schedule is being found starting from
the least release date. On each step we construct a sub-problem of jobs, which is
already available wince corresponding release date. We find an optimal processing
order of these jobs using Smith’s rule, and then we switch to the next release date.
We propose a number of properties of the optimal schedule, which prove the
optimality of the schedule constructed by Algorithm in polynomial time.
Track 2: Optimization 1:
1. Michael Katz and Carmel Domshlak, "Pushing the envelope of abstraction-based
admissible heuristics"
The field of automated, domain-independent planning seeks to build general-purpose
algorithms enabling a system to synthesize a course of action that will achieve certain
goals. Such algorithms perform reachability analysis in large-scale state models that
are implicitly described in a concise manner via some intuitive declarative language.
And though planning problems have been studied since the early days of Artificial
Intelligence research, recent developments (and, in particular, recent developments in
planning as heuristic search) have dramatically advanced the field, and also
substantially contributed to some related fields such as software/hardware
verification, control, information integration, etc.
The difference between various algorithms for planning as heuristic search is mainly
in the heuristic functions they define and use. Most typically, an (admissible) heuristic
function for domain-independent planning is defined as the (optimal) cost of
achieving the goals in an over-approximating abstraction of the planning problem in
hand. Such an abstraction is obtained by relaxing certain constraints that are present in
the specification of the real problem, and the desire is to obtain a provably poly-time
solvable, yet informative abstract problem. The main questions are thus:
1.What constraints should we relax to obtain such an effective over-approximating
abstraction?
2. How should we combine information provided by multiple such abstractions?
In this work we consider both these questions, and present some recent formal results
that help answering these questions (sometimes even to optimality). First, we consider
a generalization of the popular ``pattern database'' (PDB) homomorphism abstractions
to what is called ``structural patterns''. The basic idea is in abstracting the problem in
hand into provably tractable fragments of optimal planning, alleviating by that the
constraint of PDBs to use projections of only low dimensionality. We introduce a
general framework for additive structural patterns based on decomposing the problem
along certain graphical structure induced by it, suggest a concrete non-parametric
instance of this framework called fork-decomposition, and formally show that the
admissible heuristics induced by the latter abstractions provide state-of-the-art worstcase informativeness guarantees on several standard domains. Specifically, we
describe a procedure that takes a classical planning task, a forward-search state, and a
set of abstraction-based admissible heuristics, and derives an optimal additive
composition of these heuristics with respect to the given state. Most importantly, we
show that this procedure is polynomial-time for arbitrary sets of all known to us
abstraction-based heuristics.
2. Roberto Battiti and Paolo Campigotto, "Brain-machine optimization: learning
objective functions by interacting with the final user"
As a rule of thumb, ninety percent of the problem-solving effort in the real world is
spent on defining the problem, on specifying in a computable manner the function to
be optimized. After this modeling work is completed, optimization becomes in certain
cases a commodity. The implication is that much more research effort should be
devoted to design supporting techniques and tools to help the final user, often without
expertise in mathematics and in optimization, to define the function corresponding to
his real objectives. Machine learning can help, after the important context of
interfacing with a human user to provide feedback signals is taken into account, and
after the two tasks of modeling and solving are integrated into a loop so that an initial
rough definition can be progressively refined by bridging the representation gap
between the user world and the mathematical programming world.
To consider a more specific case, many real world optimization problems are
typically multi-objective optimization (MOO) problems. Efficient optimization
techniques have been developed for a wide range of MOO problems but, as
mentioned, asking a user to quantify the weights of the different objectives before
seeing the actual optimization results is in some cases extremely difficult. If the final
user is cooperating with an optimization expert, misunderstanding between the
persons may arise. This problem can become dramatic when the number of the
conflicting objectives in MOO increases. As a consequence, the final user will often
be dissatisfied with the solutions presented by the optimization expert. The
formalization of the optimization problem may be incomplete because some of the
objectives remain hidden in the mind of the final user.
Although providing explicit weights and mathematical formulas can be difficult for
the user, for sure he can evaluate the returned solutions. In most cases, the strategy to
overcome this situation consists of a joint iterative work between the final user and
the optimization expert to change the definition of the problem itself. The
optimization tool will then be re-executed over the new version of the problem.
We design a framework to automatically learn the knowledge that cannot be clearly
stated by the final user, i.e., the weights of the ``hidden" objectives. Each run of an
optimization tool provides a set of non-dominated solutions. The learning process we
investigate is based on the final user evaluation of the solution presented. Even if the
final user cannot state in a clear and aware manner the unsatisfied objectives, he can
select his favorite solutions among the ones provided by the optimization expert, after
the results are appropriately clustered and a suitably small number of questions is
answered.
In particular, our framework is based on the Reactive Search Optimization (RSO)
technique, which advocates the adoption of learning mechanisms as an integral part of
heuristic optimization schemes for solving complex optimization problems. The
ultimate goal of RSO is the complete elimination of the human intervention in the
parameters tuning process of heuristic algorithms. Our framework replaces the
optimization expert in the definition of the optimization task, by directly learning the
``hidden" objective functions from the final user evaluation of the solutions provided
by the optimization tools. The theoretical framework and some experimental results
will be presented in the extended version of this work.
3.
Lukasiak Piotr, Blazewicz Jacek, David Klatzmann, " CompuVac* development and standardized evaluation of novel genetic vaccines".
Recombinant viral vectors and virus-like particles are considered the most promising
vehicles to deliver antigens in prophylactic and therapeutic vaccines against infectious
diseases and cancer. Several potential vaccine designs exist but their cost-effective
development cruelly lacks a standardized evaluation system. On these grounds,
CompuVac is devoted to (i) rational development of a novel platform of genetic
vaccines and (ii) standardization of vaccine evaluation.
CompuVac assembles a platform of viral vectors and virus-like particles that are
among today’s most promising vaccine candidates and that are backed up by the
consortium’s complementary expertise and intellectual property, including SMEs
focusing on vaccine development.
CompuVac recognizes the lack of uniform means for side-by-side qualitative and
quantitative vaccine evaluation and will thus standardize the evaluation of vaccine
efficacy and safety by using “gold standard” tools, molecular and cellular methods in
virology and immunology, and algorithms based on genomic and proteomic
information.
“Gold standard” algorithms for intelligent interpretation of vaccine efficacy and safety
will be built into Compuvac’s interactive “Genetic Vaccine Decision Support
System”, which should generate (i) vector classification according to induced immune
response quality, accounting for gender and age, (ii) vector combination counsel for
prime-boost immunizations, and (iii) vector safety profile according to genomic
analysis.
Main objectives of CompuVac are:
•
to standardize the qualitative and quantitative evaluation of genetic vaccines
using defined “gold standard” antigens and methods
•
to rationally develop a platform of novel genetic vaccines using genomic and
proteomic information, together with our gold standards
•
to generate and make available to the scientific community a “tool box” and
an “interactive database” allowing to comparatively assess future vaccines to be
developed with our gold standards
Vector platform used in CompuVac is made of both viral vectors and VLPs that are
representative of and considered among the best in their class. Some, like the
adenoviruses, are already in clinical development. Within this consortium, we will
perform rational vector improvements and test rational prime boost combinations, all
guided by our comparative evaluation system. This should generate safer and more
efficient vectors, or vector combinations, for clinical development.
The consortium will generate an interactive platform called GeVaDS (Genetic
Vaccine Decision Support system). The platform will contain formatted data related
to defined gold standard antigens and methods used to assess immune responses i.e.
newly acquired results as well as algorithms allowing the intelligent comparison of
new vectors to previously analyzed ones.
CompuVac aims at making GeVaDS accessible to any researcher developing genetic
vaccines. Retrieving previously generated or introducing newly acquired results
obtained with validated approved methods, should allow any researcher to rationally
design and improve his or her vaccine vector as well as to comparatively assess its
efficacy and potential. Internet access:
http://www.compuvac.org (CompuVac),
http://gevads.cs.put.poznan.pl (GeVaDSs).
Monday
9:00-10:00 Plenary talk – Yefim Dinitz "Efficient algorithms for AND/OR (maxmin) scheduling problems on graphs"
The talk considers scheduling with AND/OR precedence constraints. The events are
represented as the vertices of a non-negatively weighted graph G=(V,E,d). The
precedence relation between them is given by its edges, and the delays by edge
weights. An event of type AND (max) should be scheduled anywhen after all its
preceding events, with the corresponding delays after them. An event of type OR
(min) should be scheduled anywhen after at least one of its preceding events, with the
corresponding delay after it. The early schedule is in question. This problem is a maxmin generalization of PERT. Besides, it may be considered as a mixture of the
Shortest Path and PERT problems.
We first consider the case when zero weight cycles are absent in G. Surprisingly,
some natural mixture of the classic Dijkstra and PERT algorithm works for it.
It solves the problem in time of two those algorithms, which is almost linear in the
graph size. After that, we present an O(|V||E|) algorithm solving the problem in the
general case, with zero weight cycles allowed.
We describe also a quite special history of the research on the AND/OR scheduling
problem. It began from a paper of D. Knuth in late 70th (presenting the algorithm, but
not the problem), went over particular cases and proofless publications, and involved
also names of Moehring, Adelson-Velsky, Levner, the lecturer, and others. Currently,
the first full paper on its general case is in preparation.
10:30- 11:45 Parallel sessions:
Track 1: Metaheuristics:
1. Johan Oppen, "Parametric models of local search progression"
Algorithms that search for good solutions to NP-hard combinatorial optimization
problems present a trace of current best objective values over time. The progression of
objective function values for best solutions found is reasonably modeled as stochastic
because the algorithms are often stochastic, and even when they are not, the
progression of the search varies with each instance in unpredictable ways. Some
characteristics are common to most searches: new bests are found quickly early in the
search and not as quickly later on.
We describe parametric models of this progression that are both interesting as ways to
characterize the search progression in a compact way and useful as means of
predicting search behavior. This is in contrast to non-parametric models that estimate
the probability that the search will achieve a value better than some given threshold in
a given amount of time. Both types of models have value, but parametric models offer
the promise of greater predictive power. For practical purposes, it is useful to be able
to use small instances to estimate parameters for models that can then be used to
predict the time trace performance on large instances.
2. Gregory Gutin and Daniel Karapetyan, "Local search heuristics for the
multidimensional assignment problem"
The Multidimensional Assignment Problem (MAP) (abbreviated s-AP in the case of s
dimensions) is an extension of the well-known assignment problem. The most
studied case of MAP is 3-AP, though the problems with larger values of s have also a
number of applications. We consider several known and new MAP local search
heuristics for MAP as well as their combinations and simple metaheuristics.
Computational experiments with three instance families are provided and discussed.
As a result, we select dominating local search heuristics. One of the most interesting
conclusions is that combination of two heuristics may yield a superior heuristic with
respect to both solution quality and the running time.
3. José Brandão, "A tabu search algorithm for the open vehicle routing problem
with time windows"
The problem studied here, the open vehicle routing problem with time windows
(OVRPTW), is different from the vehicle routing problem with time windows in that
the vehicles do not return to the distribution depot after delivering the goods to the
customers. We have solved the OVRPTW using a tabu search algorithm with
embedded local search heuristics, which take advantage of the specific properties of
the OVRP. The performance of the algorithm is tested using a large set of benchmark
problems.
Track 2: Networks:
1. Silvano Martello, "Jacobi, Koenig, Egerváry and the roots of combinatorial
optimization"
It is quite well-known that the first modern polynomial-time algorithm for the
assignment problem, invented by Harold W. Kuhn half a century ago, was christened
the "Hungarian method" to highlight that it derives from two older results, by Koenig
(1916) and Egerváry (1931). A recently discovered posthumous paper by Jacobi
(1804-1851) contains however a solution method that appears to be equivalent to the
Hungarian algorithm. A second historical result concerns a combinatorial
optimization problem, independently defined in satellite communication and in
scheduling theory, for which the same polynomial-time algorithm was independently
published thirty years ago by various authors. It can be shown that such algorithm
directly implements another result by Egerváry, which also implies the famous
Birkhoff-von Neumann theorem on doubly stochastic matrices.
2. Shoshana Anily and Aharona Pfeffer, " The uncapacitated swapping problem on
a line and on a circle"
We consider the uncapacitated swapping problem on a line and on a circle. Objects of
m types, which are initially positioned at n workstations on the graph, need to be
rearranged in order to satisfy the workstations’ requirements. Each workstation
initially contains one unit of a certain object type and requires one unit of possibly
another object type. We assume that the problem is balanced, i.e., the total supply
equals the total demand for each of the object types separately. A vehicle of unlimited
capacity is assumed to ship the objects in order to fulfill the requirements of all
workstations. The objective is to compute efficiently a shortest route such that the
vehicle can accomplish the rearrangement of the objects while following this route,
given designated starting and ending workstations on a line, or the location of a depot
on a circle.
We propose polynomial-time exact algorithms for solving the problems: an O(n)-time
algorithm for the linear track case, and an O(n2)-time algorithm for the circular track
case.
3. Susann Schrenk, Van-Dat Cung and Gerd Finke, "Revisiting the fixed charge
transportation problem"
The classical transportation problem is a special case of a minimum cost flow
problem with the property that the nodes of the network are either supply nodes or
demand nodes. The sum of all supplies and the sum of all demands are equal, and a
shipment between a supply node and a demand node induces a linear cost. The aim is
to determine a least cost shipment of a commodity through the network in order to
satisfy demands from available supplies. This classical transportation problem was
first presented by Hitchkock in 1941. In 1951 Dantzig proposed a standard
formulation as a linear program along with a first implementation of the simplex
method to solve it. This problem is polynomial.
The transportation problem including a fixed cost associated to each transportation
arc, named fixed charge transportation problem, is NP-Hard. This problem was first
introduced by Hirsch and Dantzig in 1954 and has largely been investigated in the
literature, Dantzig and Hirsch 1968, Sharma 1977, Balinski 1961. Most of papers until
then focused on approximation schemes to solve the fixed charge transportation
problem.
We study the special case, with no variable cost and the same fixed cost on all arcs.
This is in a way the simplest form of this problem, but we can show that this problem
remains NP-Hard. Also, minimizing the transportation costs is in this case equivalent
to having the maximum number of arcs on which there is no transport. That is
equivalent to finding a solution with a maximum degree of degeneracy. Since the
basic solutions are trees, solving the problem is equivalent to finding the tree with
maximum degree of degeneracy. Degeneracy occurs if one has a subset of the supply
nodes and a subset of the demand nodes whose total demand and total supply are
equal. We refer to this property as equal-subsum-problem. The existence of
degeneracy is an NP-Complete problem. Finding the maximum degree of degeneracy
is therefore also NP-Complete.
A new result in this context is the following: Finding the maximum degree of
degeneracy remains NP-Complete even if all equal subsums are given. The proof of
this theorem involves the maximum stable problem and the set packing problem.
12:00-12:50 Parallel sessions:
Track 1: Optimization 2:
1. Gur Mosheiov and Assaf Sarig, "Scheduling and due-date assignment problems
with job rejection"
Scheduling with rejection reflects a very common scenario, where the scheduler may
decide not to process a job if it is not profitable. We study the option of rejection in
several popular and well-known scheduling and due-date assignment problems. A
number of settings are considered: due-date and due-window assignment problems
with job-independent costs, a due-date assignment problem with job-dependent
weights and unit jobs, minimum total weighted earliness and tardiness cost with jobdependent and symmetric weights (known as TWET), and several classical
scheduling problems (minimum makespan, flow-time, earliness-tardiness) with
position-dependent processing times. All problems (excluding TWET) are shown to
have a polynomial time solution. For the (NP-hard) TWET, a pseudo-polynomial time
dynamic programming algorithm is introduced and tested numerically.
2. Ephraim Korach and Michal Stern, "TDIness in the clustering tree problem"
We consider the following problem: Given a complete graph G=(V,E) with a weight
on every edge and a given collection of subsets of V, we have to find a minimum
weight spanning tree T such that each subset of the vertices in the collection induces a
subtree in T. This problem arises and has motivation in the field of communication
networks. Consider the case where every subset is of size at most three. We present a
linear system that defines the polyhedron of all the feasible solutions. We prove that
this system is Total Dual Integrality (TDI). We present a conjecture for the general
case, where the size of each subset is not restricted to be less or equal to three. This
conjecture generalizes the above results.
Track 2: Convex combinatorial optimization problems:
1. Tiit Riismaa, "Application of convex extension of discrete-convex functions in
combinatorial optimization"
Convex extension of discrete-convex functions enables to adapt effective methods of
the convex programming for solving some discrete-convex programming problems of
hierarchy optimization. The n – dimensional real-valued function is called discreteconvex if the inequality of Jensen is valid for all convex combinations of (n+1) –
elements from the domain of definition. The use of all n+1 elements convex
combinations follows from the well-known theorem of Caratheodory .
The convex extension of given function is the majorant convex function not
exceeding the given function. Theorem. The function f can be extended to convex
function on conv X if f is discrete-convex on X.
The graph of a discrete-convex function is a part of the graph of a convex function.
The convex extension is so called point-wise maximum over all linear functions not
exceeding the given function. The convex function of a discrete-convex function is a
piecewise linear function. Each discrete-convex function has a unique convex
extension. The class of discrete-convex functions is the largest one to be extended to
the convex functions. A class of iteration methods of local searching is developed. On
each step of the iteration the calculation of the value of objective function is required
only on some vertices of some kind of unit cube. Many discrete or finite hierarchical
structuring problems can be formulated mathematically as a multi-level partitioning
procedure of a finite set of nonempty subsets. This partitioning procedure is
considered as a hierarchy where to the subsets of partitioning correspond nodes of
hierarchy and the relation of containing of subsets define the arcs of the hierarchy.
The feasible set of structures is a set of hierarchies corresponding to the full set of
multi-level partitioning of given finite set. Each tree from this set is represented by a
sequence of Boolean matrices, where each of these matrices is an adjacency matrix of
neighboring levels. The described formalism enables to state the reduced problem as a
two-phase mutually dependent optimization problem. Variable parameters of the inner
minimization problem are used for the description of connections between adjacent
levels. Variable parameters of the outer minimization problem are used for the
presentation of the number of elements on each level. The two-phase statement of
hierarchy optimization problem guarantees the possibility to extend the discreteconvex objective function to the convex function and enables to apply algorithm of
local searching for finding the global optimum. The approach is illustrated by a multilevel production system example.
2. Boaz Golany, Moshe Kress, Michal Penn and Uriel Rothblum, "Resource
allocation in temporary advantage competitions"
We consider a race between two competitors that develop competing products where
the advantage that is achieved by one competitor is limited in time until the opponent
is developing a new (better) product. We model the problem as network optimization
with side constraints and obtain some insights. We further briefly discuss the
stochastic case where the problems are modeled as (convex) non-linear optimization
problems.
Tuesday
9:00-10:15 Parallel sessions:
Track 1: Scheduling 3:
1. Liron Yedidsion, "Computational analysis of two fundamental problems in
MRP"
In this research we establish the computational complexity for two of the most basic
problem in the field of Material Requirement Planning (MRP). In particular, the
minimization of the total cost in the multi-echelon Bill of Material (BOM) system and
in the assembly system. Although both problems have been vastly studied since the
early 50s of the 20th century, their computational complexity has remained an open
question heretofore. We prove that both problems are strongly NP-hard, where the
minimization of the total cost in the multi-echelon BOM system is strongly NP-hard
even for the case where the Lead-Time of all the components is zero.
2. Sergey Sevastyanov and Bertrand M.T. Lin, "Efficient algorithm of generating all
optimal sequences for the two-machine Johnson problem"
For the classical two-machine Johnson problem we make it our aim to generate the set
of all optimal sequences of jobs (with respect to the minimum makespan objective),
assuming that on the set of those optimal sequences we could wish to optimize some
other (secondary) criterion.
Clearly, this purpose could be attained by enumerating all n! sequences of n jobs, and
this is the best possible bound for some instances (for example, for those consisting of
identical jobs). Yet for the majority of problem instances (having only a few optimal
solutions) this way cannot be recognized as acceptable, in view of its excessive timeconsumption.
We say that an algorithm pursuing the above mentioned purpose is efficient if its
running time can be estimated as O(P(|I|) |\Pi_{OPT}|), where P(|I|) is a polynomial of
the input size, while \Pi_{OPT} is the set of all optimal sequences of jobs. In other
words, if obtaining every optimal sequence requires (in average) at most a polynomial
time.
Such an efficient algorithm was designed in our paper. Its running time can be
estimated as O(n\log n|\Pi_{OPT}|). Therefore, in average we spend for generating
each new optimal sequence the same time as Johnson spent for his unique optimal
sequence. The algorithm is based on several lemmas containing basic properties of
optimal sequences.
3. Alexander Lazarev, "An approximate method for solving scheduling problems"
We consider the scheduling problems α|β|Fmax: A set of n jobs J1,…,Jn with release
dates r1,…,rn, processing times p1,…,pn and due dates d1,…,dn has to be scheduled on
a single or many machines. The job preemption is not allowed. The goal is to find a
schedule that minimizes the regular function F(C1,…,Cn), that Cj is the job j
completion time. We have suggest an approximation scheme to find approximate
optimal value of the objective function.
The approximation scheme: Let we have a instance A={rj,pj,dj} of our scheduling
problem. And for solving the instance we haven't sufficient computer resource. The
idea of the approach consists: we draw a line in 3n-dimensional space through a
point (instance) A. For example, {(rj,γpj,dj)|j\in N}, where γ \in [0,+ ∞). At γ =1 we
have initial instance A. Then on the given line "the reasonable" interval for Γ 1 and Γ2,
Γ1 < γ < Γ2, Γ1 < 1 < Γ2, is chosen. Usually at boundary points (instances) of an
interval the solution of scheduling problem is found for comprehensible time. Let H
be "limiting top" size of computer possibilities. For example, we choose values γ 1,…,
γk as roots of Chebyshev's polynomial on the interval $[ Γ1, Γ2], as they "nestle" on
the interval edges where we can find the solutions of the problem. For the instances
corresponding γ1,…, γk for which enough computing possibilities we find value of the
objective function F1,…,Fk. Then through points (γ1,F1),…, (γk,Fk) we construct
Lagrange's polynomial. Then the criterion function error Δ in our initial instance (with
γ =1) will not exceed size Δ ≤ C |(1- γ1)…(1- γk)|/(k+1)!, where C is constant. As a
result we do not construct the schedule of the solution of the problem, we only
estimate value of objective function. Than it is more points k, and the it is less
distance between Γ2 and Γ1 the less size Δ.
Track 2: Heuristics 1:
1. Mark Sh. Levin, "Combinatorial optimization problems in system configuration
and reconfiguration"
In recent years, systems configuration and reconfiguration problems are used as a
special significant part of design and management for many applications (e.g.,
software, hardware, manufacturing, communications, supply chain systems, solving
strategies, modular planning, combinatorial chemistry).
In the paper several basic system configuration/reconfiguration problems are
examined: (i) searching for (selection of) a set (structure) of system components, (ii)
searching for a set of compatible system components, (iii) allocation of system
components, (iv) design of system hierarchy, and (v) reconfiguration of a system
(e.g., change of system structure or system hierarchy).
The following combinatorial optimization problems are considered as underlying
models (including multicriteria formulations): problem of representatives,
multicriteria multiple choice problem, multicriteria allocation problem, graph coloring
and recoloring problems, morphological clique problem (with compatibility of system
components), multipartite clique or clustering of multipartite graph, hierarchical
clustering, and minimal spanning trees (including Steiner tree) problems. Solving
approaches and applications are discussed.
2. Aleksandra Swiercz, Jacek Blazewicz, Marek Figlerowicz, Piotr Gawron, Marta
Kasprzak, Darren Platt and Ɓukasz Szajkowski, "A new method for assembling of
DNA sequences"
Genome sequencing is the process of recovering a DNA sequence of a genome. It is
well known from its high complexity both on biological and computational levels.
The process is composed of a few phases: sequencing, assembling and finishing.
In the first phase one can obtain short DNA sequences of length up to a few hundreds
of nucleotides (in case of Sanger method). In case of new approaches to DNA
sequencing (e.g. 454 sequencing) shorter DNA sequences are obtained with relatively
good qualities and in short time. As the output of the 454 sequencing one gets
together with sequences (of length around 100-200 nucleotides), qualities, i.e. the
confidence rates of each nucleotide.
In the next phase, the assembling, short DNA sequences coming from the sequencing
phase are combined into a longer sequence or contiguous sequences, called contigs. In
the last phase, the finishing, which can be omitted in case of small genomes, the
contigs are put into the proper place of the chromosome.
The novel heuristic algorithm, SR-ASM, has been proposed for the assembly
problem, which deals well with data coming from the 454 sequencing. The algorithm
is based on a graph model. In the graph the input DNA sequences are on vertices, and
arcs connect two vertices, if two respective sequences overlap. The algorithm is
composed of three parts. In the first part the graph is constructed. At the beginning, a
fast heuristics looks for the pairs of sequences which possibly overlap, and marks
them as promising. Next, the score of the alignment is computed for every promising
pair, and an arc connecting the two respective vertices from the pair is added to the
graph with the calculated score. In the second part one searches for a path in the graph
which passes through the most number of vertices. Usually, it is not possible to find
one path due to errors in the input data. Thus instead of one path several paths are
returned. In the last part of the algorithm, consensus sequences are constructed from
the paths in the graph on the base of multiple alignment procedure.
The usefulness of our algorithm has been tested on raw data coming from the
experiment on sequencing of the genome of the Prochlorococcus marinus bacteria
performed at the Joint Genome Institute. As the output of the SR-ASM algorithm one
gets small number of long contigs which are highly similar to the genome of the
bacteria. The results are compared with the outputs of other available assemblers.
10:45-12:00 Parallel sessions:
Track 1: Graphs:
1. Dietmar Cieslik, "The Steiner ratio"
Steiner's Problem is the "Problem of shortest connectivity", that means, given a finite
set of points in a metric space X, search for a network interconnecting these points
with minimal length. This shortest network must be a tree and is called a Steiner
Minimal Tree (SMT). It may contain vertices different from the points which are to be
connected. Such points are called Steiner points. If we do not allow Steiner points,
that means, we only connect certain pairs of the given points, we get a tree which is
called a Minimum Spanning Tree (MST). Steiner's Problem is very hard as well in
combinatorial as in computational sense, but, on the other hand, the determination of
an MST is simple. Consequently, we are interested in the greatest lower bound for the
ratio between the lengths of these trees, which is called the Steiner ratio (of the space
X). We look for estimates and exact values for the Steiner ratio in several metric
spaces, given old and new results.
2. Martin Charles Golumbic and Robert E. Jamison, "Tolerance and RankTolerance in Graphs: Mathematical and Algorithmic Problems"
In this talk, we set out a plan for future investigation of two global themes which
focus on special families and properties of graphs.
These involve the notion of measured intersection known as tolerance and a more
general framework called rank-tolerance. Tolerance and rank-tolerance graphs were
originally motivated by work spawned by the classical family of interval intersection
graphs.
In a rank-tolerance representation of a graph, each vertex is assigned two parameters:
a rank, which represents the size of that vertex, and a tolerance which represents an
allowed extent of conflict with other vertices. Two vertices are adjacent if and only if
their joint rank exceeds (or equals) their joint tolerance. By varying the coupling
functions used to obtain the joint rank or joint tolerance, a variety of graph classes
arise, many of which have interesting structure.
Applications, algorithms and their complexity have driven much of the research work
in these areas. The research problems posed here provide challenges to those
interested in structured families of graphs.
These include trying to characterize the class SP of sum-product graphs, where the
tolerance coupling function is the sum} of the two tolerances and the rank coupling
function is the product of the two ranks, and investigating the class of mix graphs,
which are obtained by interpolating between the min and max functions.
3. Peter Adams, Moshe Rosenfeld and Hoa Vu Dinh, "Spanning graph designs"
Classical balanced incomplete block designs (BIBD) explore decompositions of the
edges of the complete graph K_n into "blocks" (complete subgraphs) of prescribed
size(s) so that every edge of K_n appears in the same number of blocks. A natural
generalization of these designs is to extend "blocks" to other graphs. For example,
decompositions of K_n into cycles has been studied by many authors. But the most
general case, the Oberwolfach Problem is still open.
Of particular interest are spanning graph designs where the blocks are graph on n
vertices. We study two groups of spanning graph designs: spanning cubic graph
designs and spanning designs inspired by equiangular lines in R^d.
For a cubic graph G of order n to decompose K_n it is necessary that n = 6k+4. We
show various constructions of such decompositions and conjecture that for n 'large
enough" all cubic graphs of order 6n + 4 decompose k-{6n+4}.
The second group of spanning graph designs involve G(2n, n-1). These are graphs of
order 2n regular of degree n - 1. The implied spanning graph design will yield 2n-1
copies of G so that every edge of K_{2n} will appear in exactly n-1 copies of G. For
instance, one can construct 11 labeled copies of the Icosahedron (a G(12,5)) so that
every edge of K_12 will appear in exactly 5 copies. These designs turn out to be
related to Hadamard matrices, conference matrices and other interesting combinatorial
designs.
Track 2: Heuristics 2:
1. Michal Penn and Eli Troupiansky, " A heuristic algorithm for solving the
heterogeneous open vehicle routing problem"
Given a capacitated directed graph with a depot vertex and demand vertices the VRP
(Vehicle Routing Problem) is aimed at minimizing the cost of supplying the required
demand. That is, minimizing the number of capacitated vehicles needed and the cost
of the tour. The problem is known to be NP-hard and vast of the research effort is
aimed at various heuristic methods for solving the problem.
We consider a variant of the VRP, termed Open VRP (OVRP), where the
vehicles are not required to return to the depotss at the end of their tours. If, in
addition, the vehicles are of different capacities and costs, the problem is denoted as
Heterogeneous OVRP.
We present a greedy heuristic for solving the Heterogeneous OVRP. Our
greedy algorithm consists of two stages – construction of an initial feasible solution
and post-optimization for improving the initial solution. We compare our heuristic to
some known algorithms for the homogeneous case and to Cross-Entropy (CE) in the
general heterogeneous case. Our preliminary results show that the greedy heuristic out
performs CE.
2. Dorit Ron, Ilya Safro and Achi Brandt, "Fast multilevel algorithms for linear and
planar ordering problems"
The purpose of linear ordering problems is to minimize some functional over all
possible permutations. These problems are widely used and studied in many practical
and theoretical applications. We have developed linear-time multilevel solvers for a
variety of such graph problems, e.g., the linear arrangement problem, the bandwidth
problem and more. In a multilevel framework, a hierarchy of decreasing size graphs is
constructed. Starting from the given graph G0, create by coarsening the sequence
G1,...,Gk, then solve the coarsest level directly, and finally uncoarsen (interpolate) the
solution back to G0. The experimental results for such problems turned out to be
better than every known result in almost all cases, while the short running time of the
algorithms enables applications on very large graphs.
Next we have considered the generalization of such problems to two dimensions. In
many theoretical and industrial fields, this class of problems is often addressed and
actually poses a computational bottleneck, e.g., graph visualization, facility location
problem, VLSI layout, etc. In particular, we have tried to minimize various
functionals (discrete or continuous) applied to the two-dimensional objects
(associated with the graph nodes) at hand, while maintaining equi-density demand
over the planar domain. We have concentrated on answering the following three
issues: (a) minimize the total length of the connections between these objects (b)
minimize the overlap between the objects and (c) the two-dimensional space should
be well utilized.
We present a multilevel solver for a model that describes the core part of those
applications, namely, the problem of minimizing a quadratic energy functional under
planar constraints that bound the allowed amount of material (total areas of objects) in
various subdomains of the entire domain under consideration. Given an initial
arrangement for a particular graph Gi in the hierarchy, rearrange the objects under
consideration into a more evenly distributed state over the entire defined domain. This
process is done by introducing a sequence of finer and finer grids over the domain and
demanding at each scale equi-density, that is, meeting equality or inequality
constraints at each grid square, stating how much material it may (at most) contain.
Since many variables are involved and since the needed updates may be large, we
introduce a new set of displacement variables attached to the introduced grid points,
which enables collective moves of many original variables at a time, at different
scales including large displacements. The use of such multiscale moves has two main
purposes: to enable processing in various scales and to efficiently solve the (large)
system of equations of energy minimization under equi-density demands. The system
of equations of the finer scales, when many unknowns are involved, is solved by a
combination of well-known multigrid techniques. The entire algorithm solves the
nonlinear minimization problem by applying successive steps of corrections, each
using a linearized system of equations.
We demonstrate the performance of our solver on some instances of the graph
visualization problem showing efficient usage of the given domain, and on a set of
simple (artificial) placement (VLSI) examples.
3. L.F. Escudero and S. Muñoz, "A greedy procedure for solving the line design
problem for a rapid transit network"
Given a set of potential station locations and a set of potential links between them, the
well-known extended rapid transit network design problem basically consists in
selecting which stations and links to construct without exceeding the available budget,
and determining an upper bounded number of noncircular lines from them, to
maximise the total expected number of users.
A modification of this problem has been proposed in the literature to allow the
definition of circular lines provided that whichever two locations are linked by one
line at most, and a two-stage approach has also been presented for solving this new
problem. The model considered in the first stage makes possible to select the stations
and links to be constructed without exceeding the available budget, in such a way that
the total expected number of users is maximised. Once they have been selected, in the
second stage each one of these links is assigned to a unique line, so that the number of
lines passing through each selected station is minimised.
In this work we introduce an improvement in the model considered in the first stage
above to obtain a connected rapid transit network. We also present a modification of
the algorithm proposed for solving the line design problem of the second stage above;
it is a greedy heuristic procedure that attempts to minimise the number of transfers
that should be done by the users to arrive at their destinations, without increasing the
number of lines that pass through each station.
The computational experiments that will be reported will show that this greedy
procedure can significatively reduce the total number of transfers required for the
solutions obtained by the approach taken from the literature.
12:10-13:10 plenary talk – Julian Molina "Metaheuristics for multi-objective
combinatorial optimization"
Meta-heuristic methods are widely used to solve single-objective combinatorial
optimization problems, especially when dealing with real applications. However, a
multiple criteria approach is usually demanded when solving real-world applications
requiring the consideration of several conflicting points of view corresponding to
multiple objectives.
As it happens within the field of single-objective combinatorial optimization, metaheuristics have shown to be efficient for solving multi-objective optimization models.
It is one of the most active and growing branching in the field of multi-objective
optimisation in current years.
This talk provides an overview of the research in this field and visits its principal
aspects, including main methods, performance measures, test functions, and
preference inclusion techniques. New trends in the field will be outlined, including
new hybrid procedures and links to exact techniques.
Download