Column generation
Страница в процессе перевода. |
This tutorial was generated using Literate.jl. Download the source as a .jl
file.
The purpose of this tutorial is to demonstrate the column generation algorithm. As an example, it solves the Cutting stock problem.
This tutorial uses the following packages:
using JuMP
import DataFrames
import HiGHS
import Plots
import SparseArrays
Background
The cutting stock problem is about cutting large rolls of paper into smaller pieces.
We denote the set of possible sized pieces that a roll can be cut into by . Each piece has a width, , and a demand, . The width of the large roll is .
Our objective is to minimize the number of rolls needed to meet all demand.
Here’s the data that we are going to use in this tutorial:
struct Piece
w::Float64
d::Int
end
struct Data
pieces::Vector{Piece}
W::Float64
end
function Base.show(io::IO, d::Data)
println(io, "Data for the cutting stock problem:")
println(io, " W = $(d.W)")
println(io, "with pieces:")
println(io, " i w_i d_i")
println(io, " ------------")
for (i, p) in enumerate(d.pieces)
println(io, lpad(i, 4), " ", lpad(p.w, 5), " ", lpad(p.d, 3))
end
return
end
function get_data()
data = [
75.0 38
75.0 44
75.0 30
75.0 41
75.0 36
53.8 33
53.0 36
51.0 41
50.2 35
32.2 37
30.8 44
29.8 49
20.1 37
16.2 36
14.5 42
11.0 33
8.6 47
8.2 35
6.6 49
5.1 42
]
return Data([Piece(data[i, 1], data[i, 2]) for i in axes(data, 1)], 100.0)
end
data = get_data()
Data for the cutting stock problem:
W = 100.0
with pieces:
i w_i d_i
------------
1 75.0 38
2 75.0 44
3 75.0 30
4 75.0 41
5 75.0 36
6 53.8 33
7 53.0 36
8 51.0 41
9 50.2 35
10 32.2 37
11 30.8 44
12 29.8 49
13 20.1 37
14 16.2 36
15 14.5 42
16 11.0 33
17 8.6 47
18 8.2 35
19 6.6 49
20 5.1 42
Mathematical formulation
To formulate the cutting stock problem as a mixed-integer linear program, we assume that there is a set of large rolls to use. Then, we introduce two classes of decision variables:
is a binary variable that indicates if we use roll , and counts how many pieces of size that we cut from roll .
Our mixed-integer linear program is therefore:
The objective is to minimize the number of rolls that we use, and the two constraints ensure that we respect the total width of each large roll and that we satisfy demand exactly.
The JuMP formulation of this model is:
I = length(data.pieces)
J = 1_000 # Some large number
model = Model(HiGHS.Optimizer)
set_silent(model)
@variable(model, x[1:I, 1:J] >= 0, Int)
@variable(model, y[1:J], Bin)
@objective(model, Min, sum(y))
@constraint(model, [i in 1:I], sum(x[i, :]) >= data.pieces[i].d)
@constraint(
model,
[j in 1:J],
sum(data.pieces[i].w * x[i, j] for i in 1:I) <= data.W * y[j],
);
Unfortunately, we can’t solve this formulation for realistic instances because it takes a very long time to solve. (Try removing the time limit.)
set_time_limit_sec(model, 5.0)
optimize!(model)
solution_summary(model)
* Solver : HiGHS
* Status
Result count : 1
Termination status : TIME_LIMIT
Message from the solver:
"kHighsModelStatusTimeLimit"
* Candidate solution (result #1)
Primal status : FEASIBLE_POINT
Dual status : NO_SOLUTION
Objective value : 4.11000e+02
Objective bound : 2.93000e+02
Relative gap : 2.87105e-01
* Work counters
Solve time (sec) : 5.07014e+00
Simplex iterations : 17513
Barrier iterations : -1
Node count : 0
However, there is a formulation that solves much faster, and that is to use a column generation scheme.
Column generation theory
The key insight for column generation is to recognize that feasible columns in the matrix of variables encode cutting patterns.
For example, if we look only at the roll , then a feasible solution is:
-
(1 unit of piece #1)
-
(1 unit of piece #13)
-
All other
Another solution is
-
(19 unit of piece #20)
-
All other
Cutting patterns like and are infeasible because the combined length is greater than .
Since there are a finite number of ways that we could cut a roll into a valid cutting pattern, we could create a set of all possible cutting patterns , with data indicating how many units of piece we cut in pattern . Then, we can formulate our mixed-integer linear program as:
Unfortunately, there will be a very large number of these patterns, so it is often intractable to enumerate all columns .
Column generation is an iterative algorithm that starts with a small set of initial patterns, and then cleverly chooses new columns to add to the main MILP so that we find the optimal solution without having to enumerate every column.
Choosing the initial set of patterns
For the initial set of patterns, we create a trivial cutting pattern which cuts as many units of piece as will fit.
patterns = map(1:I) do i
n_pieces = floor(Int, data.W / data.pieces[i].w)
return SparseArrays.sparsevec([i], [n_pieces], I)
end
20-element Vector{SparseArrays.SparseVector{Int64, Int64}}:
[1] = 1
[2] = 1
[3] = 1
[4] = 1
[5] = 1
[6] = 1
[7] = 1
[8] = 1
[9] = 1
[10] = 3
[11] = 3
[12] = 3
[13] = 4
[14] = 6
[15] = 6
[16] = 9
[17] = 11
[18] = 12
[19] = 15
[20] = 19
We can visualize the patterns as follows:
"""
cutting_locations(data::Data, pattern::SparseArrays.SparseVector)
A function which returns a vector of the locations along the roll at which to
cut in order to produce pattern `pattern`.
"""
function cutting_locations(data::Data, pattern::SparseArrays.SparseVector)
locations = Float64[]
offset = 0.0
for (i, c) in zip(SparseArrays.findnz(pattern)...)
for _ in 1:c
offset += data.pieces[i].w
push!(locations, offset)
end
end
return locations
end
function plot_patterns(data::Data, patterns)
plot = Plots.bar(;
xlims = (0, length(patterns) + 1),
ylims = (0, data.W),
xlabel = "Pattern",
ylabel = "Roll length",
)
for (i, p) in enumerate(patterns)
locations = cutting_locations(data, p)
Plots.bar!(
plot,
fill(i, length(locations)),
reverse(locations);
bar_width = 0.6,
label = false,
color = "#90caf9",
)
end
return plot
end
plot_patterns(data, patterns)
The base problem
Using the initial set of patterns, we can create and optimize our base model:
model = Model(HiGHS.Optimizer)
set_silent(model)
@variable(model, x[1:length(patterns)] >= 0, Int)
@objective(model, Min, sum(x))
@constraint(model, demand[i in 1:I], patterns[i]' * x >= data.pieces[i].d)
optimize!(model)
@assert is_solved_and_feasible(model)
solution_summary(model)
* Solver : HiGHS
* Status
Result count : 1
Termination status : OPTIMAL
Message from the solver:
"kHighsModelStatusOptimal"
* Candidate solution (result #1)
Primal status : FEASIBLE_POINT
Dual status : NO_SOLUTION
Objective value : 4.21000e+02
Objective bound : 4.21000e+02
Relative gap : 0.00000e+00
* Work counters
Solve time (sec) : 1.93834e-04
Simplex iterations : 0
Barrier iterations : -1
Node count : 0
This solution requires 421 rolls. This solution is sub-optimal because the model does not contain the full set of possible patterns.
How do we find a new column that leads to an improved solution?
Choosing new columns
Column generation chooses a new column by relaxing the integrality constraint on and looking at the dual variable associated with demand constraint .
For example, the dual of demand[13]
is:
unset_integer.(x)
optimize!(model)
@assert is_solved_and_feasible(model; dual = true)
π_13 = dual(demand[13])
0.25
Using the economic interpretation of the dual variable, we can say that a one unit increase in demand for piece will cost an extra rolls. Alternatively, we can say that a one unit increase in the left-hand side (for example, due to a new cutting pattern) will save us rolls. Therefore, we want a new column that maximizes the savings associated with the dual variables, while respecting the total width of the roll:
If this problem, called the pricing problem, has an objective value greater than , then we estimate than adding y
as the coefficients of a new column will decrease the objective by more than the cost of an extra roll.
Here is code to solve the pricing problem:
function solve_pricing(data::Data, π::Vector{Float64})
I = length(π)
model = Model(HiGHS.Optimizer)
set_silent(model)
@variable(model, y[1:I] >= 0, Int)
@constraint(model, sum(data.pieces[i].w * y[i] for i in 1:I) <= data.W)
@objective(model, Max, sum(π[i] * y[i] for i in 1:I))
optimize!(model)
@assert is_solved_and_feasible(model)
number_of_rolls_saved = objective_value(model)
if number_of_rolls_saved > 1 + 1e-8
# Benefit of pattern is more than the cost of a new roll plus some
# tolerance
return SparseArrays.sparse(round.(Int, value.(y)))
end
return nothing
end
solve_pricing (generic function with 1 method)
If we solve the pricing problem with an artificial dual vector:
solve_pricing(data, [1.0 / i for i in 1:I])
20-element SparseArrays.SparseVector{Int64, Int64} with 3 stored entries:
[1 ] = 1
[17] = 1
[20] = 3
the solution is a roll with 1 unit of piece #1, 1 unit of piece #17, and 3 units of piece #20.
If we solve the pricing problem with a dual vector of zeros, then the benefit of the new pattern is less than the cost of a roll, and so the function returns nothing
:
solve_pricing(data, zeros(I))
Iterative algorithm
Now we can combine our base model with the pricing subproblem in an iterative column generation scheme:
while true
# Solve the linear relaxation
optimize!(model)
@assert is_solved_and_feasible(model; dual = true)
# Obtain a new dual vector
π = dual.(demand)
# Solve the pricing problem
new_pattern = solve_pricing(data, π)
# Stop iterating if there is no new pattern
if new_pattern === nothing
@info "No new patterns, terminating the algorithm."
break
end
push!(patterns, new_pattern)
# Create a new column
push!(x, @variable(model, lower_bound = 0))
# Update the objective coefficient of the new column
set_objective_coefficient(model, x[end], 1.0)
# Update the non-zeros in the coefficient matrix
for (i, count) in zip(SparseArrays.findnz(new_pattern)...)
set_normalized_coefficient(demand[i], x[end], count)
end
println("Found new pattern. Total patterns = $(length(patterns))")
end
Found new pattern. Total patterns = 21
Found new pattern. Total patterns = 22
Found new pattern. Total patterns = 23
Found new pattern. Total patterns = 24
Found new pattern. Total patterns = 25
Found new pattern. Total patterns = 26
Found new pattern. Total patterns = 27
Found new pattern. Total patterns = 28
Found new pattern. Total patterns = 29
Found new pattern. Total patterns = 30
Found new pattern. Total patterns = 31
Found new pattern. Total patterns = 32
Found new pattern. Total patterns = 33
Found new pattern. Total patterns = 34
Found new pattern. Total patterns = 35
[ Info: No new patterns, terminating the algorithm.
We found lots of new patterns. Here’s pattern 21:
patterns[21]
20-element SparseArrays.SparseVector{Int64, Int64} with 3 stored entries:
[9 ] = 1
[13] = 2
[17] = 1
Let’s have a look at the patterns now:
plot_patterns(data, patterns)
Looking at the solution
Let’s see how many of each column we need:
solution = DataFrames.DataFrame([
(pattern = p, rolls = value(x_p)) for (p, x_p) in enumerate(x)
])
filter!(row -> row.rolls > 0, solution)
Row | pattern | rolls |
---|---|---|
Int64 | Float64 | |
1 | 1 | 38.0 |
2 | 2 | 44.0 |
3 | 3 | 30.0 |
4 | 21 | 0.5 |
5 | 22 | 10.2 |
6 | 23 | 14.65 |
7 | 24 | 23.1 |
8 | 25 | 11.25 |
9 | 26 | 21.35 |
10 | 28 | 4.3 |
11 | 29 | 19.55 |
12 | 30 | 11.25 |
13 | 31 | 17.45 |
14 | 33 | 36.0 |
15 | 34 | 11.4 |
16 | 35 | 41.0 |
Since we solved a linear program, some of our columns have fractional solutions. We can create a integer feasible solution by rounding up the orders. This requires 341 rolls:
sum(ceil.(Int, solution.rolls))
341
Alternatively, we can re-introduce the integrality constraints and resolve the problem:
set_integer.(x)
optimize!(model)
@assert is_solved_and_feasible(model)
solution = DataFrames.DataFrame([
(pattern = p, rolls = value(x_p)) for (p, x_p) in enumerate(x)
])
filter!(row -> row.rolls > 0, solution)
Row | pattern | rolls |
---|---|---|
Int64 | Float64 | |
1 | 1 | 38.0 |
2 | 2 | 44.0 |
3 | 3 | 30.0 |
4 | 21 | 1.0 |
5 | 22 | 9.0 |
6 | 23 | 19.0 |
7 | 24 | 19.0 |
8 | 25 | 13.0 |
9 | 26 | 17.0 |
10 | 28 | 2.0 |
11 | 29 | 19.0 |
12 | 30 | 13.0 |
13 | 31 | 18.0 |
14 | 33 | 36.0 |
15 | 34 | 15.0 |
16 | 35 | 41.0 |
This now requires 334 rolls:
sum(solution.rolls)
333.99999999999994
Note that this may not be the global minimum because we are not adding new columns during the solution of the mixed-integer problem model
(an algorithm known as branch and price). Nevertheless, the column generation algorithm typically finds good integer feasible solutions to an otherwise intractable optimization problem.