Abstract
In this paper, we present a new perspective on cut generation in the context of Benders decomposition. The approach, which is based on the relation between the alternative polyhedron and the reverse polar set, helps us to improve established cut selection procedures for Benders cuts, like the one suggested by Fischetti et al. (Math Program Ser B 124(1–2):175–182, 2010). Our modified version of that criterion produces cuts which are always supporting and, unless in rare special cases, facetdefining. We discuss our approach in relation to the state of the art in cut generation for Benders decomposition. In particular, we refer to Paretooptimality and facetdefining cuts and observe that each of these criteria can be matched to a particular subset of parametrizations for our cut generation framework. As a consequence, our framework covers the method to generate facetdefining cuts proposed by Conforti and Wolsey (Math Program Ser A 178:1–20, 2018) as a special case. We conclude the paper with a computational evaluation of the proposed cut selection method. For this, we use different instances of a capacity expansion problem for the european power system.
Introduction
Consider a generic optimization problem with two subsets of variables x and y where x is restricted to lie in some set \(S \subseteq \mathbb {R}^n\) and x and y are jointly constrained by a set of m linear inequalities. Such a problem can be written in the following form:
The interaction matrix \(H \in \mathbb {R}^{m \times n}\) captures the influence of the xvariables on the ysubproblem: For fixed \(x^* \in \mathbb {R}^n\), (1.1) reduces to an ordinary linear program with constraints \(Ay \le bHx^*\), where \(A \in \mathbb {R}^{m \times k}, y \in \mathbb {R}^k\), and \(b  Hx^* \in \mathbb {R}^m\).
We are interested in cases where the size of the complete problem (1.1) leads to infeasibly high computation times (or memory demands), but both the problem over S and the problem resulting from fixing x can separately be solved much more efficiently due to their special structures. To deal with such problems, Benders (1962) introduced a method that works by iterating between these two “easier” problems:
For a problem of the form (1.1), let the function \(z: \mathbb {R}^n \rightarrow \mathbb {R}\cup \left\{ {\pm \infty }\right\} \) represent the value of the optimal ypart of the objective function for a fixed vector x:
The corresponding epigraph of \(z\) is
Writing \({{\,\mathrm{epi}\,}}_S(z) := {{\,\mathrm{epi}\,}}(z) \cap (S \times \mathbb {R})\), this provides us with an alternative representation of the optimization problem (1.1):
This representation suggests the following iterative algorithm: Start by finding a solution \((x^*,\eta ^*) \in S \times \mathbb {R}\) that minimizes \(c^\top x + \eta \) without any additional constraints (adding a generous lower bound for \(\eta \) to make the problem bounded). If \((x^*,\eta ^*) \in {{\,\mathrm{epi}\,}}(z)\), then \((x^*,\eta ^*) \in {{\,\mathrm{epi}\,}}_S(z)\) (since \(x^* \in S\)) and the solution is optimal. Otherwise, we add constraints violated by \((x^*,\eta ^*)\) but satisfied by all \((x',\eta ') \in {{\,\mathrm{epi}\,}}(z)\) and iterate. This is of course just an ordinary cutting plane algorithm and the crucial question is how to select a separating inequality in each iteration.
The original Benders algorithm uses feasibility cuts (cuts with coefficient 0 for the variable \(\eta \)) and optimality cuts (cuts with nonzero coefficient for the variable \(\eta \)), depending on whether or not the subproblem that results from fixing the xvariables is feasible (see, e. g., Vanderbeck and Wolsey 2010). Fischetti et al. (2010), on the other hand, present a unified perspective that covers both cases: They begin by observing that the subproblem can be seen as a pure feasibility problem, represented by the set
This polyhedron will be empty if and only if \((x^*,\eta ^*) \notin {{\,\mathrm{epi}\,}}(z)\) and any Farkas certificate for emptiness of (1.4) can be used to derive an additional valid inequality. The set of such certificates (up to positive scaling)
is called alternative polyhedron.^{Footnote 1} Thus \(P(x^*,\eta ^*) = \emptyset \) if and only if \((x^*,\eta ^*) \in {{\,\mathrm{epi}\,}}(z)\) and every point \((\gamma , \gamma _0) \in P(x^*,\eta ^*)\) induces an inequality \(\gamma ^\top (bHx)+\gamma _0\eta \ge 0\) that is valid for \({{\,\mathrm{epi}\,}}(z)\) but violated by \((x^*,\eta ^*)\).
This characterization is very useful and has been demonstrated empirically to work well in Fischetti et al. (2010). However, it exposes some fundamental issues, which are demonstrated by the following example (Fig. 1).
Example 1.1
Consider the following optimization problem:
Note that the constraint \(4x4y \le 14\) is redundant and does not support the feasible region. Suppose that we want to decompose the problem into its xpart and its ypart.
Writing the components of \((\gamma ,\gamma _0)\) in order \((\gamma _1,\gamma _2,\gamma _3, \gamma _0)\), we obtain
The three vertices of P(0, 0) are
As Gleeson and Ryan (1990) showed, each of these points corresponds to a minimal infeasible subsystem of (1.6) with the objective function written in inequality form \(x+y \le 0\). Consequently, each vertex yields one of the original inequalities as a cut. This notably includes the redundant inequality \(4x4y \le 14\), which does not support the feasible region but is derived from the vertex \(P_3\) in the alternative polyhedron (and furthermore minimizes the linear objective \(\mathbbm {1}\)).
A cut generated from a point in the alternative polyhedron may thus be very weak, not even supporting the set \({{\,\mathrm{epi}\,}}(z)\). This is true even if we use a vertex of the alternative polyhedron and even if that vertex minimizes a given linear objective such as the vector \(\mathbbm {1}\) as suggested in Fischetti et al. (2010).
In the following, we present an improved approach for cut generation in the context of Benders decomposition. Our method can be parametrized by the selection of an objective vector in primal space and produces facet cuts without any additional computational effort for all but a subdimensional set of parametrizations. In addition, our method is more robust with respect to the formulation of the problem than the original approach from Fischetti et al. (2010). In particular it always generates supporting cuts, avoiding the problem pointed out in the context of Example 1.1.
Our method is based on the relation between the alternative polyhedron as introduced above, which is commonly used in the context of Benders cut generation, and the reverse polar set, originally introduced by Balas and Ivanescu (1964) in the context of transportation problems.
We show that the alternative polyhedron can be viewed as an extended formulation of the reverse polar set, providing us with a parametrizable method to generate cuts with different wellknown desirable properties, most notably facetdefining cuts. As a special case, we obtain an (arguably simpler) alternative proof for the method to generate facetdefining cuts proposed by Conforti and Wolsey (2018), if applied to Benders decomposition. Our work links their approach more directly to previous work on cut selection, within the context of Benders decomposition (e. g., Fischetti et al. (2010)) as well as more generally for separation of convex sets (e. g., Cornuéjols and Lemaréchal (2006)).
Before we proceed by investigating different representations of the set of possible Benders cuts, it is useful to record a general characterization of the set of normal vectors for cuts separating a point from \({{\,\mathrm{epi}\,}}(z)\) as defined in (1.3). In the following, we say that a halfspace \(H^{\le }_{(\pi ,\alpha )} := \left\{ {x \in \mathbb {R}^n {\bigg \vert }\pi ^\top x \le \alpha }\right\} \) is xseparating for a convex set \(C \subset \mathbb {R}^n\) and a point \(x \in \mathbb {R}^n \setminus C\) if \(x \notin H^{\le }_{(\pi ,\alpha )} \supset C\).
Theorem 1.2
Let \(z\) be defined as in (1.2) such that \({{\,\mathrm{epi}\,}}(z) \!\ne \! \emptyset \) and let \((x^*,\eta ^*),(\pi , \pi _0)\! \in \! \mathbb {R}^n \times \mathbb {R}\). Then \((\pi ,\pi _0)\) is the normal vector of a \((x^*,\eta ^*)\)separating halfspace for \({{\,\mathrm{epi}\,}}(z)\) if and only if there exists a vector \(\gamma \in \mathbb {R}_{\ge 0}^m\) satisfying
Proof
Let \(h_{{{\,\mathrm{epi}\,}}(z)}(\pi ,\pi _0) := \sup \left\{ {\pi ^\top x + \pi _0 \eta {\bigg \vert }(x,\eta ) \in {{\,\mathrm{epi}\,}}(z)}\right\} \) be the support function of \({{\,\mathrm{epi}\,}}(z)\) evaluated at \((\pi ,\pi _0)\). The vector \((\pi ,\pi _0)\) is the normal vector of a \((x^*,\eta ^*)\)separating halfspace for \({{\,\mathrm{epi}\,}}(z)\) if and only if
By the definition of \({{\,\mathrm{epi}\,}}(z)\) (which is closed and polyhedral) and then by strong LP duality, we obtain
Note that in order for the equality \(\gamma _0=\pi _0\) to hold and (1.13) to be feasible (and hence (1.12) to be bounded), we need that \(\pi _0 \le 0\). Thus the optimality of any pair \((\gamma ,\gamma _0)\) for (1.13) is equivalent to to the fulfillment of conditions (1.7) to (1.10). \(\square \)
As one can see from the proof above, any \(\gamma \) satisfying (1.8) to (1.10) is an upper bound for \(h_{{{\,\mathrm{epi}\,}}(z)}\). This means that given a certificate \(\gamma \) to prove that \((\pi ,\pi _0)\) is a normal vector of an \((x^*,\eta ^*)\)separating halfspace \(H^{\le }_{((\pi ,\pi _0),\alpha )}\), we immediately obtain a corresponding right hand side \(\alpha := \gamma ^\top b\). Furthermore, the definition of the support function \(h_{{{\,\mathrm{epi}\,}}(z)}\) immediately tells us when this righthand side is actually optimal and the resulting halfspace supports \({{\,\mathrm{epi}\,}}(z)\):
Remark 1.1
Let \((x^*,\eta ^*) \in \mathbb {R}^n \times \mathbb {R}\) and let \((\pi ,\pi _0)\) be the normal vector of an \((x^*,\eta ^*)\)separating halfspace for \({{\,\mathrm{epi}\,}}(z)\). If \(\gamma \) minimizes \(\gamma ^\top b\) among all possible certificates in Theorem 1.2, then the halfspace \(H^{\le }_{((\pi ,\pi _0), \gamma ^\top b)}\) supports the set \({{\,\mathrm{epi}\,}}(z)\).
Benders cuts from the reverse polar set
While it would be sufficient in the context of Benders decomposition to obtain an arbitrary \((x^*,\eta ^*)\)separating halfspace whenever the set in (1.4) is empty, the alternative polyhedron \(P(x^*,\eta ^*)\) actually completely characterizes the set of all possible normal vectors of such halfspaces:
Corollary 2.1
The alternative polyhedron (1.5) completely characterizes all normal vectors of \((x^*,\eta ^*)\)separating halfspaces for \({{\,\mathrm{epi}\,}}(z)\). In particular:

(a)
Let \((\gamma ,\gamma _0) \in P(x^*,\eta ^*)\). Then \(\gamma ^\top Hx \gamma _0 \eta \le \gamma ^\top b\) is violated by \((x^*,\eta ^*)\), but satisfied by all \((x,\eta ) \in {{\,\mathrm{epi}\,}}(z)\).

(b)
Let \((\pi ,\pi _0)\) be the normal vector of a \((x^*,\eta ^*)\)separating halfspace for \({{\,\mathrm{epi}\,}}(z)\). Then there exist \((\gamma ,\gamma _0) \in P(x^*,\eta ^*)\) and \(\lambda \ge 0\) such that \((\gamma ^\top H, \gamma _0) = \lambda \cdot (\pi ^\top ,\pi _0)\).
Observe, however, that in contrast to Remark 1.1, Corollary 2.1 does not guarantee that the cut generated from a point in the alternative polyhedron is supporting (as seen in Example 1.1 not even if that point is a vertex): A given vector \((\gamma ,\gamma _0) \in P(x^*,\eta ^*)\) might not minimize \(\gamma ^\top b\) among all points in \(P(x^*,\eta ^*)\) which lead to the same cut normal.
Alternatively, as argued by Cornuéjols and Lemaréchal (2006), the reverse polar of a convex set characterizes the set of normals of cuts that separate the origin from the set. The reverse polar was originally introduced in Balas and Ivanescu (1964) and can be defined as follows:
Definition 2.1
Let \(C \subseteq \mathbb {R}^n\) be a convex set. Then the reverse polar set \(C^\) of C is defined as
It is thus a subset of the polar cone
We can use Theorem 1.2 and an appropriate positive scaling^{Footnote 2} to obtain the following description of the reverse polar set.^{Footnote 3} Keeping in mind that we want to separate the point \((x^*,\eta ^*)\) rather than the origin, we must translate the set such that \((x^*,\eta ^*)\) becomes the origin (Fig. 2).
Note furthermore that, as a consequence of Remark 1.1, we can compute for any given normal vector \((\pi ,\pi _0)\) a supporting inequality (if one exists) by solving problems (1.12) or (1.13).
We thus have at our disposal two alternative characterizations of the set of possible normal vectors of \((x^*,\eta ^*)\)separating halfspaces: The alternative polyhedron and the reverse polar set. Despite their similarity, subtle differences exist between both representations that affect their usefulness for the generation of Benders cuts.
It should be noted at this point that we are not the first ones to notice the similarity between the approaches of Cornuéjols and Lemaréchal (2006) and Fischetti et al. (2010). Indeed, the work of Cornuéjols and Lemaréchal (2006) is explicitly cited in Fischetti et al. (2010), albeit only in a remark about the possibility to exchange normalization and objective function in optimization problems over the alternative polyhedron (see Corollary 2.4 below).
Before we proceed, we introduce a variant of the alternative polyhedron, the relaxed alternative polyhedron, which is also used in Gleeson and Ryan (1990). We will see that it is equivalent to the original alternative polyhedron for almost all purposes, but can more easily be connected to the reverse polar set:
Definition 2.2
Let a problem of the form (1.1) and a point \((x^*, \eta ^*) \in \mathbb {R}^n \times \mathbb {R}\) be given. The relaxed alternative polyhedron \(P^\le (x^*,\eta ^*)\) is defined as
To motivate the above definition, observe that optimization problems over the original and the relaxed alternative polyhedron are equivalent, provided that the optimization problem over the relaxed alternative polyhedron has a finite nonzero optimum:
Remark 2.1
Let \(z\) be defined as in (1.2) and let \((x^*,\eta ^*) \in \mathbb {R}^n \times \mathbb {R}\). Let \(({\tilde{\omega }}, {\tilde{\omega }}_0) \in \mathbb {R}^m \times \mathbb {R}\) be such that \(\max \left\{ {{\tilde{\omega }}^\top \gamma + {\tilde{\omega }}_0\gamma _0 {\bigg \vert }\gamma ,\gamma _0 \in P^\le (x^*,\eta ^*)}\right\} < 0\). Then the sets of optimal solutions for \({\tilde{\omega }}^\top \gamma + {\tilde{\omega }}_0\gamma _0\) over \(P^\le (x^*,\eta ^*)\) and \(P(x^*,\eta ^*)\) are identical. Furthermore, every vertex of \(P^\le (x^*,\eta ^*)\) is also a vertex of \(P(x^*,\eta ^*)\).
The following key theorem now almost becomes a trivial observation. However, to our knowledge, the relation between the alternative polyhedron and the reverse polar set has not been made explicit in a similar fashion before (for a set S, we write \(AS := \left\{ {Ax {\bigg \vert }x \in S}\right\} \)).
Theorem 2.1
Let \(z\) be defined as in (1.2) and \((x^*, \eta ^*) \in \mathbb {R}^n \times \mathbb {R}\). Then
Proof
\(\square \)
One common scenario for Benders decomposition is where the master problem is significantly smaller than the subproblem. In this case Theorem 2.1 implies that the relaxed alternative polyhedron is an extended formulation for the reverse polar set, which in particular is always polynomial in size.
We revisit Example 1.1 to illustrate this observation.
Example 1.1
(continuing from p. 3) In the situation of the optimization problem (1.6), observe that the relaxed alternative polyhedron is
It can be verified that \(P^\le (0,0)\) is hence a 3dimensional unbounded polyhedron in \(\mathbb {R}^4\) with extremal rays \({{\,\mathrm{pos}\,}}(0,1,0,1)\), \({{\,\mathrm{pos}\,}}(1,0,0,1)\), \({{\,\mathrm{pos}\,}}(0,0,\nicefrac 14, 1)\) and with the original alternative polyhedron P(0, 0) as the only bounded facet. We can now use Theorem 2.1 to derive the reverse polar set:
The set \({{\,\mathrm{epi}\,}}(z)^\) is visualized in Fig. 3.
Cutgenerating linear programs
One way to select a particular cut normal from the reverse polar set or the alternative polyhedron is by maximizing a linear objective function over these sets. Using Theorem 2.1, we can derive the precise relation between optimization problems over the reverse polar set and the alternative polyhedron.
Corollary 2.2
Let \(z\) be defined as in (1.2), \((x^*,\eta ^*), (\omega ,\omega _0) \in \mathbb {R}^n \times \mathbb {R}\) and
Then \((\pi ,\pi _0)\) is an optimal solution to the problem
if and only if there exists \(\gamma \) such that \(H^\top \gamma = \pi \) and \((\gamma ,\pi _0)\) is an optimal solution to the problem
Furthermore, the objective values of both optimization problems are identical.
Proof
Let \((\pi ,\pi _0)\) be an optimal solution to (2.3). By Theorem 2.1, there exists a vector \(\gamma \) with \(H^\top \gamma = \pi \) such that \((\gamma , \pi _0) \in P^\le (x^*,\eta ^*)\). Now, let \((\gamma ', \gamma '_0)\) be an arbitrary point in \(P^\le (x^*,\eta ^*)\). By Theorem 2.1, \((H^\top \gamma ',\gamma '_0) \in ({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))^\) and thus from the optimality of \((\pi ,\pi _0)\) for (2.3) we obtain
which proves the optimality of \((\gamma ,\pi _0)\) for (2.4).
Similarly, let \((\gamma ,\gamma _0)\) be an optimal solution to (2.4), \(\pi := H^\top \gamma \), and \(\pi _0 := \gamma _0\), which means by Theorem 2.1, \((\pi ,\pi _0) \in ({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))^\).
Now, let \((\pi ',\pi '_0)\) be an arbitrary point in \(({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))^\). By Theorem 2.1, there exists \(\gamma '\) with \(H^\top \gamma ' = \pi '\) such that \((\gamma ',\pi '_0) \in P^\le (x^*,\eta ^*)\). Using the optimality of \((\gamma ,\gamma _0)\) for (2.4), we obtain
which proves the optimality of \((\pi ,\pi _0)\) for (2.3). \(\square \)
Note that the optimization problem stated in (2.4) is technically more general, since there is no reason to limit ourselves to objective functions of the form (2.2) a priori. If we choose a different objective function, we still obtain a valid cut. However, since there may be no objective function \((\omega , \omega _0)\) such that the resulting cut normal is optimal for (2.3), we lose some of the properties associated with optimal solutions from the reverse polar set.
Indeed, this is the approach that Fischetti et al. (2010) take: They use the problem in (2.4) with \({\tilde{\omega }}_m = 0\) for all m which correspond to rows of zeros in the interaction matrix H and \({\tilde{\omega }}_m = 1\) for all other m, as well as \({\tilde{\omega }}_0 = 1\) (or \({\tilde{\omega }}_0 = \kappa \) for some scaling factor \(\kappa > 0\)). In general, there exists no vector \((\omega ,\omega _0)\) such that this choice can be obtained by (2.2).
We now take a closer look at the role of objective functions in the context of Example 1.1:
Example 1.1
(continuing from p. 3) In the situation of the optimization problem (1.6), remember that the point \(P_3\) actually minimizes the 1norm over P(0, 0) and is hence the unique result of the (unscaled) selection procedure from Fischetti et al. (2010). On the other hand the transformation from Theorem 2.1 actually maps this point, which lead to a nonsupporting cut, to the interior of the reverse polar set. It will therefore never appear as an optimal solution of any linear optimization problem.
In order to obtain a supporting cut, we only have to make sure that the used objective can be written in the form \((H\omega , \omega _0)\). In our example, if we choose the objective function over the alternative polyhedron from the set
then the point \(P_3 \in P(0,0)\) is never optimal.
To illustrate this further, we solve the optimization problem using \((\omega ,\omega _0) := (1,1)\) as an example, which implies that
Assuming that we begin with 0 as an initial lower bound for both x and \(\eta \) to make the problem bounded, we thus obtain \((x^1 ,\eta ^1) := (0,0)\) as an initial tentative solution. The cutgenerating problem is
with optimal solution \(P_1 = (\nicefrac {1}{5},0,0, \nicefrac {1}{5})\). The resulting cut is \(\gamma ^\top Hx \gamma _0 \eta \le \gamma ^\top b\), which resolves to
the (linearly scaled) first inequality from (1.6).
Adding this inequality to the master problem, we obtain \((x^2 ,\eta ^2) := (\nicefrac {5}{2},0)\) as the next tentative solution. Now,
which leads to the cutgenerating problem
from which we obtain the next cut:
the (linearly scaled) second inequality from (1.6).
In the next iteration, we obtain \((x^*, \eta ^*) = (\nicefrac {4}{3},\nicefrac {7}{3})\), the optimal solution. Correspondingly,
which certifies optimality of the solution \((x,y)=(\nicefrac {4}{3},\nicefrac {7}{3})\).
We have thus solved the optimization problem in two iterations, whereas the selection procedure from Fischetti et al. (2010) would have selected the point \(P_3\) in the first iteration, leading to a cut that corresponds to the the redundant inequality in the original problem. It would thus require at least one additional iteration to solve the problem.
On the other hand, we will see that the fact that our approach yields two facetdefining cuts is not a coincidence: The following corollary shows that the generated cuts are always at least supporting and we will see in Sect. 3.2 that the generated cuts are actually almost always facetdefining.
One interesting difference between the alternative polyhedron and the reverse polar set, which can be verified using the above example, is their different behavior with respect to algebraic operations on the set of inequalities: If, for instance, we scale one of the inequalities by a positive factor, the reverse polar set remains unchanged (just as the feasible region defined by the set of inequalities). On the other hand, the alternative polyhedron does change and, as a consequence, might yield a different optimal solution with respect to a given objective. In fact, by scaling the inequalities of (1.6) appropriately, we can actually avoid that the point \(P_3\) is optimal for the approach from Fischetti et al. (2010). If an objective function is used which does not take this scaling into account, such as the vector of zeros and ones proposed by Fischetti et al. (2010), then the selected cut might change depending on the scaling factor. Even selecting a suitable manual scaling factor \(\kappa \) as mentioned above cannot fix this, since it cannot scale individual constraints against each other.
Combining our results from this section, we obtain the following statement:
Corollary 2.3
Let \(z\) be defined as in (1.2) and \((x^*,\eta ^*), (\omega ,\omega _0) \in \mathbb {R}^n \times \mathbb {R}\), \(({\tilde{\omega }}, {\tilde{\omega }}_0) := (H\omega , \omega _0)\), and \((\gamma ,\gamma _0) \in P(x^*,\eta ^*)\) be maximal with respect to the objective \(({\tilde{\omega }}, {\tilde{\omega }}_0)\) such that \(\tilde{\omega }^\top \gamma + {\tilde{\omega }}_0 \gamma _0 < 0\). Then the inequality \(\gamma ^\top H x  \gamma _0 \eta \le \gamma ^\top b\) supports \({{\,\mathrm{epi}\,}}(z)\).
Proof
Let \((\pi ,\pi _0) := (H^\top \gamma , \gamma _0)\). Then, by Remark 1.1, the statement is true if \(\gamma \) minimizes \(\gamma ^\top b\) among all possible certificates for the vector \((\pi ,\pi _0)\) in Theorem 1.2. It is easy to verify that \(\gamma \) is indeed a valid certificate for \((\pi ,\pi _0)\) in Theorem 1.2. For a contradiction, we hence assume that it does not minimize \(\gamma ^\top b\). Let thus \(\gamma ' \ge 0\) be an alternative certificate for \((\pi ,\pi _0)\) with \(\gamma '^\top b < \gamma ^\top b\). Then from (1.7) to (1.10) we obtain that \(\gamma '^\top A\pi _0 d^\top = 0\) and \(\gamma '^\top H = \pi ^\top \). Furthermore, since \((\gamma , \gamma _0) \in P^\le (x^*,\eta ^*)\),
We can thus scale \((\gamma ', \gamma _0)\) by an appropriate factor \(\lambda \in (0,1)\) to obtain that \(\lambda \cdot (\gamma ', \gamma _0) \in P^\le (x^*,\eta ^*)\) and
On the other hand, by Remark 2.1, if \((\gamma ,\gamma _0)\) maximizes the objective \(({\tilde{\omega }}, {\tilde{\omega }}_0)\) over \(P(x^*,\eta ^*)\), then it is also maximal within \(P^\le (x^*,\eta ^*)\), a contradiction. By Remark 1.1, this implies that the inequality \(\gamma ^\top H x  \gamma _0 \eta \le \gamma ^\top b\) does indeed support \({{\,\mathrm{epi}\,}}(z)\). \(\square \)
A critical requirement for Corollary 2.3 is that \({\tilde{\omega }}^\top \gamma + \tilde{\omega }_0 \gamma _0 < 0\). Cornuéjols and Lemaréchal (2006, Theorem 2.3) establish some criteria on the objective function for which optimization problems over the reverse polar set are bounded. We have simplified the notation for our purposes and rephrased the relevant parts of the theorem according to our terminology.
Theorem 2.2
(Cornuéjols and Lemaréchal (2006)) Let \((x^*,\eta ^*) \notin {{\,\mathrm{epi}\,}}(z)\), \((\omega ,\omega _0) \in \mathbb {R}^n \times \mathbb {R}\), and
Then
Furthermore, if \((\omega ,\omega _0) \in ({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))\), then \(z^* \le 1\).
Note in particular that the last part of the above statement implies \(z^* < 0\) whenever \((\omega ,\omega _0) \in {{\,\mathrm{pos}\,}}({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*)) \setminus \left\{ {0}\right\} \), which provides us with a large variety of objective functions for which \( \omega ^\top \gamma + \omega _0 \gamma _0 < 0\) holds in the optimal solution. By Corollaries 2.2 and 2.3, this means that the cut which results from maximizing these objectives over the reverse polar set is guaranteed to be supporting.
Alternative representations
We now derive an alternative representation of the optimization problem in (2.4), which will turn out to be much more useful in practice. For instance, the structure of the resulting problem will be very similar to the original subproblem, which makes it easy to use existing solution algorithms for the subproblem in a cutgenerating program.
Cornuéjols and Lemaréchal (2006, Theorem 4.2) prove that linear optimization problems over the reverse polar set can be evaluated in terms of the support function of the original set (in our case \({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*)\)). This can also be applied to the alternative polyhedron, as mentioned (without proof) by Fischetti et al. (2010). The following lemma makes a statement similar to Cornuéjols and Lemaréchal (2006, Theorem 4.2), which is applicable to a wider range of settings. For the proof, we refer to Stursberg (2019, Theorem 3.20).
Lemma 2.1
(Cornuéjols and Lemaréchal (2006)) Let \(K \subseteq \mathbb {R}^n\) be a cone and \(c_1, c_2 \in \mathbb {R}^n\). Consider the optimization problems
and
Then the following hold:

(a)
If \(x^*\) is an optimal solution for (2.5) with objective value \(\xi > 0\), then \(\frac{1}{\xi } \cdot x^*\) is an optimal solution for (2.6) with objective value \(\frac{1}{\xi }\).

(b)
Conversely, if \(x^*\) is an optimal solution for (2.6) with objective value \(\xi < 0\), then \(\frac{1}{\xi } \cdot x^*\) is an optimal solution for (2.6) with objective value \(\frac{1}{\xi }\).
This lemma allows us to solve optimization problems of the form (2.4) by instead resorting to the optimization problem
Let \(({\tilde{\omega }},{\tilde{\omega }}_0) \in \mathbb {R}^m \times \mathbb {R}\) and let \((\gamma ^*,\gamma _0^*)\) denote an optimal solution with value \(\xi > 0\) for (2.7) to (2.9). Applying Lemma 2.1 with \(c_1 := (Hx^*b,\eta ^*)\), \(c_2 := ({\tilde{\omega }},{\tilde{\omega }}_0)\) and \(K := \left\{ {(\gamma ,\gamma _0) \ge 0 {\bigg \vert }\gamma ^\top A + \gamma _0 d^\top = 0}\right\} \), we obtain that \(\frac{1}{\xi } \cdot (\gamma ^*,\gamma _0^*)\) is an optimal solution with value \(\frac{1}{\xi }\) for (2.4).
The structural similarity of (2.7) to (2.9) and the original problem becomes more apparent when we consider the dual problem:
Corollary 2.4
Let \(({\tilde{\omega }},{\tilde{\omega }}_0) \in \mathbb {R}^m \times \mathbb {R}\) and \((\lambda ,x,y)\) be an optimal solution for the problem
with \(\lambda > 0\). Denote the corresponding dual solution by \((\gamma ,\gamma _0)\). Then \(\frac{1}{\lambda }(\gamma ,\gamma _0)\) is an optimal solution for
with objective value \(\frac{1}{\lambda }\).
Note that, together with our observations in the context of the definition of the alternative polyhedron (1.5), this means in particular that

(a)
Whenever (2.10) to (2.12) has objective value 0, then the alternative polyhedron is empty and \((x^*,\eta ^*) \in {{\,\mathrm{epi}\,}}(z)\), and

(b)
Whenever (2.10) to (2.12) is feasible with (finite) objective value greater than 0, then (2.3) and (2.4) have objective values strictly less than 0, which means that the requirements for Remark 2.1 and Corollary 2.3 are satisfied.
Remark 2.2
If \(({\tilde{\omega }}, {\tilde{\omega }}_0) := (H\omega , \omega _0)\), then the optimization problem (2.10) to (2.12) becomes
The difference between the formulations from Corollary 2.4 and Remark 2.2 lies in how they relax the original problem: In (2.10) to (2.12), the relaxation works on the level of individual inequalities by relaxing their righthand sides, whereas in (2.13) to (2.15) it works on the level of the master solution \((x^*,\eta ^*)\), allowing us to choose a possibly more advantageous value for the vector x itself.
Cut selection
As we have seen in the previous section, Benders decomposition can be viewed as an instance of a classical cutting plane algorithm (Theorem 1.2). The Benders subproblem takes the role of the separation problem and the alternative polyhedron that is commonly used to select a Benders cut is a higherdimensional representation of the reverse polar set, which characterizes all possible cut normals (Theorem 2.1).
Finally, Corollary 2.4 and Remark 2.2 show that selecting a cut normal by a linear objective over the reverse polar set or the alternative polyhedron can be interpreted as two different relaxations (2.10) to (2.12) and (2.13) to (2.15) of the original Benders feasibility subproblem (1.4). The former relaxation provides more flexibility with respect to the choice of parameters and coincides with the latter for a particular selection of the objective function.
Cut selection is one of four major areas of algorithmic improvements for Benders decomposition that recent work has focused on [see, e. g., the very extensive literature review by Rahmaniani et al. (2017)]. As a consequence, a number of selection criteria for Benders cuts have previously been explicitly proposed in the literature. Many of them also arise naturally from our discussion and analysis of the Benders decomposition algorithm above. We will first present these criteria in the way they typically appear in the literature and then link them to the reverse polar set and/or the alternative polyhedron.
Minimal infeasible subsystems
The work of Fischetti et al. (2010) is based on the premise that “one is interested in detecting a ‘minimal source of infeasibility’” whenever the feasibility subproblem (1.4) is empty. They hence suggest to generate Benders cuts based on Farkas certificates that correspond to minimal infeasible subsystems (MIS) of (1.4). Fischetti et al. (2010) empirically study the performance of MIScuts on a set of multicommodity network design instances. Their results suggest that MISbased cut selection outperforms the standard implementation of Benders decomposition by a factor of at least 2–3. Furthermore, this advantage increases substantially when focusing on harder instances (e.g. those which could not be solved by the standard implementation within 10 hours).
We define this criterion as follows:
Definition 3.1
Let \(z\) be defined as in (1.2) and let \((\pi ,\pi _0) \in \mathbb {R}^n \times \mathbb {R}\). We say that \((\pi ,\pi _0)\) satisfies the MIS criterion if there exists \((\gamma , \gamma _0) \ge 0\) such that

(a)
\(\pi =H^\top \gamma , \pi _0=\gamma _0\)

(b)
The inequalities of (1.4) corresponding to the rows of H which are multiplied by the nonzero components of \((\gamma ,\gamma _0)\) in the equations in a) form a minimal infeasible subsystem of (1.4).
Note that we have defined the MIS criterion as a property of a normal vector, rather than a property of a cut. The reason for this is that the cut normal is the only relevant choice to make, given that an optimal righthand side for each cut normal is provided by Corollary 2.3. Accordingly, we will call any cut with a normal vector that satisfies the MIS criterion a MIScut.
Gleeson and Ryan (1990) show that the set of \((\gamma , \gamma _0)\) that appear in the above definition is exactly (up to homogeneity) the set of vertices of the alternative polyhedron:
Theorem 3.1
(Gleeson and Ryan (1990)) Let \((x^*,\eta ^*) \in \mathbb {R}^n \times \mathbb {R}\). For each vertex v of the (relaxed) alternative polyhedron (1.5), the set of constraints corresponding to the nonzero entries of v forms a minimal infeasible subsystem of (1.4). Conversely, for every minimal infeasible subsystem, there exists a vertex of the alternative polyhedron.
This theorem immediately provides a characterization of cut normals which satisfy MIS in terms of the alternative polyhedron, which is also used in Fischetti et al. (2010). However, we can furthermore use Theorem 2.1, to transfer one direction of the characterization to the reverse polar set:
Corollary 3.1
Let \(z\) be defined as in (1.2) and \((x^*,\eta ^*) \in \mathbb {R}^n \times \mathbb {R}\). If there is a vertex \((\gamma , \gamma _0)\) of \(P(x^*,\eta ^*)\) such that \((\pi ,\pi _0)=(H^\top \gamma ,\gamma _0)\), then the vector \((\pi ,\pi _0)\) satisfies the MIS criterion. Furthermore, if \((\pi ,\pi _0)\) is a vertex of \(({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))^\), then it satisfies the MIS criterion.
Note that the reverse direction of the last sentence is generally not true, i.e. there might be minimal infeasible subsystems that do not correspond to vertices of the reverse polar set. As an example to illustrate this, as well as the above definition overall, consider Example 1.1 and specifically Fig. 3: Each of the vectors \({\tilde{P}}_1, {\tilde{P}}_2, {\tilde{P}}_3\) satisfies the MIS criterion, since they can be related by the equations \(\pi =H^\top \gamma , \pi _0=\gamma _0\) to vertices of the alternative polyhedron, which can be identified with minimal infeasible subsystems of (1.4).
Facetdefining cuts
In cutting plane algorithms for polyhedra, facetdefining cuts are commonly considered to be very useful since they form the smallest family of inequalities which completely describe (the convex hull of) the feasible solutions. A cuttingplane algorithm that can separate (distinct) facet inequalities in each iteration is not necessarily computationally efficient, but at least it is automatically guaranteed to terminate after a finite number of iterations. Also in practical applications, facet cuts have turned out to be extremely useful, e.g. in the context of branchandcut algorithms for integer programs such as the Traveling Salesman Problem. This is why the description of facetdefining inequalities has been a large and very active area of research for decades (see Balas 1975; Nemhauser and Wolsey 1988; Cook et al. 1998; Korte and Vygen 2008 and, as mentioned before, Conforti and Wolsey 2018).
Remember that a halfspace \(H^{\le }_{(\pi ,\alpha )}\) is facetdefining for a set C if \(C \subseteq H^{\le }_{(\pi ,\alpha )}\) and \(H_{(\pi ,\alpha )} \cap C\) contains \(\dim (C)\) many affinely independent points. Analogously to the MIS criterion above, we define the Facet criterion for a normal vector in the context of Benders decomposition as follows:
Definition 3.2
Let \(z\) be defined as in (1.2) and \((\pi ,\pi _0) \in \mathbb {R}^n \times \mathbb {R}\setminus \left\{ {0}\right\} \). We say that \((\pi ,\pi _0)\) satisfies the Facet criterion if there exists \(\alpha \in \mathbb {R}\) such that \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) is either facetdefining for \({{\,\mathrm{epi}\,}}(z)\) or the corresponding hyperplane \(H_{((\pi ,\pi _0),\alpha )}\) contains \({{\,\mathrm{epi}\,}}(z)\).
Note that, in deviation from the common definition of a facetdefining cut, the above definition requires that the halfspace supports at least \(\dim (C)\) affinely independent points. In other words, in the case where \({{\,\mathrm{epi}\,}}(z)\) is not fulldimensional, we also allow that \({{\,\mathrm{epi}\,}}(z)\) is entirely contained in the hyperplane which represents the boundary of \(H^{\le }_{((\pi ,\pi _0),\alpha )}\). In this situation, the comparison of different cut normals is inherently difficult: Since there is no clear way to tell if a cut supporting a facet of \({{\,\mathrm{epi}\,}}(z)\) or one fully containing the set is the stronger cut, the Facet criterion captures arguably the strongest statement about a cut in relation to \({{\,\mathrm{epi}\,}}(z)\) that we can make in general: In no case would we want to select a cut that supports neither a facet nor fully contains the set \({{\,\mathrm{epi}\,}}(z)\).
On the other hand, by this definition, the “trivial” cut normal (0, 0) would be facetdefining (since the hyperplane \(H_{((0,0),0)}\) contains all of \(\mathbb {R}^n \times \mathbb {R}\) and thus also \({{\,\mathrm{epi}\,}}(z)\)). Since this is not very useful, we have to exclude this choice explicitly and thus choose \((\pi ,\pi _0) \in \mathbb {R}^n \times \mathbb {R}\setminus \left\{ {0}\right\} \) in Definition 3.2.
For an example to illustrate the above definition, we refer to Fig. 4 where it is discussed together with the property of Paretooptimality, which will be defined later.
The following result was proven by Cornuéjols and Lemaréchal (2006, Theorem 6.2), containing a minor error in the case where the set P is subdimensional.^{Footnote 4} We therefore restate a corrected version of the important parts below, a corresponding proof can be found in Stursberg (2019, Theorem 3.30).
Theorem 3.2
(Cornuéjols and Lemaréchal (2006)) Let \(P \subseteq \mathbb {R}^n\) be a polyhedron, \(x^* \notin P\) and
Then, there exists an \(x^*\)separating halfspace with normal vector \(\pi \ne 0\) supporting an rdimensional face of P if and only if there exists a vertex \(\pi ^*\) of \({{\,\mathrm{lin}\,}}(Px^*) \cap (Px^*)^\) and some \(\lambda > 0\) such that \(\lambda \pi \in \pi ^* + {{\,\mathrm{lin}\,}}(Px^*)^\bot \).
Most notably, for the case where P is fulldimensional (i. e., \(\dim (P)=n\)) the above theorem implies that there exists an \(x^*\)separating halfspace with normal vector \(\pi \) supporting a facet of P if and only if there exists a vertex \(\pi ^*\) of \((Px^*)^\) and \(\lambda \ge 0\) such that \(\lambda \pi = \pi ^*\).
In this case, every cut generated from a vertex of the reverse polar set defines a facet of \({{\,\mathrm{epi}\,}}(z)\). If an explicit \(\mathcal {H}\)representation of the reverse polar set is available, we can thus easily obtain a facetdefining cut, e. g., by linear programming.
Note that since \(P^\le (x^*,\eta ^*)\) is linefree (i. e. its lineality space, the maximal linear subspace it contains, is \(\left\{ {0}\right\} \)), Theorem 2.1 implies that for every vertex of the reverse polar set there exists a vertex of the relaxed alternative polyhedron (and hence of the original alternative polyhedron) that leads to the same cut normal. In other words, if the normal of an \(x^*\)separating halfspace satisfies the Facet criterion, then it also satisfies the MIS criterion.
On the other hand, Theorem 2.1 is not sufficient to guarantee that selecting a vertex of the alternative polyhedron yields a facetdefining cut: As Example 1.1 shows, a vertex of \(P^\le (x^*,\eta ^*)\), is not necessarily mapped to a vertex of the reverse polar set under the transformation from Theorem 2.1. This exposes a useful hierarchy of subsets of the alternative polyhedron according to the properties of the cut normals which they yield: Selecting a vertex of the alternative polyhedron already guarantees that the resulting cut normal satisfies the MIS criterion and the points that lead to cut normals satisfying the Facet criterion constitute a subset of these vertices. The approach of selecting MIScuts may thus be viewed as a heuristic method to find Facetcuts.
Although cuts satisfying the MIS criterion in general do not satisfy the Facet criterion, we can obtain some information on when this is the case in the situation of Corollary 2.2, i. e. if the objective function \(({\tilde{\omega }},{\tilde{\omega }}_0)\) used to select the cut via problem (2.4) satisfies \(({\tilde{\omega }},{\tilde{\omega }}_0) = (H\omega ,\omega _0)\) for some valid objective \((\omega ,\omega _0)\) for problem (2.3).
In this case it turns out that we actually obtain a Facetcut for all objectives \((\omega ,\omega _0)\) except those from a lowerdimensional subspace. More precisely, we can prove the following characterization of the relationship between vertices of the alternative polyhedron and cut normals satisfying the Facet criterion. Like Conforti and Wolsey (2018, Proposition 6), the following theorem provides a method to generate facetdefining cuts using a single linear program. In contrast to Conforti and Wolsey (2018), however, our theorem uses a linear program over the (relaxed) alternative polyhedron and thus creates a link to wellestablished cut selection methods in the context of Benders decomposition, such as that proposed by Fischetti et al. (2010):
Theorem 3.3
Let \(z\) be defined as in (1.2), \((x^*, \eta ^*) \in (\mathbb {R}^n \times \mathbb {R}) \setminus {{\,\mathrm{epi}\,}}(z)\), and \((\omega ,\omega _0) \in {{\,\mathrm{cl}\,}}({{\,\mathrm{pos}\,}}({{\,\mathrm{epi}\,}}(z)  (x^*,\eta ^*)))\). Then, there exists an optimal vertex \((\gamma ^*,\gamma _0^*) \in P^\le (x^*,\eta ^*)\) with respect to the objective function \((H\omega ,\omega _0)\) such that the resulting cut normal \((H^\top \gamma ^*,\gamma _0^*)\) is \((x^*,\eta ^*)\)separating and satisfies the Facet criterion.
Proof
Let \(L := {{\,\mathrm{lin}\,}}({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))\) and observe that L is orthogonal to the lineality space of \(({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^\). From Theorem 2.2, the reverse polar set \(({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^\) is bounded in the direction of \((\omega ,\omega _0)\). We may therefore choose an optimal solution \((\pi ,\pi _0)\) from the intersection \(({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^ \cap L\). While the reverse polar need not be linefree, note that \(({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^ \cap L\) is indeed linefree and we can therefore choose \((\pi ,\pi _0)\) to be extremal in \(({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^ \cap L\). By Corollary 2.2, there exists \(\gamma '\) with \(H^\top \gamma ' = \pi \) such that \((\gamma ',\pi _0)\) is an optimal solution to the problem
Denote by \(P^*\) the face of optimal solutions of (3.2) and observe that
Let \((\gamma ^*,\gamma ^*_0)\) be a vertex of \(P^* \cap \left\{ {(\gamma ,\gamma _0) {\bigg \vert }(H^\top \gamma , \gamma _0)  (\pi ,\pi _0) \in L^\bot }\right\} \) (which exists, since \(P^\le (x^*,\eta ^*)\) is linefree). Then \((\gamma ^*,\gamma ^*_0)\) is obviously optimal for (3.2). Since furthermore \((H^\top \gamma ^*,\gamma _0^*) = (\pi ,\pi _0) + v\) for some \(v \in L^\bot \), this means by Theorem 3.2 that it satisfies the Facet criterion.
It remains to show that \((\gamma ^*,\gamma ^*_0)\) is a vertex of \(P^*\), thus showing that it is also a vertex of \(P^\le (x^*,\eta ^*)\). To see this, let \((\gamma ^1,\gamma ^1_0), (\gamma ^2,\gamma ^2_0) \in P^*\) such that \((\gamma ^*,\gamma ^*_0) \in {{\,\mathrm{relint}\,}}([(\gamma ^1,\gamma ^1_0),(\gamma ^2,\gamma ^2_0)])\)^{Footnote 5} However, it follows that
and by Theorem 2.1, \([(H^\top \gamma ^1,\gamma ^1_0),(H^\top \gamma ^2,\gamma ^2_0)] \subseteq ({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^\). As \((\pi ,\pi _0)\) is extremal in the set \(({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^ \cap L\), this implies that both \((H^\top \gamma ^1,\gamma ^1_0),(H^\top \gamma ^2,\gamma ^2_0) \in \left\{ {(\pi ,\pi _0) + v {\bigg \vert }v \in L^\bot }\right\} \) which in turn implies that \((\gamma ^1,\gamma ^1_0), (\gamma ^2,\gamma ^2_0) \in P^* \cap \left\{ {(\gamma ,\gamma _0) {\bigg \vert }(H^\top \gamma , \gamma _0)  (\pi ,\pi _0) \in L^\bot }\right\} \). As \((\gamma ^*,\gamma ^*_0)\) is extremal in \(P^* \cap \left\{ {(\gamma ,\gamma _0) {\bigg \vert }(H^\top \gamma , \gamma _0)  (\pi ,\pi _0) \in L^{\bot }}\right\} \), we obtain that \((\gamma ^1,\gamma ^1_0)=(\gamma ^2,\gamma ^2_0)=(\gamma ^*,\gamma ^*_0)\), which proves extremality of \((\gamma ^*,\gamma ^*_0)\) in \(P^*\). \(\square \)
In particular, Theorem 3.3 implies the following: If \((\gamma ',\gamma _0') \in P^\le (x^*,\eta ^*)\) is an optimal vertex with respect to the objective function \((H\omega ,\omega _0)\) such that the resulting cut normal \((H^\top \gamma ',\gamma _0')\) does not satisfy the Facet criterion, then the optimal solution for maximizing \((H\omega ,\omega _0)\) over \(P^\le (x^*,\eta ^*)\) is not unique. Furthermore, by Corollary 2.2, this implies that the same is true for maximizing the objective \((\omega , \omega _0)\) over \(({{\,\mathrm{epi}\,}}(z) (x^*,\eta ^*))^\).
We can summarize our results as follows: While any Facetcut is also an MIScut, the reverse is not always true. However, if we optimize the objective \((H\omega ,\omega _0)\) over the alternative polyhedron, then there exists only a subdimensional set of choices for the vector \((\omega ,\omega _0)\) for which the resulting cut might not satisfy the Facet criterion (those, for which the optimum over the reverse polar set is nonunique).
This suggests that these cases should be “rare” in practice, especially if we choose (or perturb) \((\omega ,\omega _0)\) randomly from some fulldimensional set. This argument, why a cut obtained for a generic vector \((\omega ,\omega _0)\) can be expected to be facetdefining, is identical to the concept of “almost surely” finding facetdefining cuts proposed by Conforti and Wolsey (2018).
Looking back at Remark 2.2, this similarity should not come as a surprise: With \((\omega , \omega _0) = ({\bar{x}}  x^*, {\bar{\eta }}  \eta ^*)\) for a point \((\bar{x}, {\bar{\eta }}) \in {{\,\mathrm{relint}\,}}({{\,\mathrm{epi}\,}}(z))\), the resulting cutgenerating LP is almost identical. In fact, the point \(({\bar{x}}, {\bar{\eta }})\) in this case takes the role of the point that the origin is relocated into in the approach from Conforti and Wolsey (2018). Observe, however, that while Conforti and Wolsey (2018) require that point to lie in the relative interior of \({{\,\mathrm{epi}\,}}(z)\), we can actually expect a cut satisfying the Facet criterion from any \((\omega , \omega _0)\) for which the optimal objective over the reverse polar is strictly negative. By Theorem 2.2, one sufficient (but not necessary) criterion for this is to choose \((\omega , \omega _0) = ({\bar{x}}  x^*, {\bar{\eta }}  \eta ^*)\) for an arbitrary point \(({\bar{x}}, {\bar{\eta }}) \in {{\,\mathrm{relint}\,}}({{\,\mathrm{epi}\,}}(z))\).
Paretooptimality
The first systematic work on the general selection of Benders cuts to our knowledge was undertaken by Magnanti and Wong (1981). The paper, which has proven very influential and still being referred to regularly, focuses on the property of Paretooptimality. It can intuitively be described as follows: A cut is Paretooptimal if there is no other cut valid for \({{\,\mathrm{epi}\,}}(z)\) which is clearly superior, which dominates the first cut.
In this setting, any cut that does not support \({{\,\mathrm{epi}\,}}(z)\) is obviously dominated. Between supporting cuts, there is no general criterion for domination. We can, however, compare cuts where the cut normal \((\pi ,\pi _0)\) satisfies \(\pi _0 \ne 0\) (this is also the case covered by Magnanti and Wong (1981)):
Definition 3.3
For a problem of the form (1.1), we say that an inequality \(\pi ^\top x +\pi _0 \eta \le \alpha \) with \(\pi _0 < 0\) is dominated by another inequality \(\pi '^\top x + \pi '_0 \eta \le \alpha '\) if \(\pi '_0 < 0\) and
with strict inequality for at least one \(x \in S\).
If \(\pi _0 < 0\) and \(\pi ^\top x +\pi _0 \eta \le \alpha \) is not dominated by any valid inequality for \({{\,\mathrm{epi}\,}}(z)\), then we call it Paretooptimal.
Remember that the set S contains all points \(x \in \mathbb {R}^n\) that are feasible for an optimization problem of the form (1.1) if we ignore the linear constraints \(Hx + Ay \le b\). By the above definition, a cut dominates another cut if the minimum value of \(\eta \) that it enforces is at least as good for all \(x \in S\) and strictly better for at least one \(x \in S\) (see Fig. 4).
Analogously to the previous criteria, we define the Pareto criterion for a cut normal:
Definition 3.4
For a problem of the form (1.1) with \(z\) as defined as in (1.2), let \((\pi ,\pi _0) \in \mathbb {R}^n \times \mathbb {R}\). We say that \((\pi ,\pi _0)\) satisfies the Pareto criterion if there exists a scalar \(\alpha \in \mathbb {R}\) such that the inequality \(\pi ^\top x +\pi _0 \eta \le \alpha \) is Paretooptimal.
This criterion is very reasonable: If a cut is not Paretooptimal, then there exists a different cut which is also valid for \({{\,\mathrm{epi}\,}}(z)\), but leads to a strictly tighter approximation. We would hence prefer to generate a stronger, Paretooptimal cut right away.
The following theorem provides us with a characterization of Paretooptimal cuts. It is based on the idea of Magnanti and Wong (1981, Theorem 1), which is formulated under the assumption that the subproblem is always feasible (which implies that \(\pi _0 < 0\) for any cut normal \((\pi ,\pi _0)\)). While the original theorem is only concerned with sufficiency, we extend the result in a natural way to obtain a criterion that gives a complete characterization of Paretooptimal cuts. We use the following separation lemma:
Lemma 3.1
(Rockafellar (1970)) Let \(C \subseteq \mathbb {R}^n\) be a nonempty convex set and \(K \subseteq \mathbb {R}^n\) a nonempty polyhedron such that \({{\,\mathrm{relint}\,}}(C) \cap K = \emptyset \). Then, there exists a hyperplane separating C and K which does not contain C.
Using this lemma, we obtain the following theorem:
Theorem 3.4
For a problem of the form (1.1), let \((\pi , \pi _0) \in \mathbb {R}^n \times \mathbb {R}\) with \(\pi _0 < 0\). The inequality \(\pi ^\top x +\pi _0 \eta \le \alpha \) is Paretooptimal if and only if \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) is a halfspace supporting \({{\,\mathrm{epi}\,}}(z)\) in a point \((x^*,\eta ^*) \in {{\,\mathrm{epi}\,}}(z) \cap {{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}(S)) \times \mathbb {R}\).
Proof
For the if part, suppose for a contradiction that \((\pi ^\top ,\pi _0)(x,\eta )^\top \le \alpha \) is not Paretooptimal, i. e. there exist some \(\pi ',\pi _0',\alpha '\) such that the inequality \((\pi '^\top ,\pi '_0)(x,\eta )^\top \le \alpha '\) dominates the former inequality. This means that for all \(x \in S\) (and hence all \(x \in {{\,\mathrm{conv}\,}}(S)\)), it holds that
and furthermore
Finally, since \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) supports \({{\,\mathrm{epi}\,}}(z)\) in \((x^*,\eta ^*)\),
and hence equality must hold everywhere in the above inequality chain. Now, as \(x^* \in {{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}(S))\), we can choose \(\lambda > 1\) such that \({\tilde{x}} := {\bar{x}} + \lambda (x^*  {\bar{x}}) \in {{\,\mathrm{conv}\,}}(S)\). But then
contradicting (3.4).
For the onlyif part, we first note that if \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) does not support \({{\,\mathrm{epi}\,}}(z)\), then it is obviously dominated by \(H^{\le }_{((\pi ,\pi _0),\alpha ')}\) with \(\alpha ' := \alpha '+\varepsilon \) for some \(\varepsilon >0\). Therefore, let \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) be such that it supports \({{\,\mathrm{epi}\,}}(z)\), but not in points from the set \({{\,\mathrm{epi}\,}}(z) \cap {{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}(S)) \times \mathbb {R}\). Denote by \(S^* := \left\{ {x \in \mathbb {R}^n {\bigg \vert }\exists \eta : (x,\eta )\in {{\,\mathrm{epi}\,}}(z) \cap H_{((\pi ,\pi _0),\alpha )}}\right\} \) the set of points where \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) supports \({{\,\mathrm{epi}\,}}(z)\).
Since \({{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}(S)) \cap S^* = \emptyset \), we can use Lemma 3.1 to obtain a hyperplane separating \({{\,\mathrm{conv}\,}}(S)\) and \(S^*\) which does not contain S. Hence, there exist \(\pi ^*, \alpha ^*\) such that \({\pi ^*}^\top x \ge \alpha ^*\) for all \(x \in S^*\) and \({\pi ^*}^\top x \le \alpha ^*\) for all \(x \in {{\,\mathrm{conv}\,}}(S)\), where the second inequality is strict for some \(x \in {{\,\mathrm{conv}\,}}(S)\) and thus also for some \(x^* \in S\).
Let \(\varepsilon >0\), \(\pi ' := \pi  \varepsilon \pi ^*\) and \(\alpha ' := \alpha  \varepsilon \alpha ^*\). If \(\varepsilon \) is sufficiently small, then the inequality \((\pi '^\top ,\pi _0)(x,\eta )^\top \le \alpha '\) is valid for \({{\,\mathrm{epi}\,}}(z)\): All \((x,\eta ) \in {{\,\mathrm{epi}\,}}(z)\) with \(x \notin S^*\) satisfy the original inequality strictly and for all \(x \in S^*\),
since the original inequality is valid for \({{\,\mathrm{epi}\,}}(z)\).
Finally, we claim that the inequality \((\pi ^\top ,\pi _0)(x,\eta )^\top \le \alpha \) is dominated by the inequality \((\pi '^\top ,\pi _0)(x,\eta )^\top \le \alpha '\): For all \(x \in S\), it holds that
Since the last inequality is strict for \(x^* \in S\), this proves the statement. \(\square \)
For the case where S is convex, the previous theorem immediately implies the following statement:
Corollary 3.2
Let S be convex. Then, \(\pi ^\top x +\pi _0 \eta \le \alpha \) is Paretooptimal if and only if \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) supports a face F of \({{\,\mathrm{epi}\,}}_S(z)\) such that \(F \not \subset {{\,\mathrm{relbd}\,}}(S) \times \mathbb {R}\).
Magnanti and Wong (1981) also propose an algorithm that computes a Paretooptimal cut by solving the cutgenerating problem twice. While their algorithm is defined for the original Benders optimality cuts, it can be adapted to work with other cut selection criteria, as well. Sherali and Lunday (2013) present a method based on multiobjective optimization to obtain a cut that satisfies a weaker version of Paretooptimality by solving only a single instance of the cutgenerating LP. Papadakos (2008) notes that, given a point in the relative interior of \({{\,\mathrm{conv}\,}}(S)\), a Paretooptimal cut can be generated using a single run of the cutgenerating problem. Also, under certain conditions on the problem, other points not in the relative interior allow this, as well. However, the approach suggested by the authors adds Paretooptimal cuts independently from master or subproblem solutions, together with subproblemgenerated cuts, which are generally not Paretooptimal. This means that the Paretooptimal cuts which are added may not even cut off the current tentative solution. The upcoming Theorem 3.5 will lead to an approach that reconciles both objectives, generating cuts that are both Paretooptimal and cut off the current tentative solution.
We use a result by Cornuéjols and Lemaréchal (2006) on the set of points exposed by a cut normal \((\pi ,\pi _0)\) to derive a method that always obtains a Paretooptimal cut. The following lemma has been slightly generalized and rewritten to match our setting and notation, but it follows the general idea of Cornuéjols and Lemaréchal (2006, Theorem 3.4).
Lemma 3.2
(Cornuéjols and Lemaréchal (2006)) Let \((x^*,\eta ^*) \in \mathbb {R}^n \times \mathbb {R}\setminus {{\,\mathrm{epi}\,}}(z)\) and \((\omega , \omega _0) \in {{\,\mathrm{pos}\,}}({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))\) and let
where \(h_Q(\omega ,\omega _0) := \sup \left\{ {\omega ^\top x + \omega _0 \eta {\bigg \vert }(x,\eta ) \in Q}\right\} \) is the support function of the set Q.
If \((\pi ,\pi _0)\) is optimal in \(Q := ({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))^\) with respect to the objective \((\omega , \omega _0)\) then there exists \(\alpha \in \mathbb {R}\) such that \(H^{\le }_{((\pi ,\pi _0),\alpha )}\) supports \({{\,\mathrm{epi}\,}}(z)\) in \(({\bar{x}}, {\bar{\eta }})\).
Proof
The case of \((\omega , \omega _0) \in ({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))\) was proven by Cornuéjols and Lemaréchal (2006, Theorem 3.4). If \((\omega , \omega _0) \in {{\,\mathrm{pos}\,}}({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))\), then there is \(\mu > 0\) such that \(\mu \cdot (\omega ,\omega _0) \in ({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))\). Note that if \((\pi ,\pi _0)\) is optimal with respect to \((\omega , \omega _0)\), then also with respect to \(\mu \cdot (\omega ,\omega _0)\). Thus it follows from Cornuéjols and Lemaréchal (2006, Theorem 3.4) that there exists \(\alpha \in \mathbb {R}\) such that \(H^\le _{(\pi , \pi _0),\alpha }\) supports \({{\,\mathrm{epi}\,}}_S(z)\) in
\(\square \)
We can now prove the theorem already mentioned above.
Theorem 3.5
Let \((x^*,\eta ^*) \in S \times \mathbb {R}\), \((\omega , \omega _0) \in {{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}({{\,\mathrm{epi}\,}}_S(z)(x^*,\eta ^*)))\), \((\pi ,\pi _0)\) be optimal in \(({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))^\) with respect to the objective \((\omega , \omega _0)\), and \(\pi _0 < 0\). Then \((\pi ,\pi _0)\) satisfies the Pareto criterion.
Proof
Let \(Q := ({{\,\mathrm{epi}\,}}(z)(x^*,\eta ^*))^\) again and \(\lambda := (h_Q(\omega ,\omega _0))^{1}\). Since, in particular, \((\omega , \omega _0) \in {{\,\mathrm{epi}\,}}_S(z)(x^*,\eta ^*)\) it follows from the definition of the reverse polar set that \(h_Q(\omega ,\omega _0) \le 1\) and thus \(\lambda \in [0,1]\).
For \(({\bar{x}}, {\bar{\eta }})\) from Lemma 3.2, we obtain \(({\bar{x}}, {\bar{\eta }}) = \lambda \left( (\omega ,\omega _0) + (x^*,\eta ^*)\right) + (1\lambda ) (x^*,\eta ^*)\) is a convex combination of \((\omega , \omega _0) + (x^*,\eta ^*) \in {{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}({{\,\mathrm{epi}\,}}_S(z))) \subseteq {{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}(S)) \times \mathbb {R}\) and \((x^*,\eta ^*) \in S \times \mathbb {R}\). Therefore, \({\bar{x}} \in {{\,\mathrm{relint}\,}}({{\,\mathrm{conv}\,}}(S))\) and thus by Theorem 3.4 the cut defined by \((\pi ,\pi _0)\) is Paretooptimal. \(\square \)
The results from this section are summarized in Table 1.
Computational results
To validate the theoretical results presented in this paper, we have compared our refined cut selection approach to that presented by Fischetti et al. (2010) on a set of instances of the Capacity Expansion Problem for electrical power systems.
To map the range of potential instances we use a total of 14 test instances, spanning from a small, closely connected model of the Bavarian power system to a large, realistic model of the (rather sparsely connected) European power system consisting of 102 demand regions with 587 (aggregated) generation units and 195 existing and potential transmission lines (Schaber et al. 2012), that were investigated in the context of two joint research projects.
For both models, we optimize capacity expansion and hourly dispatch based on demand data and data for the availability of renewable energy sources in hourly resolution for a period of one year. Due to their inherent structure, where subproblems for individual timesteps are loosely coupled by capacity expansion decisions and storage constraints, this type of problem is generally wellsuited for Benders decomposition. To give an indication of the size of the resulting optimization problems, instances 112 each consist of \(\approx \) 800,000 variables, \(\approx \) 1,200,000 constraints and \(\approx \) 3,700,000 nonzero entries in the constraint matrix. Instances 13 and 14 consist of \(\approx \) 1,500,000 variables, \(\approx \) 2,600,000 constraints and \(\approx \) 7,600,000 nonzero entries in the constraint matrix.
To demonstrate the benefits of our refined cut selection approach, we continuously update the weight vector \(({\tilde{\omega }},{\tilde{\omega }}_0)\) so that in each iteration it satisfies the conditions of Theorem 3.5 and thus the resulting cut meets the most advantageous of the criteria developed in this paper (corresponding to the last row in Table 1). We call this approach adaptive cuts.
As a benchmark, we use a version of the approach proposed by Fischetti et al. (2010) that strengthens the resulting cuts without additional computational effort using the information represented by the matrix H (thereby in particular making sure that the obtained cut is always supporting). We denote this approach by the term static cuts.
More specifically, we compare the following approaches:
 “adaptive cuts”:

\(({\tilde{\omega }},{\tilde{\omega }}_0) = (H, 1)^\top ({\bar{x}}  x^*, {\bar{\eta }}  \eta ^*)\) where \(({\bar{x}}, \bar{\eta })\) is a (suboptimal) feasible solution computed based on a current upper bound from the Benders Decomposition algorithm.
 “static cuts”:

\(({\tilde{\omega }},{\tilde{\omega }}_0) = (H, 1)^\top (\mathbbm {1},1)\).
As mentioned above, the computational effort for the cutgenerating LP in both approaches is almost identical: The only difference is that in the “adaptive cuts” approach, we use a different objective function, which can be obtained from the result of the previous iteration via a simple matrixvector multiplication.
We have implemented both approaches in C++ using Gurobi 7.5. For our computations, we used ten CPU cores running at 2.4 GHz with 45 GB of main memory.
To solve master problem and subproblems, we use the dual simplex algorithm (i. e., the version of the simplex algorithm that maintains dual feasibility while pivoting between bases). In each iteration, we warmstart all problems using the optimal basis from the previous iteration and solve all subproblems in parallel on the available cores. Beyond this, we run the solution algorithm with default settings, i. e., we did not undertake any computational optimizations with respect to either the algorithm itself or the solution method of master and subproblems. In particular, also the specific update mechanism use for the adaptive cuts approach should be seen as an illustrative example rather than a performanceoptimized prescription.
As a performance measure, we use the time to reach different thresholds for the relative duality gap, i. e., the gap between upper and lower bound relative to the optimal objective value. This takes into account that in practical applications, one is often satisfied with a solution that is guaranteed to be within a certain tolerance of the optimal solution (e. g., 0.1 %), rather than a strictly optimal solution. Our results show that for any desired gap, the adaptive cuts selection approach performs substantially better than the static cuts approach.
The results can be inspected in detail in Table 2: The adaptive cuts approach generally reaches any given optimality threshold by a factor of 2–3 faster than the static cuts approach. While the general result is very consistent across all instances, differences in the magnitude of the advantage exist: The benefit of the adaptive cuts approach tends to be larger in more difficult instances which overall take longer to solve.
To further visualize our results, Fig. 5 compares the two approaches with respect to the progression of the average duality gap over all instances based on the (larger) European power system model. The plot confirms our observation from Table 2: The adaptive cuts approach reduces the duality gap by a factor of 2–3 faster than the static cuts approach.
Outlook
We conclude with an outlook on interesting research questions raised by the results presented in this paper.
Our theoretical results clearly point to a computational advantage from improving the parametrization of cutgenerating LPs in Benders decomposition and we have demonstrated this advantage in the context of Capacity Expansion Problems for electrical power systems. This holds despite the fact that we have not performed any “fine tuning” of parameters beyond what is immediately implied by our theoretical results. A broader computational study of such optimizations (for which we point out some ideas below), as well as of general performance across other types of problem instances would certainly be worthwile.
In a generic implementation of Benders decomposition, feasible solutions are used primarily to decide when the algorithm has converged sufficiently close to the optimal solution. By Theorem 2.2, however, any such solution can furthermore be used to derive a subproblem objective which satisfies the prerequisites of both Corollary 2.3 and Theorem 3.3. Together with Theorem 3.5, they thus result in the generation of cuts which are always supporting, almost always support a facet and (if \(\pi _0 < 0\)) are also paretooptimal. In our computational experiments, these cuts proved to be very useful in improving the performance of a Benders decomposition algorithm. Since information from a feasible solution can thus be used within the cut generation, it makes sense to investigate more closely the possibilities how such a solution can be obtained during the algorithm. This is likely to be very problemspecific, but some general ideas could be:

How is the information from feasible solutions computed in different iterations best aggregated? Does it make sense to use e. g. a stabilization approach or a convex combination with some other choices for \((\omega ,\omega _0)\), e. g. from previous iterations? This corresponds to the method used by Papadakos (2008) in their empirical study.

More broadly, what different methods can be used to generate feasible solutions and what effect do different feasible solutions have on cut generation and the computational performance of the algorithm?
Furthermore, if a feasible solution is not available as the basis for a subproblem objective, the cutgenerating problem might be unbounded/infeasible. On the other hand, the approach from Fischetti et al. (2010) with \({\tilde{\omega }} = \mathbbm {1}\) yields a cutgenerating LP that is always feasible, but the resulting cut might be weaker. How can both approaches be combined in a bestpossible way? For instance, is choosing \({\tilde{\omega }} = H\omega + \varepsilon \cdot \mathbbm {1}\) as the relaxation term and letting \(\varepsilon \) go to zero a good choice?
Finally, our approach provides a clear geometric interpretation of the interaction between parametrization of the cutgenerating LP and the resulting cut normals. How can this be used to leverage apriori knowledge about the problem (or information obtained through a fast preprocessing algorithm) to improve the selection of a subproblem objective \((\omega ,\omega _0)\) from a set of cuts satisfying the same quality criteria (e.g. that are all facetdefining)?
Notes
 1.
Note that the specific form of the alternative polyhedron depends on the formulation of the basic optimization problem (1.1). In our case, a point \(P(x^*, \eta ^*)\) certifies infeasibility of (1.4) by the following observation: Let y be any vector contained in the set from (1.4). Then, for any \((\gamma ,\gamma _0) \ge 0\) it holds that \(\gamma ^\top A y + \gamma _0 d^\top y \le \gamma ^\top (bHx^*)+\gamma _0\eta ^*\). On the other hand, any \((\gamma , \gamma _0) \in P(x^*,\eta ^*)\) implies \(\gamma ^\top A y + \gamma _0 d^\top y = 0 > 1 = \gamma ^\top (bHx^*)+\gamma _0\eta ^*\), independently of the choice of y.
 2.
 3.
Since we use additional \(\gamma \) variables for the description, it is essentially an extended formulation.
 4.
The result can also be derived from Balas (1998, Theorem 4.5) in the context of disjunctive cuts.
 5.
\({{\,\mathrm{relint}\,}}(S)\) (\({{\,\mathrm{relbd}\,}}(S)\)) denote the interior (boundary) relative to the affine hull of S.
References
Balas E (1975) Facets of the knapsack polytope. Math Program 8(1):146–164
Balas E (1998) Disjunctive programming: properties of the convex hull of feasible points. Discrete Appl Math 89(1):3–44
Balas E, Ivanescu PL (1964) On the generalized transportation problem. Manag Sci 11(1):188–202
Benders JF (1962) Partitioning procedures for solving mixedvariables programming problems. Numer Math 4(1):238–252
Conforti M, Wolsey LA (2018) Facet‘ separation with one linear program. Math Program Ser A 178:1–20
Cook W, Cunningham W, Pullyblank W, Schrijver A (1998) Combinatorial optimization. Series in discrete mathematics and optimization. Wiley
Cornuéjols G, Lemaréchal C (2006) A convexanalysis perspective on disjunctive cuts. Math Program Ser A 106(3):567–586
Fischetti M, Salvagnin D, Zanette A (2010) A note on the selection of Benders cuts. Math Program Ser B 124(1–2):175–182
Gleeson J, Ryan J (1990) Identifying minimally infeasible subsystems of inequalities. ORSA J Comput 2(1):61–63
Korte B, Vygen J (2008) Combinatorial optimization, 4th edn. Springer
Magnanti TL, Wong R (1981) Accelerating Benders decomposition: algorithmic enhancement and model selection criteria. Oper Res 29(3):464–484
Nemhauser GL, Wolsey LA (1988) Integer and combinatorial optimization. Wiley
Papadakos N (2008) Practical enhancements to the MagnantiWong method. Oper Res Lett 36(4):444–449
Rahmaniani R, Crainic TG, Gendreau M, Rei W (2017) The Benders decomposition algorithm: a literature review. Eur J Oper Res 259(3):801–817
Rockafellar RT (1970) Convex analysis. Princeton University Press
Schaber K, Steinke F, Hamacher T (2012) Transmission grid extensions for the integration of variable renewable energies in Europe: Who benefits where? Energy Policy 43:123–135
Sheral HD, Lunday BJ (2013) On generating maximal nondominated Benders cuts. Ann Oper Res 210(1):57–72
Stursberg P (2019) On the mathematics of energy system optimization. PhD thesis. Technische Universität München
Vanderbeck F, Wolsey LA (2010) Reformulation and Decomposition of Integer Programs. In: Jünger M, Liebling TM, Naddef D, Nemhauser GL, Pulleyblank WR, Reinelt G, Rinaldi G, Wolsey LA (eds) 50 years of integer programming 19582008: from the early years to the stateoftheart. Springer, pp 431–502. isbn: 9783540682790
Acknowledgements
The authors thank their colleagues Magdalena Stüber and Michael Ritter for their invaluable support as well as the reviewers and the associate editor for their helpful comments and the corresponding discussion, which have contributed to further improve the paper. Paul Stursberg acknowledges funding from Deutsche Forschungsgemeinschaft (DFG) through TUM International Graduate School of Science and Engineering (IGSSE), GSC 81, and by the German Federal Ministry for Economic Affairs and Energy (FKZ 03ET4029) on the basis of a decision by the German Bundestag.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Brandenberg, R., Stursberg, P. Refined cut selection for benders decomposition: applied to network capacity expansion problems. Math Meth Oper Res 94, 383–412 (2021). https://doi.org/10.1007/s00186021007568
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00186021007568
Keywords
 Benders decomposition
 Decomposition methods
 Cutting planes
 Reverse polar set
 Alternative polyhedron
 Pareto optimal cuts
 Facet defining cuts