Polynomial Optimization:
Enhancing RLT relaxations with Conic Constraints
Abstract
Conic optimization has recently emerged as a powerful tool for designing tractable and guaranteed algorithms for non-convex polynomial optimization problems. On the one hand, tractability is crucial for efficiently solving large-scale problems and, on the other hand, strong bounds are needed to ensure high quality solutions. In this research, we investigate the strengthening of RLT relaxations of polynomial optimization problems through the addition of nine different types of constraints that are based on linear, second-order cone, and semidefinite programming to solve to optimality the instances of well established test sets of polynomial optimization problems. We describe how to design these conic constraints and their performance with respect to each other and with respect to the standard RLT relaxations. Our first finding is that the different variants of nonlinear constraints (second-order cone and semidefinite) are the best performing ones in around of the instances. Additionally, we present a machine learning approach to decide on the most suitable constraints to add for a given instance. The computational results show that the machine learning approach significantly outperforms each and every one of the nine individual approaches.
Keywords. Reformulation-Linearization Technique (RLT), Global Optimization, Polynomial Programming, Conic Optimization, Machine Learning.
1 Introduction
The large volume of theoretical and computational research on conic optimization has led to important advances over the last few years in the efficiency and robustness of the associated algorithmic procedures to solve them. Leading state-of-the-art mixed integer linear programming (MILP) solvers such as Gurobi (Gurobi Optimization, 2022), CPLEX (IBM Corp., 2022), and Xpress (FICO, 2022) have recently added functionalities that allow to efficiently solve second-order cone programming (SOCP) problems and, further, Mosek (MOSEK ApS, 2022, Andersen and Andersen, 2000) has positioned itself as a reliable solver for general semidefinite programming (SDP) problems.
Importantly, despite the convex nature of conic optimization problems, they are also proving to be a powerful tool in the design of branch-and-bound algorithms for general non-convex mixed integer nonlinear programming (MINLP) problems and, specially, polynomial optimization problems. The latter are very general and encompass many problems arising in operations research: integer programming, linear programming, mixed integer programming, and quadratic optimization, to name a few. The developments of the last several decades have shown that conic optimization is a central tool in addressing non-convexities. These advances have been more prominent in the case of polynomial optimization problems, and even more so in the particular case of (quadratically-constrained) quadratic optimization problems, where a variety of convex relaxations have been thoroughly studied (Shor, 1987, Ghaddar et al., 2011b , Burer and Ye, 2020, Bonami et al., 2019, Elloumi and Lambert, 2019). These conic-based relaxations include semidefinite programming and second-order cone programming as their powerhouses. They often provide fast and guaranteed approaches for computing bounds on the global value of the non-convex optimization problem at hand. For instance, Lasserre, (2001) introduced semidefinite relaxations corresponding to liftings of the polynomial programs into higher dimensions. The construction is motivated by results related to representations of non-negative polynomials as sum-of-squares (Parrilo, 2003) and the dual theory of moments. This led to the development of systematic approaches for solving polynomial optimization problems to global optimality, the main limitation of these approaches currently being that they are computationally prohibitive in general.
Nowadays, state-of-the-art solvers tackle MINLP problems and, in particular, polynomial optimization ones through different relaxations to be solved at the nodes of the branch-and-bound tree. A crucial direction of current research focuses on integrating tighter and tighter relaxations while preserving reasonable computational properties, with relaxations that build upon different families of conic constraints becoming increasingly important. One relaxation for polynomial optimization problems is based on the Reformulation-Linearization Technique (RLT) and the most common approach to strengthen this relaxation comes from the identification of linear SDP-based cuts that tighten the RLT relaxation as the algorithm progresses (Sherali et al., 2012, Baltean-Lugojan et al., 2019, González-Rodríguez et al., 2020). In this paper we are interested in the alternative approach of directly adding SOCP and SDP constraints to the relaxations as in Burer and Vandenbussche, (2008) and Buchheim and Wiegele, (2013). Probably the main reason why this approach has received less attention so far is that the resulting algorithms need to rely on SDP solvers which, until recently, were probably not reliable enough and too expensive computationally in some instances. One of the primary goals of this paper is to show that the situation has changed, and that general branch-and-bound schemes based on the solution of SOCPs and SDPs at each and every node of the branch-and-bound tree can be very competitive and even superior to previous approaches.
In order to achieve the above goal, in this paper we introduce and compare linear SDP-based constraints as well as SOCP and SDP constraints. Importantly, their definition ensures that they are efficient in the sense of preserving the size and sparsity of the original RLT-based relaxation for polynomial optimization problems introduced in Sherali and Tuncbilek, (1992) and further refined in Dalkiran and Sherali, (2013). We consider a total of nine different versions of constraints to be added: one based on linear SDP-based cuts, four based on SOCP constraints, and four based on SDP constraints. These conic constraints are then integrated into the polynomial optimization solver RAPOSa (González-Rodríguez et al., 2020), whose core is an RLT-based branch-and-bound algorithm. As a second step, we then develop thorough computational studies on well established benchmarks including randomly generated instances from Dalkiran and Sherali, (2016), MINLPLib instances (Bussieck et al., 2003), and QPLIB instances (Furini et al., 2018), which provide a wide variety of classes of polynomial optimization problems. One of the main findings is that, in around of the instances, the best performance is achieved by one of the versions explicitly incorporating SOCP and SDP conic constraints. The remaining is split quite evenly between the baseline RLT and the version that incorporates SDP-based linear cuts. Importantly, the analysis also allowed to identify particular classes of problems where one specific family of SOCP/SDP conic constraints is consistently superior to the linear versions.
To the best of our knowledge, our contribution represents one of very few implementations of branch-and-bound schemes with conic relaxations for broad classes of problems, with the added value of the generality of the resulting scheme, since it can be applied to any given polynomial optimization problem. The most related approaches are Burer and Vandenbussche, (2008) for non-convex quadratic problems with linear constraints and Buchheim and Wiegele, (2013) for unconstrained mixed-integer quadratic problems. In Burer and Vandenbussche, (2008), the authors present a computational analysis in which they compare the relative performance of different SDP relaxations, with the main highlight being that “only a small number of nodes are required” to fully solve the problems at hand. Buchheim and Wiegele, (2013) introduce a new branch-and-bound algorithm, Q-MIST (Quadratic Mixed-Integer Semidefinite programming Technique), and develop a thorough computational analysis in which they show that, for a wide variety of families of instances within the scope of their algorithm, Q-MIST notably outperforms Couenne (Belotti et al., 2009). It is worth noting that, if looking at specific classes of optimization problems, then one can find additional successful implementations of branch-and-bound algorithms with the inclusion of tailor-made conic constraints, such as Ghaddar and Jabr, (2019) for optimal power flow problems and combinatorial optimization problems, Krislock et al., (2017) and Rendl et al., (2010) for maximum cut problems, Piccialli et al., (2022) for minimum sum-of-squares clustering, and Ghaddar et al., 2011a for maximum -cut problems.
Last, but not least, in our computational experiments we also observe that there is a lot of variability in the best performing version of conic constraints for the different instances, with each version beating the rest for a non negligible number of instances. This observation motivates the last contribution of this paper, in which we exploit this variability by learning to choose the best version among our portfolio of different constraints. Building upon the framework in Ghaddar et al., (2022), we show that the resulting machine learning version significantly outperforms each and every one of the underlying versions. This last contribution naturally fits into the rapidly emerging strand of research on “learning to optimize”, whose advances are nicely presented in the survey papers Lodi and Zarpellon, (2017) and Bengio et al., (2021). The closest approach to this last part of our contribution is Baltean-Lugojan et al., (2019), in which deep neural networks are used to rank SDP-based cuts for quadratic problems. Then, only the top scoring cuts are added, aiming to obtain a good balance between the tightness and the complexity of the relaxations. A common feature of the SDP-based cuts used in Baltean-Lugojan et al., (2019) and those in this paper is that they are designed with a big emphasis in sparsity considerations (number of nonzeros in the constraints), with the goal of obtaining computationally efficient relaxations. From the point of view of the learning process the approaches are quite different, since they focus on one type of constraints (the SDP-based cuts) and they want to learn the best SDP-based cuts to add at a given node, whereas we want learn to choose for any given instance which conic constraints to include in the relaxations.
The contribution of this paper can be summarized as follows. First, we define and show the potential of different SOCP and SDP strengthenings of the classic RLT relaxations for general polynomial optimization problems in a branch-and-bound scheme. Second, we design a machine learning approach to learn the best strenghtening to use on a given instance and obtain promising results for the resulting algorithm.
The remainder of this paper is organized as follows. In Section 2 we present a brief overview of the classic RLT scheme. In Section 3 we describe the different families of conic constraints that will be integrated within the baseline RLT implementation. In Section 4 we present a first series of computational results. Then, in Section 5 we show how the conic constraints can be further exploited within a machine learning framework. Finally, we conclude in Section 6 and discuss future research directions.
2 Foundations of the RLT Technique
The Reformulation-Linearization Technique was originally developed in Sherali and Tuncbilek, (1992). It was designed to find global optima in polynomial optimization problems of the following form:
(PO) |
where denotes the set of variables, each is a polynomial of degree and is a hyperrectangle containing the feasible region. Then, is the degree of the problem and represents all possible monomials of degree .
The Reformulation-Linearization Technique consists of a branch and bound algorithm based on solving linear relaxations of the polynomial problem (PO). These linear relaxations are built by working on a lifted space, where each monomial of the original problem is replaced with a corresponding RLT variable. For example, associated to monomials of the form and one would define the RLT variables and , respectively. More generally, RLT variables are defined as
(1) |
where is a multiset containing the information about the multiplicity of each variable in the underlying monomial. Then, at each node of the branch-and-bound tree, one would solve the corresponding linear relaxation. Whenever we get a solution of a linear relaxation in which the identities in (1) hold, we get a feasible solution of (PO). Otherwise, the violations of these identities are used to choose the branching variable.
In order to get tighter relaxations and ensure convergence, new constraints, called bound-factor constraints, must be added. They are of the following form:
(2) |
Thus, for each pair of multisets and such that and , the corresponding bound-factor constraint is added to the linear relaxation.
Dalkiran and Sherali, (2013) show that it is not necessary to add all bound-factor constraints to the linear relaxations, since certain subsets of them are enough to ensure the convergence of the algorithm. More precisely, they proved that convergence to a global optimum only requires the inclusion in the linear relaxation of those bound-factor constraints where is a monomial that appears in (PO), regardless of its degree. Further, they also showed that convergence is also preserved if, whenever the bound-factor constraints associated to a monomial are present, all bound-factor constraints associated to monomials are removed. Motivated by these results, -sets are defined as those monomials of degree greater than one present in (PO) which, moreover, are not included in any other monomial (multiset inclusion). Consider the polynomial programming problem
(3) |
The monomials with degree greater than one are ,, , , and . Since is included in , it is removed. Therefore, the -sets are , , , and .
The analysis developed in this paper builds upon the above theoretical results from Dalkiran and Sherali, (2013) and, therefore, only the bound-factor constraints associated to -sets are incorporated to the linear relaxations of (PO). Such relaxations are less tight but, on the other hand, they are smaller in size and, hence, faster to solve. Note that the use of -sets reduces not only the number of constraints but, more importantly, also the number of RLT variables. As it can be seen in Dalkiran and Sherali, (2013) and González-Rodríguez et al., (2020), the performance of the RLT algorithm is clearly superior when the -set approach is followed.
3 Conic Enhancements of RLT
As already discussed earlier, the use of semidefinite programming to improve the performance of branch-and-bound schemes is not new, and Baltean-Lugojan et al., (2019) provides a thorough and up-to-date review of the field. Typically, the goal is to rely on semidefinite programming to tighten the relaxations of the original non-convex optimization problems targeted by a branch-and-bound algorithm. We start this section by reviewing the main ingredient of such methods and then present various families of SDP-driven constraints that can be incorporated into the RLT relaxations in an efficient way. In particular, they should preserve the underlying dimensionality and sparsity of the problem, which is crucial for these approaches to be competitive.
The main ingredient that has to be specified is the matrix or matrices on which positive semidefiniteness is to be imposed. To each (multi-)set of variables , with , one can associate a vector (resulting from the concatenation of all the variables in , including repetitions). To any such vector one can associate matrix , which is trivially positive semidefinite. Now, let be the matrix obtained when each monomial in is replaced by the corresponding RLT variable in the lifted space. The constraint is a valid cut because it never removes feasible solutions of (PO) and, hence, it does not compromise convergence of the RLT algorithm to a global optimum. In practice, vectors of the form are often preferred, since they result in matrices containing also variables in the original space and not only RLT variables, leading to tighter relaxations. In Sherali et al., (2012), for instance, the authors discuss different ways of defining for a given . Specifically, they mainly work with , , and , where is defined by concatenating also monomials of degree greater than one, while ensuring that no monomial in the resulting matrix has degree larger than , the degree of (PO).
We now move to the definition of the specific SDP-driven constraints for the RLT algorithm that constitute the subject of study in this paper.
3.1 Linear SDP-based Constraints
In Sherali et al., (2012), the authors associate linear cuts to the constraints of the form as follows. At each node of the branch-and-bound tree, the positive semidefiniteness of the chosen matrices is assessed at the solution of the corresponding relaxation. Given a negative eigenvalue of one such matrix with as its associated eigenvector, then the valid cut can be added to the linear relaxation to separate the current solution. A potential drawback of these cuts is that they may be very “dense”, in the sense of involving a large number of variables, which may increase the solving time of the relaxations. Thus, as already discussed in Sherali et al., (2012), it is important to carefully choose matrices.
The sparsity of the cuts is particularly important if the RLT algorithm is being run with the -set approach since, in general, the resulting cuts might involve monomials not contained in any -set. This would require to include additional RLT variables in the relaxations (and the corresponding bound-factor constraints), increasing the size and potentially reducing the sparsity of the relaxations. Here we follow González-Rodríguez et al., (2020), where the authors consider, at each node, all matrices obtained from vectors associated with the different -sets of (PO), which lead to sparse cuts that essentially preserve the dimensionality of the resulting relaxations. The authors present a detailed computational analysis, comparing different approaches to add an inherit cuts. We adopt the best performing version, which consists of using vector and inheriting all cuts from one node to all its descendants.
In Sherali et al., (2012) different methods are discussed to efficiently look for the negative eigenvalues of the matrices. One such approach, that we follow here, consists of dividing each matrix in overlapping submatrices (each matrix shares its first 5 rows with the preceding one) and add a valid cut for each negative eigenvalue they have. This approach, on top of being computationally cheap, leads to even sparser cuts.
The above discussion regarding the adequacy of building constraints based on -sets, in order to obtain sparser constraints (number of nonzero coefficients) and to preserve the size (number of variables) of the resulting relaxations, also applies to the conic constraints defined in the following subsections which, therefore, also build upon -sets.
3.2 SDP Constraints
We next describe two approaches to tighten the classic RLT relaxations by directly adding semidefinite constraints. They just differ in the matrices on which positive semidefiniteness is imposed.
- Approach 1.
-
For each -set , is used to define the matrix and the constraint . Thus, . Note that, whenever contains only one variable (possibly multiple times), this would result in a trivial constraint and, therefore, these constraints are disregarded with one exception: if , then is replaced with . With this exception, this approach is mathematically equivalent for quadratic problems to the SOCP approach we present in Section 3.3 below.
- Approach 2.
-
For each -set , is used to define the matrix and the constraint .
Preliminary analysis have shown that constraints building upon , although they lead to tighter relaxations, generate significantly bigger matrices and increase the complexity of the resulting SDP problems. To provide an example of both approaches, consider the polynomial optimization problem in Equation (3). Then, the semidefinite constraints added with the above approaches are the following ones:
Approach 1 |
. |
Approach 2 |
. |
3.3 SOCP Constraints
We now describe the second-order cone constraints which, with respect to the SDP ones, lead to looser relaxations but, on the other hand, can be solved more efficiently by state-of-the-art optimization solvers. For each -set and each pair of variables present in , , we define the following second-order cone constraint:
(4) |
We argue now why these constraints are valid cuts, i.e., they never remove solutions feasible to (PO). Constraint (4) can be equivalently rewritten as . Then, given a solution of a linear relaxation satisfying the RLT identities in (1), the above condition reduces to , which is trivially true. Note that constraints in (4) are trivially true if and, hence, whenever we have a variable appearing twice or more in , we instead add the second-order constraint
(5) |
which is equivalent to and, for solutions satisfying (1), is again trivially true. Consider again the polynomial optimization problem in Equation (3). Then, the SOCP constraints added are the following ones:
3.4 Binding SOCP and SDP Constraints
Since solving SOCP or SDP problems is usually more time-consuming than solving linear programming problems, we define a new approach in order to reduce the time needed for solving the resulting RLT relaxation with SOCP or SDP constraints. This consists of checking which conic constraints (second-order cone or semidefinite) are binding after solving the first relaxation, i.e., this is done only once, at the root node. Thereafter, only these binding constraints are used to tighten the future linear relaxations. This approach significantly reduces the number of second-order cone or semidefinite constraints in the relaxations and, although these new relaxations are not as tight, one might expect that the binding constraints at the root node tend to be the most important ones in subsequent relaxations, at least in the first phase of the algorithm. We assess the trade-off between the difficulty of solving the relaxations and how tight they are in the computational analysis in the next section.
4 Computational Results
4.1 Testing Environment
All the computational analyses reported in this paper have been performed on the supercomputer Finisterrae III, provided by Galicia Supercomputing Centre (CESGA). Specifically, we use nodes powered with 32 cores Intel Xeon Ice Lake 8352Y CPUs with 256GB of RAM connected through an Infiniband HDR network, and 1TB of SSD.
Regarding the datasets, we use three different sets of problems. The first one, DS, is taken from Dalkiran and Sherali, (2016) and consists of 180 instances of randomly generated polynomial programming problems of different degrees, number of variables, and density. The second dataset comes from the well known benchmark MINLPLib (Bussieck et al., 2003), a library of Mixed-Integer Nonlinear Programming problems. We have selected from MINLPLib those instances that are polynomial programming problems with box-constrained and continuous variables, resulting in a total of 166 instances. The third dataset comes from another well known benchmark, QPLIB (Furini et al., 2018), a library of quadratic programming instances, for which we made a selection analogous to the one made for MINLPLib, resulting in a total of 63 instances. Hereafter we refer to the first dataset as DS, to the second one as MINLPLib, and to the third one as QPLIB.111Instances from DS dataset can be downloaded at https://raposa.usc.es/files/DS-TS.zip, instances from MINLPLib dataset can be downloaded at https://raposa.usc.es/files/MINLPLib-TS.zip, and instances from QPLIB dataset can be downloaded at https://raposa.usc.es/files/QPLIB-TS.zip.
We develop our analysis by building upon the global solver for polynomial optimization problems RAPOSa (González-Rodríguez et al., 2020). Regarding the auxiliary solvers, RAPOSa uses i) Gurobi for the linear relaxations ii) Gurobi or Mosek for the SOCP relaxations, and iii) Mosek for the SDP relaxations. The main objective of the thorough numerical analysis developed in this and in the following section is to assess the performance of different SOCP/SDP conic-driven versions of RAPOSa with respect to two more traditional ones: basic RLT and RLT with linear SDP-based cuts. More precisely, the full set of ten different versions is as follows:
-
•
RLT: standard RLT algorithm (with -sets).
-
•
SDP-Cuts: linear SDP-based cuts added to RLT.
-
•
SOCPG: SOCP constraints added to the RLT relaxation and solved with Gurobi.
-
•
SOCPG,B: same as above, but using only constraints that were binding at the root node.
-
•
SOCPM: SOCP constraints added to the RLT relaxation and solved with Mosek.
-
•
SOCPM,B: same as above, but using only constraints that were binding at the root node.
-
•
SDP1: SDP constraints added to the RLT following Approach 1.
-
•
SDP1,B: same as above, but using only constraints that were binding at the root node.
-
•
SDP2: SDP constraints added to the RLT following Approach 2.
-
•
SDP2,B: same as above, but using only constraints that were binding at the root node.
For each instance and each one of the above versions, we run RAPOSa with a time limit of one hour.
4.2 Numerical Results and Analysis
The main goal of the numerical analysis in this section is to show the potential of SOCP/SDP conic-constraints to improve upon the performance of more classic implementations of RLT, such as RLT and SDP-Cuts. The measures used to evaluate the performance of the different versions of RAPOSa are and , two performance indicators introduced in Ghaddar et al., (2022) that capture the pace at which a given algorithm closes the gap or, more precisely, the pace at which it increases the lower bound along the branch-and-bound tree. To compute we use the following formula:
(6) |
Then, is just a normalized version of with values in . It is computed, for each version of the solver/algorithm, by dividing the best (smallest) pace among all versions to be compared by the pace of the current one.
As thoroughly discussed in Ghaddar et al., (2022), and are natural measures that allow to compare the performance of different solvers/algorithms on all the instances of a test set at once, regardless of their difficulty and of how many versions of the underlying solver/algorithm have solved them to optimality. This is different from more common approaches, where the running time is used to evaluate performance on instances solved by all versions, the optimality gap is used for those instances solved by none, and where some decision has to be made regarding those instances solved by some but not all of the versions of the solver/algorithm.
Figure 1 shows, for the different sets of problems, the percentage of instances in which each one of the ten versions is the best one. We can see that, quite consistently across the three test sets, a version with either SOCP or SDP constraints is the best one in around of the instances. This is in itself one of the main highlights of this paper: RLT versions incorporating nonlinear SOCP/SDP conic constraints can improve the performance of RLT-based algorithms in half of the instances of the sets of problems under consideration.

Figure 1 contains more valuable information. First, the versions with SOCP constraints are the best ones much more often than those with SDP constraints. Regarding the solvers for SOCP versions, Gurobi performs notably better in MINLPLib instances, whereas Mosek is overwhelmingly better on DS and is also the best one for QPLIB instances. When comparing binding versions with their non-binding counterparts we can see that, for SDP versions, SDP1,B and SDP2,B are the best ones significantly more often than SDP1 and SDP2, respectively. In the case of SOCP versions, there is no clear winner between binding and non-binding versions. Finally, regarding the RLT versions without conic constraints, RLT and SDP-Cuts, we can see that each of them turns out to be the best option in around of the instances, with SDP-Cuts looking preferable in DS and QPLIB, whereas RLT is the best one three times as much in MINLPLib.
Importantly, Figure 1 and the preceding discussion show that there is a lot of variability, with all ten versions showing up as the best choice for a non-negligible percentage of instances. Further, this variability also follows different patterns for the different sets of problems, which motivates the approach taken in Section 5 below, where we use machine learning techniques to try to learn to choose in advance the most promising RLT version for a given instance.

A natural question given the results in Figure 1 is whether or not there are specific subclasses of instances in the different test sets where a certain RLT version is noticeably dominant. A deeper analysis of the results shows that this is indeed the case, as presented in Figure 2. First, when looking at instances in DS with high density (larger that ) we can see that the versions relying on SOCP constraints and Mosek as a solver, SOCPM and SOCPM,B, are the best performing ones in around of the instances; RLT goes down to around , SDP-Cuts to around , and SDP2 and SDP2,B split the remaining . Second, we selected from MINLPLib set all instances whose name contains the word “water”, which are problems related to the design of water networks (Castro and Teles, 2013, Teles et al., 2012) and wastewater treatment systems (Castro et al., 2009, 2007). We have that in 15 out of the 28 resulting instances, the best option is one of the following SDP-based ones: SDP1,B, SDP2, and SDP2,B. Again, the instances in which RLT or SDP-Cuts are the best option are less than . The reasons behind the notoriously good performance of the versions based on second order cone constraints for high density problems in DS and of those based on positive semidefinite constraints for “water”-related instances in MINLPLib are definitely worth studying more deeply, but such an analysis is beyond the scope of this paper.
All | DS high density | water | |
---|---|---|---|
RLT (across all instances) | 12.69 | 0.82 | 33994.03 |
Optimal version (instance by instance) | 6.28 | 0.21 | 2546.88 |
Improvement | 50.5% | 74.4% | 92.5% |
Despite the results shown in Figure 1 and Figure 2, it is important to ensure that the variability is not spurious. For instance, it might be that all RLT versions performed very similarly to one another, which would turn most of the above discussion meaningless. In Table 1 we present a concise summary of the geometric mean of for the complete set of instances and for the two special subclasses identified above. The first row measures the performance of RLT while the second row measures the performance of a hypothetical RLT version capable of choosing the best performing RLT version in each and every instance. The first column shows that, on aggregate on the whole set of instances, this hypothetical and optimal version would divide the pace by two, i.e., a substantial improvement of . This improvement is much more pronounced for high density problems in DS, in which the pace gets divided by four ( improvement), and even more so for “water”-related instances in MINLPLib where the pace becomes more than ten times smaller ( improvement).
5 Machine Learning for different Conic Constraints
In this section we want to exploit the wide variability in the performance of the different RLT versions shown above to try to learn in advance which one should be chosen for a given instance. The goal is to design a machine learning procedure that can be trained using the different features of the instances and then choose the most promising RLT version when confronted with a new instance. The performance of the hypothetical “Optimal version” in Table 1 represents an upper bound on the improvement that can be attained by such a machine learning version.
We follow the framework in Ghaddar et al., (2022), where the authors use learning techniques to improve the performance of the RLT-based solver RAPOSa by learning to choose between different branching rules. The improvements reported there are substantial, with the machine learning version delivering improvements of up to with respect to the best original branching rule. Table 2 presents the full list of the input variables (features).222VIG and CMIG stand for two graphs that can be associated to any given polynomial optimization problem: variables intersection graph and constraints-monomials intersection graph, and whose precise definitions is given in Ghaddar et al., (2022). They capture diverse characteristics of each instance and are a key ingredient of the machine learning framework, which consists of predicting the performance () of each branching rule on a new instance based on a regression analysis of its performance on the training instances. Then, the rule with a highest predicted performance for the given instance is chosen.
Variables | No. of variables, variance of the density of the variables |
Average/median/variance of the ranges of the variables | |
Average/variance of the no. of appearances of each variable | |
Pct. of variables not present in any monomial with degree greater than one | |
Pct. of variables not present in any monomial with degree greater than two | |
Constraints | No. of constraints, Pct. of equality/linear/quadratic constraints |
Monomials | No. of monomials |
Pct. of linear/quadratic monomials, Pct. of linear/quadratic RLT variables | |
Average pct. of monomials in each constraint and in the objective function | |
Coefficients | Average/variance of the coefficients |
Other | Degree and density of PO |
No. of variables divided by no. of constrains/degree | |
No. of RLT variables/monomials divided by no. of constrains | |
Graphs | Density, modularity, treewidth, and transitivity of VIG and CMIG |
For the learning we rely on quantile regression models since as argued in Ghaddar et al., (2022), the presence of outliers and of an asymmetric behavior of (negative skewness) makes quantile regression more suitable than conventional regression models (based on the conditional mean). More precisely, we use quantile regression forests as the core tool for our analysis. Random forests were introduced in Breiman, (2001) as ensemble methods aggregate the information on several individual decision trees to make a single prediction. Quantile regression forests, introduced in Meinshausen, (2006), are a generalization of random forests that compute an estimation of the conditional distribution of the response variable by taking into account all the observations in every leaf of every tree and not just their average. As discussed in Ghaddar et al., (2022), one advantage of random forests is that the reporting of the results can be done for the complete data set in terms of Out-Of-Bag predictions; there is no need to do the reporting with respect to the underlying training and test sets. The statistical analysis is conducted in programming language R (R Core Team, 2021), using library ranger (Wright and Ziegler, 2017). Finally, the learning is conducted jointly on all sets of instances.
Table 3 shows the remarkable improvement obtained with the machine learning (ML) version of RLT that chooses, for each given instance, the most promising version of the 10-version portfolio. Considering all instances, the ML version improves with respect to RLT, being the upper bound for learning . In instances with high densities from DS test set, it improves a remarkable out of the optimal , and in “water”-related instances in MINLPLib it improves out of . It is worth noting that the size of these improvements is comparable, and even superior, to those obtained in Ghaddar et al., (2022) when this learning scheme was introduced to learn to choose between branching rules, which shows the robustness of the proposed approach.
All | DS high density | water | |
---|---|---|---|
RLT (across all instances) | 12.69 | 0.82 | 33994.03 |
ML-based version | 8.51 | 0.25 | 14727.22 |
Optimal version (instance by instance) | 6.28 | 0.21 | 2546.88 |
Improvement after learning | 32.9% | 69.5% | 56.7% |
Optimal improvement (upper bound for learning) | 50.5% | 74.4% | 92.5% |

Figure 3 represents, side by side, how often each of the ten RLT versions is selected by the ML version and by the optimal one. We can see that the behavior of the former mimics quite well that of the latter in the three sets of instances. Given that the learning is conducted jointly on the whole set of instances, the fact that the ML version adapts to the instances in DS, MINLPLib, and QPLib, is reassuring about the quality of the learning process. In particular, the SOCP versions with Mosek are primarily chosen for DS test set and the SDP versions are mainly chosen for QPLIB test set, the one in which they are optimal more often. In Figure 4 we further explore the behavior for the high density problems in DS and for “water”-related instances in MINLPLib. We see that, again, the ML version mimics the patterns of the optimal version. The dominant version in DS, SOCPM, is chosen in almost of the instances. Similarly, the three SDP dominant versions in “water”-related instances are chosen almost of the time. The dominant version in DS, SOCPM, is chosen in almost of the instances.

Figure 5 presents boxplots summarizing the performance according to for all instances and for each individual test set. Recall that, by definition, values close to 1 mean that the corresponding version is almost the best one, whereas values close to 0 mean that its pace is much worse than the best one. We can see that, although RLT and SDP cuts versions are, on aggregate, the best ones in all three test sets, they are significantly outperformed by the ML version in the three of them. The improvement is particularly noticeable in DS, where the learning criterion is, by far, better than all underlying versions. This observation is further reinforced by the performance profiles (Dolan and Moré, (2002)) represented in Figure 6. Again, the ML version clearly outperforms all others, specially in DS instances.


6 Conclusions and Future Research
The main contribution of this paper is to show that the solution to global optimality via branch-and-bound schemes of non-convex optimization problems and, in particular, polynomial optimization ones, can benefit from tightening the underlying relaxations with conic constraints. We explore different families of such constraints, building upon either second-order cones or positive semidefiniteness. We also show that the potential of these conic constraints can be successfully exploited by embedding it into a learning framework. The main goal is to predict which is the most promising type of constraints to add to the RLT relaxation at each node of the underlying branch-and-bound algorithm when confronted with a new instance. The results in Section 4 show that the versions with SOCP/SDP conic constraints deliver consistently good results for instances in specific subclasses of problems: high density problems in DS and of those based on positive semidefinite constraints for “water”-related instances in MINLPLib.
As a future step, one may wonder to what extent one might get an even superior performance if the learning analysis was further specialized for the current setting: for example including features capturing some “conic” characteristics of the polynomial optimization problems and fine-tuning the regression techniques. Furthermore, an important direction for future research is to improve the understanding on the structure of these problems and the specificities that lead to the superior performance of SOCP and SDP constraints, respectively. Another direction is to investigate the number of SOCP and SDP constraints added at different nodes of the branch-and-bound tree. Additionally, we aim to extend this framework for various relaxations of polynomial optimization problems besides RLT.
Acknowledgements
This research has been funded by FEDER and the Spanish Ministry of Science and Technology through projects MTM2014-60191-JIN, MTM2017-87197-C3, and PID2021-124030NB-C32. Brais González-Rodríguez acknowledges support from the Spanish Ministry of Education through FPU grant 17/02643. Raúl Alvite-Pazó and Samuel Alvite-Pazó acknowledge support from CITMAga through proyect ITMATI-R-7-JGD. Bissan Ghaddar’s research is supported by Natural Sciences and Engineering Research Council of Canada Discovery Grant 2017-04185 and by the David G. Burgoyne Faculty Fellowship.
References
- Andersen and Andersen, (2000) Andersen, E. D. and Andersen, K. D. (2000). The Mosek Interior Point Optimizer for Linear Programming: An Implementation of the Homogeneous Algorithm. Springer US, Boston, MA.
- Baltean-Lugojan et al., (2019) Baltean-Lugojan, R., Bonami, P., Misener, R., and Tramontani, A. (2019). Scoring positive semidefinite cutting planes for quadratic optimization via trained neural networks. Technical report, Optimization-online 7942.
- Belotti et al., (2009) Belotti, P., Lee, J., Liberti, L., Margot, F., and Wächter, A. (2009). Branching and bounds tightening techniques for non-convex MINLP. Optimization Methods and Software, 24(4-5):597–634.
- Bengio et al., (2021) Bengio, Y., Lodi, A., and Prouvost, A. (2021). Machine learning for combinatorial optimization: a methodological tour d’horizon. European Journal of Operational Research, 290(2):405–421.
- Bonami et al., (2019) Bonami, P., Lodi, A., Schweiger, J., and Tramontani, A. (2019). Solving quadratic programming by cutting planes. SIAM Journal on Optimization, 29(2):1076–1105.
- Breiman, (2001) Breiman, L. (2001). Random forests. Machine Learning, 45(1):5–32.
- Buchheim and Wiegele, (2013) Buchheim, C. and Wiegele, A. (2013). Semidefinite relaxations for non-convex quadratic mixed-integer programming. Mathematical Programming, 141(1):435–452.
- Burer and Vandenbussche, (2008) Burer, S. and Vandenbussche, D. (2008). A finite branch-and-bound algorithm for nonconvex quadratic programming via semidefinite relaxations. Mathematical Programming, 113(2):259–282.
- Burer and Ye, (2020) Burer, S. and Ye, Y. (2020). Exact semidefinite formulations for a class of (random and non-random) nonconvex quadratic programs. Mathematical Programming, 181(1):1–17.
- Bussieck et al., (2003) Bussieck, M. R., Drud, A. S., and Meeraus, A. (2003). MINLPLib-a collection of test models for mixed-integer nonlinear programming. INFORMS Journal on Computing, 15:114–119.
- Castro et al., (2007) Castro, P. M., Matos, H. A., and Novais, A. Q. (2007). An efficient heuristic procedure for the optimal design of wastewater treatment systems. Resources, conservation and recycling, 50(2):158–185.
- Castro and Teles, (2013) Castro, P. M. and Teles, J. P. (2013). Comparison of global optimization algorithms for the design of water-using networks. Computers & chemical engineering, 52:249–261.
- Castro et al., (2009) Castro, P. M., Teles, J. P., and Novais, A. Q. (2009). Linear program-based algorithm for the optimal design of wastewater treatment systems. Clean Technologies and Environmental Policy, 11(1):83–93.
- Dalkiran and Sherali, (2013) Dalkiran, E. and Sherali, H. D. (2013). Theoretical filtering of RLT bound-factor constraints for solving polynomial programming problems to global optimality. Journal of Global Optimization, 57(4):1147–1172.
- Dalkiran and Sherali, (2016) Dalkiran, E. and Sherali, H. D. (2016). RLT-POS: Reformulation-linearization technique-based optimization software for solving polynomial programming problems. Mathematical Programming Computation, 8:337–375.
- Dolan and Moré, (2002) Dolan, E. D. and Moré, J. J. (2002). Benchmarking optimization software with performance profiles. Mathematical Programming, 91:201–213.
- Elloumi and Lambert, (2019) Elloumi, S. and Lambert, A. (2019). Global solution of non-convex quadratically constrained quadratic programs. Optimization methods and software, 34(1):98–114.
- FICO, (2022) FICO (2022). FICO Xpress Optimization Suite. Available at: https://www.fico.com/en/products/fico-xpress-optimization.
- Furini et al., (2018) Furini, F., Traversi, E., Belotti, P., Frangioni, A., Gleixner, A., Gould, N., Liberti, L., Lodi, A., Misener, R., Mittelmann, H., Sahinidis, N., Vigerske, S., and Wiegele, A. (2018). QPLIB: a library of quadratic programming instances. Mathematical Programming Computation, 1:237–265.
- (20) Ghaddar, B., Anjos, M. F., and Liers, F. (2011a). A branch-and-cut algorithm based on semidefinite programming for the minimum k-partition problem. Annals of Operations Research, 188(1):155–174.
- Ghaddar et al., (2022) Ghaddar, B., Gómez-Casares, I., González-Díaz, J., González-Rodríguez, B., Pateiro-López, B., and Rodríguez-Ballesteros, S. (2022). Learning for spatial branching: An algorithm selection approach. Technical report.
- Ghaddar and Jabr, (2019) Ghaddar, B. and Jabr, R. A. (2019). Power transmission network expansion planning: A semidefinite programming branch-and-bound approach. European Journal of Operational Research, 274(3):837–844.
- (23) Ghaddar, B., Vera, J. C., and Anjos, M. F. (2011b). Second-order cone relaxations for binary quadratic polynomial programs. SIAM Journal on Optimization, 21(1):391–414.
- González-Rodríguez et al., (2020) González-Rodríguez, B., Ossorio-Castillo, J., González-Díaz, J., González-Rueda, Á. M., Penas, D. R., and Rodríguez-Martínez, D. (2020). Computational advances in polynomial optimization: RAPOSa, a freely available global solver. Technical report, Optimization-online 7942.
- Gurobi Optimization, (2022) Gurobi Optimization (2022). Gurobi Optimizer Reference Manual. Available at: http://www.gurobi.com.
- IBM Corp., (2022) IBM Corp. (2022). IBM ILOG CPLEX Optimization Studio. CPLEX User’s Manual. Available at: https://www.ibm.com/es-es/products/ilog-cplex-optimization-studio.
- Krislock et al., (2017) Krislock, N., Malick, J., and Roupin, F. (2017). Biqcrunch: A semidefinite branch-and-bound method for solving binary quadratic problems. ACM Trans. Math. Softw., 43(4).
- Lasserre, (2001) Lasserre, J. B. (2001). Global optimization with polynomials and the problem of moments. SIAM Journal on optimization, 11(3):796–817.
- Lodi and Zarpellon, (2017) Lodi, A. and Zarpellon, G. (2017). On learning and branching: a survey. Top, 25(2):207–236.
- Meinshausen, (2006) Meinshausen, N. (2006). Quantile regression forests. Journal of Machine Learning Research, 7:983–999.
- MOSEK ApS, (2022) MOSEK ApS (2022). Introducing the MOSEK Optimization Suite 9.3.20.
- Parrilo, (2003) Parrilo, P. A. (2003). Semidefinite programming relaxations for semialgebraic problems. Mathematical programming, 96(2):293–320.
- Piccialli et al., (2022) Piccialli, V., Sudoso, A. M., and Wiegele, A. (2022). Sos-sdp: an exact solver for minimum sum-of-squares clustering. INFORMS Journal on Computing.
- R Core Team, (2021) R Core Team (2021). R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria.
- Rendl et al., (2010) Rendl, F., Rinaldi, G., and Wiegele, A. (2010). Solving max-cut to optimality by intersecting semidefinite and polyhedral relaxations. Mathematical Programming, 121(2):307–335.
- Sherali et al., (2012) Sherali, H. D., Dalkiran, E., and Desai, J. (2012). Enhancing RLT-based relaxations for polynomial programming problems via a new class of -semidefinite cuts. Computational Optimization and Applications, 52(2):483–506.
- Sherali and Tuncbilek, (1992) Sherali, H. D. and Tuncbilek, C. H. (1992). A global optimization algorithm for polynomial programming problems using a reformulation-linearization technique. Journal of Global Optimization, 2(1):101–112.
- Shor, (1987) Shor, N. Z. (1987). An approach to obtaining global extremums in polynomial mathematical programming problems. Cybernetics, 23(5):695–700.
- Teles et al., (2012) Teles, J. P., Castro, P. M., and Matos, H. A. (2012). Global optimization of water networks design using multiparametric disaggregation. Computers & Chemical Engineering, 40:132–147.
- Wright and Ziegler, (2017) Wright, M. N. and Ziegler, A. (2017). ranger: A fast implementation of random forests for high dimensional data in C++ and R. Journal of Statistical Software, 77(1):1–17.