Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100060
Derek H. Smith , Roberto Montemanni
{"title":"The missing Moore graph as an optimization problem","authors":"Derek H. Smith , Roberto Montemanni","doi":"10.1016/j.ejco.2023.100060","DOIUrl":"https://doi.org/10.1016/j.ejco.2023.100060","url":null,"abstract":"","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":""},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49742908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2022.100056
Vanessa Lange , Hans Daduna
We investigate a location-allocation-routing problem where trucks deliver goods from a central production facility to a set of warehouses with fixed locations and known demands. Due to limited capacities congestion occurs and results in queueing problems. The location of the center is determined to maximize the utilization of the given resources (measured in throughput) and the minimal number of trucks is determined to satisfy the overall demand generated by the warehouses. Main results for this integrated decision problem on strategic and tactical/operational level are: (i) The location decision is reduced to a standard Weber problem with weighted distances. (ii) The joint decision for location and fleet size is separable. (iii) The location of the center is robust against perturbations of several system parameters on the operational/tactical level. Additionally, we consider minimization of travel times as optimization target. By numerical examples we demonstrate the consequences of neglecting available information on long-term (rough) demand structure.
{"title":"The Weber problem in logistic and services networks under congestion","authors":"Vanessa Lange , Hans Daduna","doi":"10.1016/j.ejco.2022.100056","DOIUrl":"10.1016/j.ejco.2022.100056","url":null,"abstract":"<div><p>We investigate a location-allocation-routing problem where trucks deliver goods from a central production facility to a set of warehouses with fixed locations and known demands. Due to limited capacities congestion occurs and results in queueing problems. The location of the center is determined to maximize the utilization of the given resources (measured in throughput) and the minimal number of trucks is determined to satisfy the overall demand generated by the warehouses. Main results for this integrated decision problem on strategic and tactical/operational level are: (i) The location decision is reduced to a standard Weber problem with weighted distances. (ii) The joint decision for location and fleet size is separable. (iii) The location of the center is robust against perturbations of several system parameters on the operational/tactical level. Additionally, we consider minimization of travel times as optimization target. By numerical examples we demonstrate the consequences of neglecting available information on long-term (rough) demand structure.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100056"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46090888","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100073
Ahmadreza Marandi , Virginie Lurkin
Price differentiation is a common strategy in many markets. In this paper, we study a static multiproduct price optimization problem with demand given by a discrete mixed multinomial logit model. By considering a mixed logit model that includes customer specific variables and parameters in the utility specification, our pricing problem reflects well the discrete choice models used in practice. To solve this pricing problem, we design an efficient iterative optimization algorithm that asymptotically converges to the optimal solution. To this end, a linear optimization (LO) problem is formulated, based on the trust-region approach, to find a “good” feasible solution and approximate the problem from below. A convex optimization problem is designed using a convexification technique to approximate the optimization problem from above. Then, using a branching method, we tighten the optimality gap. The effectiveness of our algorithm is illustrated on several cases, and compared against solvers and existing state-of-the-art methods in the literature.
{"title":"An exact algorithm for the static pricing problem under discrete mixed logit demand","authors":"Ahmadreza Marandi , Virginie Lurkin","doi":"10.1016/j.ejco.2023.100073","DOIUrl":"10.1016/j.ejco.2023.100073","url":null,"abstract":"<div><p>Price differentiation is a common strategy in many markets. In this paper, we study a static multiproduct price optimization problem with demand given by a discrete mixed multinomial logit model. By considering a mixed logit model that includes customer specific variables and parameters in the utility specification, our pricing problem reflects well the discrete choice models used in practice. To solve this pricing problem, we design an efficient iterative optimization algorithm that asymptotically converges to the optimal solution. To this end, a <em>linear optimization</em> (LO) problem is formulated, based on the trust-region approach, to find a “good” feasible solution and approximate the problem from below. A convex optimization problem is designed using a convexification technique to approximate the optimization problem from above. Then, using a branching method, we tighten the optimality gap. The effectiveness of our algorithm is illustrated on several cases, and compared against solvers and existing state-of-the-art methods in the literature.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100073"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43244012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100064
Coralia Cartis , Panayotis Mertikopoulos
{"title":"Optimization Challenges in Data Science – Special Issue Editorial","authors":"Coralia Cartis , Panayotis Mertikopoulos","doi":"10.1016/j.ejco.2023.100064","DOIUrl":"10.1016/j.ejco.2023.100064","url":null,"abstract":"","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100064"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46448319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100079
Sara Venturini , Andrea Cristofari , Francesco Rinaldi , Francesco Tudisco
Graph Semi-Supervised learning is an important data analysis tool, where given a graph and a set of labeled nodes, the aim is to infer the labels to the remaining unlabeled nodes. In this paper, we start by considering an optimization-based formulation of the problem for an undirected graph, and then we extend this formulation to multilayer hypergraphs. We solve the problem using different coordinate descent approaches and compare the results with the ones obtained by the classic gradient descent method. Experiments on synthetic and real-world datasets show the potential of using coordinate descent methods with suitable selection rules.
{"title":"Laplacian-based semi-Supervised learning in multilayer hypergraphs by coordinate descent","authors":"Sara Venturini , Andrea Cristofari , Francesco Rinaldi , Francesco Tudisco","doi":"10.1016/j.ejco.2023.100079","DOIUrl":"https://doi.org/10.1016/j.ejco.2023.100079","url":null,"abstract":"<div><p>Graph Semi-Supervised learning is an important data analysis tool, where given a graph and a set of labeled nodes, the aim is to infer the labels to the remaining unlabeled nodes. In this paper, we start by considering an optimization-based formulation of the problem for an undirected graph, and then we extend this formulation to multilayer hypergraphs. We solve the problem using different coordinate descent approaches and compare the results with the ones obtained by the classic gradient descent method. Experiments on synthetic and real-world datasets show the potential of using coordinate descent methods with suitable selection rules.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100079"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440623000230/pdfft?md5=6f643343ec1f2c6f80d72a373b75a084&pid=1-s2.0-S2192440623000230-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91964081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A multilevel programming problem is an optimization problem that involves multiple decision makers, whose decisions are made in a sequential (or hierarchical) order. If all objective functions and constraints are linear and some decision variables in any level are restricted to take on integral or discrete values, then the problem is called a multilevel mixed integer linear programming problem (ML-MILP). Such problems are known to have disconnected feasible regions (called inducible regions), making the task of constructing an optimal solution challenging. Therefore, existing solution approaches are limited to some strict assumptions in the model formulations and lack universality. This paper presents a branch-and-cut (B&C) algorithm for the global solution of such problems with any finite number of hierarchical levels, and containing both continuous and discrete variables at each level of the decision-making hierarchy. Finite convergence of the proposed algorithm to a global solution is established. Numerical examples are used to illustrate the detailed procedure and to demonstrate the performance of the algorithm. Additionally, the computational performance of the proposed method is studied by comparing it with existing method through some selected numerical examples.
{"title":"Branch-and-cut solution approach for multilevel mixed integer linear programming problems","authors":"Ashenafi Awraris , Berhanu Guta Wordofa , Semu Mitiku Kassa","doi":"10.1016/j.ejco.2023.100076","DOIUrl":"https://doi.org/10.1016/j.ejco.2023.100076","url":null,"abstract":"<div><p>A multilevel programming problem is an optimization problem that involves multiple decision makers, whose decisions are made in a sequential (or hierarchical) order. If all objective functions and constraints are linear and some decision variables in any level are restricted to take on integral or discrete values, then the problem is called a multilevel mixed integer linear programming problem (ML-MILP). Such problems are known to have disconnected feasible regions (called inducible regions), making the task of constructing an optimal solution challenging. Therefore, existing solution approaches are limited to some strict assumptions in the model formulations and lack universality. This paper presents a branch-and-cut (B&C) algorithm for the global solution of such problems with any finite number of hierarchical levels, and containing both continuous and discrete variables at each level of the decision-making hierarchy. Finite convergence of the proposed algorithm to a global solution is established. Numerical examples are used to illustrate the detailed procedure and to demonstrate the performance of the algorithm. Additionally, the computational performance of the proposed method is studied by comparing it with existing method through some selected numerical examples.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100076"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49742854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100080
Rafael Kendy Arakaki, Fábio Luiz Usberti
Consider an undirected graph with demands scattered over the edges and a homogeneous fleet of vehicles to service the demands. In the open capacitated arc routing problem (OCARP) the objective is to find a set of routes that collectively service all demands with the minimum cost. Each vehicle has limited capacity and it can start and finish the route at any node. The OCARP is NP-hard, and its applications include meter reading and cutting path determination problems. State-of-the-art solution methods developed for the OCARP are heuristics, which show good tradeoffs between solution quality and processing time, but do not provide optimality certificates of the obtained solutions. This work focuses on a lower bounding method for the OCARP which can be used to better assess the quality of heuristic solutions. We propose the Relaxed Flow method () which involves the resolution of a mixed integer linear formulation where all vehicles' capacities are modeled as flows on an augmented graph. A parameter k controls the model tightness and is shown to be at least as tight as the well-known Belenguer and Benavent's formulation for any . To strengthen the model, capacity cuts were included in by means of a branch-and-cut framework. Extensive computational experiments conducted on a set of benchmark instances revealed that our method outperformed previous methods. Computational experiments also demonstrated the importance of the parameterization technique to obtain good results. The previously known lower bounds were improved substantially and optimality certificates were attained in 78.9% of the instances. As far as we know this is the first parameterized lower bounding method proposed for an arc routing problem, and we argue it can be generalized to other variants of arc routing problems and general routing problems.
{"title":"A parameterized lower bounding method for the open capacitated arc routing problem","authors":"Rafael Kendy Arakaki, Fábio Luiz Usberti","doi":"10.1016/j.ejco.2023.100080","DOIUrl":"https://doi.org/10.1016/j.ejco.2023.100080","url":null,"abstract":"<div><p>Consider an undirected graph with demands scattered over the edges and a homogeneous fleet of vehicles to service the demands. In the <em>open capacitated arc routing problem</em> (OCARP) the objective is to find a set of routes that collectively service all demands with the minimum cost. Each vehicle has limited capacity and it can start and finish the route at any node. The OCARP is NP-hard, and its applications include meter reading and cutting path determination problems. State-of-the-art solution methods developed for the OCARP are heuristics, which show good tradeoffs between solution quality and processing time, but do not provide optimality certificates of the obtained solutions. This work focuses on a lower bounding method for the OCARP which can be used to better assess the quality of heuristic solutions. We propose the Relaxed Flow method (<span><math><mi>R</mi><mi>F</mi><mo>(</mo><mi>k</mi><mo>)</mo></math></span>) which involves the resolution of a mixed integer linear formulation where all vehicles' capacities are modeled as flows on an augmented graph. A parameter <em>k</em> controls the model tightness and <span><math><mi>R</mi><mi>F</mi><mo>(</mo><mi>k</mi><mo>)</mo></math></span> is shown to be at least as tight as the well-known Belenguer and Benavent's formulation for any <span><math><mi>k</mi><mo>⩾</mo><mn>0</mn></math></span>. To strengthen the model, capacity cuts were included in <span><math><mi>R</mi><mi>F</mi><mo>(</mo><mi>k</mi><mo>)</mo></math></span> by means of a branch-and-cut framework. Extensive computational experiments conducted on a set of benchmark instances revealed that our method outperformed previous methods. Computational experiments also demonstrated the importance of the parameterization technique to obtain good results. The previously known lower bounds were improved substantially and optimality certificates were attained in 78.9% of the instances. As far as we know this is the first parameterized lower bounding method proposed for an arc routing problem, and we argue it can be generalized to other variants of arc routing problems and general routing problems.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100080"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2192440623000242/pdfft?md5=47d32c2d05c48fe7d4f22b5727ef4bc9&pid=1-s2.0-S2192440623000242-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"92025778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100070
Annabella Astorino , Matteo Avolio , Antonio Fuduli
Multiple Instance Learning (MIL) is a kind of weak supervised learning, where each sample is represented by a bag of instances. The main characteristic of such problems resides in the training phase, since the class labels are provided only for each bag, whereas the instance labels are unknown.
We focus on binary MIL problems characterized by two types of instances (positive and negative): based on the standard MIL assumption, a bag is considered positive if at least one of its instances is positive and it is considered negative otherwise. Then our idea is to generate a maximum-margin polyhedral separation surface such that, for each positive bag, at least one of its instances is inside the polyhedron and all the instances of the negative bags are outside. The resulting optimization problem is a nonlinear, nonconvex and nonsmooth mixed integer program, that we heuristically solve by a Block Coordinate Descent type method, based on repeatedly applying the DC (Difference of Convex) Algorithm.
Numerical results are presented on a set of benchmark datasets.
{"title":"Maximum-margin polyhedral separation for binary Multiple Instance Learning","authors":"Annabella Astorino , Matteo Avolio , Antonio Fuduli","doi":"10.1016/j.ejco.2023.100070","DOIUrl":"10.1016/j.ejco.2023.100070","url":null,"abstract":"<div><p>Multiple Instance Learning (MIL) is a kind of weak supervised learning, where each sample is represented by a bag of instances. The main characteristic of such problems resides in the training phase, since the class labels are provided only for each bag, whereas the instance labels are unknown.</p><p>We focus on binary MIL problems characterized by two types of instances (positive and negative): based on the standard MIL assumption, a bag is considered positive if at least one of its instances is positive and it is considered negative otherwise. Then our idea is to generate a maximum-margin polyhedral separation surface such that, for each positive bag, at least one of its instances is inside the polyhedron and all the instances of the negative bags are outside. The resulting optimization problem is a nonlinear, nonconvex and nonsmooth mixed integer program, that we heuristically solve by a Block Coordinate Descent type method, based on repeatedly applying the DC (Difference of Convex) Algorithm.</p><p>Numerical results are presented on a set of benchmark datasets.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100070"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42034951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100058
Vassilios Yfantis , Simon Wenzel , Achim Wagner , Martin Ruskowski , Sebastian Engell
In this paper, two new algorithms for dual decomposition-based distributed optimization are presented. Both algorithms rely on the quadratic approximation of the dual function of the primal optimization problem. The dual variables are updated in each iteration through a maximization of the approximated dual function. The first algorithm approximates the dual function by solving a regression problem, based on the values of the dual function collected from previous iterations. The second algorithm updates the parameters of the quadratic approximation via a quasi-Newton scheme. Both algorithms employ step size constraints for the update of the dual variables. Furthermore, the subgradients from previous iterations are stored in order to construct cutting planes, similar to bundle methods for nonsmooth optimization. However, instead of using the cutting planes to formulate a piece-wise linear over-approximation of the dual function, they are used to construct valid inequalities for the update step. In order to demonstrate the efficiency of the algorithms, they are evaluated on a large set of constrained quadratic, convex and mixed-integer benchmark problems and compared to the subgradient method, the bundle trust method, the alternating direction method of multipliers and the quadratic approximation coordination algorithm. The results show that the proposed algorithms perform better than the compared algorithms both in terms of the required number of iterations and in the number of solved benchmark problems in most cases.
{"title":"Hierarchical distributed optimization of constraint-coupled convex and mixed-integer programs using approximations of the dual function","authors":"Vassilios Yfantis , Simon Wenzel , Achim Wagner , Martin Ruskowski , Sebastian Engell","doi":"10.1016/j.ejco.2023.100058","DOIUrl":"10.1016/j.ejco.2023.100058","url":null,"abstract":"<div><p>In this paper, two new algorithms for dual decomposition-based distributed optimization are presented. Both algorithms rely on the quadratic approximation of the dual function of the primal optimization problem. The dual variables are updated in each iteration through a maximization of the approximated dual function. The first algorithm approximates the dual function by solving a regression problem, based on the values of the dual function collected from previous iterations. The second algorithm updates the parameters of the quadratic approximation via a quasi-Newton scheme. Both algorithms employ step size constraints for the update of the dual variables. Furthermore, the subgradients from previous iterations are stored in order to construct cutting planes, similar to bundle methods for nonsmooth optimization. However, instead of using the cutting planes to formulate a piece-wise linear over-approximation of the dual function, they are used to construct valid inequalities for the update step. In order to demonstrate the efficiency of the algorithms, they are evaluated on a large set of constrained quadratic, convex and mixed-integer benchmark problems and compared to the subgradient method, the bundle trust method, the alternating direction method of multipliers and the quadratic approximation coordination algorithm. The results show that the proposed algorithms perform better than the compared algorithms both in terms of the required number of iterations and in the number of solved benchmark problems in most cases.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100058"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44992580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-01DOI: 10.1016/j.ejco.2023.100069
Sahand Asgharieh Ahari , Burak Kocuk
Logistic regression is one of the widely-used classification tools to construct prediction models. For datasets with a large number of features, feature subset selection methods are considered to obtain accurate and interpretable prediction models, in which irrelevant and redundant features are removed. In this paper, we address the problem of feature subset selection in logistic regression using modern optimization techniques. To this end, we formulate this problem as a mixed-integer exponential cone program (MIEXP). To the best of our knowledge, this is the first time both nonlinear and discrete aspects of the underlying problem are fully considered within an exact optimization framework. We derive different versions of the MIEXP model by the means of regularization and goodness of fit measures including Akaike Information Criterion and Bayesian Information Criterion. Finally, we solve our MIEXP models using the solver MOSEK and evaluate the performance of our different versions over a set of toy examples and benchmark datasets. The results show that our approach is quite successful in obtaining accurate and interpretable prediction models compared to other methods from the literature.
{"title":"A mixed-integer exponential cone programming formulation for feature subset selection in logistic regression","authors":"Sahand Asgharieh Ahari , Burak Kocuk","doi":"10.1016/j.ejco.2023.100069","DOIUrl":"10.1016/j.ejco.2023.100069","url":null,"abstract":"<div><p>Logistic regression is one of the widely-used classification tools to construct prediction models. For datasets with a large number of features, feature subset selection methods are considered to obtain accurate and interpretable prediction models, in which irrelevant and redundant features are removed. In this paper, we address the problem of <em>feature subset selection in logistic regression</em> using modern optimization techniques. To this end, we formulate this problem as a mixed-integer exponential cone program (MIEXP). To the best of our knowledge, this is the first time both nonlinear and discrete aspects of the underlying problem are fully considered within an exact optimization framework. We derive different versions of the MIEXP model by the means of regularization and goodness of fit measures including Akaike Information Criterion and Bayesian Information Criterion. Finally, we solve our MIEXP models using the solver <em>MOSEK</em> and evaluate the performance of our different versions over a set of toy examples and benchmark datasets. The results show that our approach is quite successful in obtaining accurate and interpretable prediction models compared to other methods from the literature.</p></div>","PeriodicalId":51880,"journal":{"name":"EURO Journal on Computational Optimization","volume":"11 ","pages":"Article 100069"},"PeriodicalIF":2.4,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44435816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}