Pub Date : 2024-09-23DOI: 10.1007/s10479-024-06157-4
Somayeh Sadeghi, Abbas Seifi
We consider a shortest-path network interdiction problem under endogenous uncertainty on successful detection. Endogenous uncertainty arises from the fact that the interdictor may decide to enforce surveillance on some critical arcs, which would affect the prior probability of success on those arcs. The evader decision is formulated as a two-stage stochastic programming problem. In a “here and now situation”, he has to choose the shortest path in the network before realizing detection scenarios. Then, in the second stage, the evader tries to minimize the expected cost of being detected over all possible scenarios. We consider binary scenarios to represent whether or not the evader is detected on each path and apply a linearization method to deal with the non-linearity in the decision-dependent probability measure. A decomposition method is used to solve the proposed model for a case study of a worldwide drug trafficking network. The case study is concerned with finding the most critical arcs for interdicting drug trafficking. Numerical results show that a tiny increase in the probability of opium seizures leads to a significant change in the expected cost when the critical arcs are interdicted. Due to the exponential number of scenarios, the model could not be solved in a reasonable time. Common scenario reduction methods are designed for exogenous uncertainty. We apply an improved bundling method to reduce the number of scenarios in case of endogenous uncertainty. Computational results show that our method reduces the model size and solution time tremendously without significantly affecting the objective value.
{"title":"A modified scenario bundling method for shortest path network interdiction under endogenous uncertainty","authors":"Somayeh Sadeghi, Abbas Seifi","doi":"10.1007/s10479-024-06157-4","DOIUrl":"10.1007/s10479-024-06157-4","url":null,"abstract":"<div><p>We consider a shortest-path network interdiction problem under endogenous uncertainty on successful detection. Endogenous uncertainty arises from the fact that the interdictor may decide to enforce surveillance on some critical arcs, which would affect the prior probability of success on those arcs. The evader decision is formulated as a two-stage stochastic programming problem. In a “here and now situation”, he has to choose the shortest path in the network before realizing detection scenarios. Then, in the second stage, the evader tries to minimize the expected cost of being detected over all possible scenarios. We consider binary scenarios to represent whether or not the evader is detected on each path and apply a linearization method to deal with the non-linearity in the decision-dependent probability measure. A decomposition method is used to solve the proposed model for a case study of a worldwide drug trafficking network. The case study is concerned with finding the most critical arcs for interdicting drug trafficking. Numerical results show that a tiny increase in the probability of opium seizures leads to a significant change in the expected cost when the critical arcs are interdicted. Due to the exponential number of scenarios, the model could not be solved in a reasonable time. Common scenario reduction methods are designed for exogenous uncertainty. We apply an improved bundling method to reduce the number of scenarios in case of endogenous uncertainty. Computational results show that our method reduces the model size and solution time tremendously without significantly affecting the objective value.</p></div>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"343 1","pages":"429 - 457"},"PeriodicalIF":4.4,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142789233","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-23DOI: 10.1007/s10479-024-06283-z
Xue Yan, Ting Wang, Xuefei Shi
Outsourcing operations have become an essential factor in enhancing the competitive advantage of software development enterprises. In this work, we examine the application of combinatorial auction in technician assignment and outsourcing service procurement, which is conducted by software enterprises to minimize the total cost of developing all the software. It gives rise to an unrelated parallel machine scheduling problem incorporating combinatorial auction (UPMSCA). Here, the jobs represent the software to be developed, and they consume the perishable time resources of the development technicians, which can be translated into monetary costs. The objective is to schedule the jobs on parallel machines or select the bid with the lowest cost. To solve the problem, we propose an arc-flow model and a set-partitioning formulation with column-based constraints. A branch-and-price algorithm with four branching rules is proposed and utilizes an effective dynamic programming algorithm to solve the pricing subproblem in the pattern-based formulation. To speed up computation, a bidirectional search method and a dominance rule are applied. Results from extensive computational tests on 100 sets of randomly generated instances demonstrate the performance of our algorithm.
{"title":"Optimal scheduling on unrelated parallel machines with combinatorial auction","authors":"Xue Yan, Ting Wang, Xuefei Shi","doi":"10.1007/s10479-024-06283-z","DOIUrl":"10.1007/s10479-024-06283-z","url":null,"abstract":"<div><p>Outsourcing operations have become an essential factor in enhancing the competitive advantage of software development enterprises. In this work, we examine the application of combinatorial auction in technician assignment and outsourcing service procurement, which is conducted by software enterprises to minimize the total cost of developing all the software. It gives rise to an unrelated parallel machine scheduling problem incorporating combinatorial auction (<span>UPMSCA</span>). Here, the jobs represent the software to be developed, and they consume the perishable time resources of the development technicians, which can be translated into monetary costs. The objective is to schedule the jobs on parallel machines or select the bid with the lowest cost. To solve the problem, we propose an arc-flow model and a set-partitioning formulation with column-based constraints. A branch-and-price algorithm with four branching rules is proposed and utilizes an effective dynamic programming algorithm to solve the pricing subproblem in the pattern-based formulation. To speed up computation, a bidirectional search method and a dominance rule are applied. Results from extensive computational tests on 100 sets of randomly generated instances demonstrate the performance of our algorithm.</p></div>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"344 2-3","pages":"937 - 963"},"PeriodicalIF":4.4,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-23DOI: 10.1007/s10479-024-06273-1
Jose A. Rodriguez-Serrano
The systematic prediction of real estate prices is a foundational block in the operations of many firms and has individual, societal and policy implications. In the past, a vast amount of works have used common statistical models such as ordinary least squares or machine learning approaches. While these approaches yield good predictive accuracy, most models work very differently from the human intuition in understanding real estate prices. Usually, humans apply a criterion known as “direct comparison”, whereby the property to be valued is explicitly compared with similar properties. This trait is frequently ignored when applying machine learning to real estate valuation. In this article, we propose a model based on a methodology called prototype-based learning, that to our knowledge has never been applied to real estate valuation. The model has four crucial characteristics: (a) it is able to capture non-linear relations between price and the input variables, (b) it is a parametric model able to optimize any loss function of interest, (c) it has some degree of explainability, and, more importantly, (d) it encodes the notion of direct comparison. None of the past approaches for real estate prediction comply with these four characteristics simultaneously. The experimental validation indicates that, in terms of predictive accuracy, the proposed model is better or on par to other machine learning based approaches. An interesting advantage of this method is the ability to summarize a dataset of real estate prices into a few “prototypes”, a set of the most representative properties.
{"title":"Prototype-based learning for real estate valuation: a machine learning model that explains prices","authors":"Jose A. Rodriguez-Serrano","doi":"10.1007/s10479-024-06273-1","DOIUrl":"10.1007/s10479-024-06273-1","url":null,"abstract":"<div><p>The systematic prediction of real estate prices is a foundational block in the operations of many firms and has individual, societal and policy implications. In the past, a vast amount of works have used common statistical models such as ordinary least squares or machine learning approaches. While these approaches yield good predictive accuracy, most models work very differently from the human intuition in understanding real estate prices. Usually, humans apply a criterion known as “direct comparison”, whereby the property to be valued is explicitly compared with similar properties. This trait is frequently ignored when applying machine learning to real estate valuation. In this article, we propose a model based on a methodology called <i>prototype-based learning</i>, that to our knowledge has never been applied to real estate valuation. The model has four crucial characteristics: (a) it is able to capture non-linear relations between price and the input variables, (b) it is a parametric model able to optimize any loss function of interest, (c) it has some degree of explainability, and, more importantly, (d) it encodes the notion of direct comparison. None of the past approaches for real estate prediction comply with these four characteristics simultaneously. The experimental validation indicates that, in terms of predictive accuracy, the proposed model is better or on par to other machine learning based approaches. An interesting advantage of this method is the ability to summarize a dataset of real estate prices into a few “prototypes”, a set of the most representative properties.</p></div>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"344 1","pages":"287 - 311"},"PeriodicalIF":4.4,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10479-024-06273-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142912891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-22DOI: 10.1007/s10479-024-06285-x
I. Gusti Agung Premananda, Aris Tjahyanto, Ahmad Mukhlason
Sports timetabling is a complex and challenging problem. The latest open benchmark dataset for the sport timetabling problem is from the International Timetabling Competition (ITC) 2021. Due to its complexity, only a few approaches have successfully generated feasible solutions for the problems in this dataset, as reported in scientific literature. To the best of our knowledge, there is only one study in the literature that has successfully generated feasible solutions for all 45 problems in the dataset. In this paper, we propose our novel efficient algorithm based on the Iterated Local Search algorithm to solve the ITC 2021 benchmark dataset. Unlike prior successful approaches that combined metaheuristics with an exact approach, our proposed approach is solely metaheuristic. Our contribution includes the design of strategies for both perturbation and local search phases, coupled with the integration of shuffling strategies. The experimental results show that our proposed algorithm is remarkably successful in generating feasible solutions for all 45 problems present in the ITC 2021 dataset.
{"title":"Efficient iterated local search based metaheuristic approach for solving sports timetabling problems of International Timetabling Competition 2021","authors":"I. Gusti Agung Premananda, Aris Tjahyanto, Ahmad Mukhlason","doi":"10.1007/s10479-024-06285-x","DOIUrl":"10.1007/s10479-024-06285-x","url":null,"abstract":"<div><p>Sports timetabling is a complex and challenging problem. The latest open benchmark dataset for the sport timetabling problem is from the International Timetabling Competition (ITC) 2021. Due to its complexity, only a few approaches have successfully generated feasible solutions for the problems in this dataset, as reported in scientific literature. To the best of our knowledge, there is only one study in the literature that has successfully generated feasible solutions for all 45 problems in the dataset. In this paper, we propose our novel efficient algorithm based on the Iterated Local Search algorithm to solve the ITC 2021 benchmark dataset. Unlike prior successful approaches that combined metaheuristics with an exact approach, our proposed approach is solely metaheuristic. Our contribution includes the design of strategies for both perturbation and local search phases, coupled with the integration of shuffling strategies. The experimental results show that our proposed algorithm is remarkably successful in generating feasible solutions for all 45 problems present in the ITC 2021 dataset.</p></div>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"343 1","pages":"411 - 427"},"PeriodicalIF":4.4,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142789226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-21DOI: 10.1007/s10479-024-06275-z
Kuo-Ching Ying, Pourya Pourhejazy, Chuan-En Sung
Volatility in the supply chain of critical products, notably the vaccine shortage during the pandemic, influences livelihoods and may lead to significant delays and long waiting times. Considering the capital- and time-intensive nature of capacity expansion plans, strategic operational production decisions are required best to address the supply-demand mismatches given the limited production resources. This research article investigates the production scenarios where the demand of one agent must be completed within a specified due date, hereinafter referred to as the deadline, while minimizing the maximum or total completion time of another agent's demand. For this purpose, the Two-Agent Proportionate Flowshop Scheduling Problem with deadlines is introduced. Two polynomial-time optimization algorithms are developed to solve these optimization problems. This study will serve as a basis for further developing this practical yet understudied scheduling problem.
{"title":"Two-agent proportionate flowshop scheduling with deadlines: polynomial-time optimization algorithms","authors":"Kuo-Ching Ying, Pourya Pourhejazy, Chuan-En Sung","doi":"10.1007/s10479-024-06275-z","DOIUrl":"10.1007/s10479-024-06275-z","url":null,"abstract":"<div><p>Volatility in the supply chain of critical products, notably the vaccine shortage during the pandemic, influences livelihoods and may lead to significant delays and long waiting times. Considering the capital- and time-intensive nature of capacity expansion plans, strategic operational production decisions are required best to address the supply-demand mismatches given the limited production resources. This research article investigates the production scenarios where the demand of one agent must be completed within a specified due date, hereinafter referred to as the <i>deadline</i>, while minimizing the maximum or total completion time of another agent's demand. For this purpose, the Two-Agent Proportionate Flowshop Scheduling Problem with deadlines is introduced. Two polynomial-time optimization algorithms are developed to solve these optimization problems. This study will serve as a basis for further developing this practical yet understudied scheduling problem.</p></div>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"343 1","pages":"543 - 558"},"PeriodicalIF":4.4,"publicationDate":"2024-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10479-024-06275-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142789186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-21DOI: 10.1007/s10479-024-06245-5
Bruce Golden, Linus Schrage, Douglas Shier, Lida Anna Apergi
Linear programming has had a tremendous impact in the modeling and solution of a great diversity of applied problems, especially in the efficient allocation of resources. As a result, this methodology forms the backbone of introductory courses in operations research. What students, and others, may not appreciate is that linear programming transcends its linear nomenclature and can be applied to an even wider range of important practical problems. The objective of this article is to present a selection, and just a selection, from this range of problems that at first blush do not seem amenable to linear programming formulation. The exposition focuses on the most basic models in these selected applications, with pointers to more elaborate formulations and extensions. Thus, our intent is to expand the modeling awareness of those first encountering linear programming. In addition, we hope this article will be of interest to those who teach linear programming and to seasoned academics and practitioners, alike.
{"title":"The unexpected power of linear programming: an updated collection of surprising applications","authors":"Bruce Golden, Linus Schrage, Douglas Shier, Lida Anna Apergi","doi":"10.1007/s10479-024-06245-5","DOIUrl":"10.1007/s10479-024-06245-5","url":null,"abstract":"<div><p>Linear programming has had a tremendous impact in the modeling and solution of a great diversity of applied problems, especially in the efficient allocation of resources. As a result, this methodology forms the backbone of introductory courses in operations research. What students, and others, may not appreciate is that linear programming transcends its linear nomenclature and can be applied to an even wider range of important practical problems. The objective of this article is to present a selection, and just a selection, from this range of problems that at first blush do not seem amenable to linear programming formulation. The exposition focuses on the most basic models in these selected applications, with pointers to more elaborate formulations and extensions. Thus, our intent is to expand the modeling awareness of those first encountering linear programming. In addition, we hope this article will be of interest to those who teach linear programming and to seasoned academics and practitioners, alike.</p></div>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"343 2021-2023)","pages":"573 - 605"},"PeriodicalIF":4.4,"publicationDate":"2024-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10479-024-06245-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142826131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.1007/s10479-024-06293-x
Jang Ho Kim, Seyoung Kim, Yongjae Lee, Woo Chang Kim, Frank J. Fabozzi
Mean–variance optimization, introduced by Markowitz, is a foundational theory and methodology in finance and optimization, significantly influencing investment management practices. This study enhances mean–variance optimization by integrating machine learning-based anomaly detection, specifically using GANs (generative adversarial networks), to identify anomaly levels in the stock market. We demonstrate the utility of GANs in detecting market anomalies and incorporating this information into portfolio optimization using robust methods such as shrinkage estimators and the Gerber statistic. Empirical analysis confirms that portfolios optimized with anomaly scores outperform those using conventional portfolio optimization. This study highlights the potential of advanced data-driven techniques to improve risk management and portfolio performance.
{"title":"Enhancing mean–variance portfolio optimization through GANs-based anomaly detection","authors":"Jang Ho Kim, Seyoung Kim, Yongjae Lee, Woo Chang Kim, Frank J. Fabozzi","doi":"10.1007/s10479-024-06293-x","DOIUrl":"10.1007/s10479-024-06293-x","url":null,"abstract":"<div><p>Mean–variance optimization, introduced by Markowitz, is a foundational theory and methodology in finance and optimization, significantly influencing investment management practices. This study enhances mean–variance optimization by integrating machine learning-based anomaly detection, specifically using GANs (generative adversarial networks), to identify anomaly levels in the stock market. We demonstrate the utility of GANs in detecting market anomalies and incorporating this information into portfolio optimization using robust methods such as shrinkage estimators and the Gerber statistic. Empirical analysis confirms that portfolios optimized with anomaly scores outperform those using conventional portfolio optimization. This study highlights the potential of advanced data-driven techniques to improve risk management and portfolio performance.</p></div>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"346 1","pages":"217 - 244"},"PeriodicalIF":4.4,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143638643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s10479-024-06252-6
Abdelali Gabih, Hakam Kondakji, Ralf Wunderlich
{"title":"Correction: Power utility maximization with expert opinions at fixed arrival times in a market with hidden gaussian drift","authors":"Abdelali Gabih, Hakam Kondakji, Ralf Wunderlich","doi":"10.1007/s10479-024-06252-6","DOIUrl":"10.1007/s10479-024-06252-6","url":null,"abstract":"","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"341 2-3","pages":"1363 - 1363"},"PeriodicalIF":4.4,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10479-024-06252-6.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142438864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-19DOI: 10.1007/s10479-024-06226-8
Lasse Bohlen, Julian Rosenberger, Patrick Zschech, Mathias Kraus
In healthcare, especially within intensive care units (ICU), informed decision-making by medical professionals is crucial due to the complexity of medical data. Healthcare analytics seeks to support these decisions by generating accurate predictions through advanced machine learning (ML) models, such as boosted decision trees and random forests. While these models frequently exhibit accurate predictions across various medical tasks, they often lack interpretability. To address this challenge, researchers have developed interpretable ML models that balance accuracy and interpretability. In this study, we evaluate the performance gap between interpretable and black-box models in two healthcare prediction tasks, mortality and length-of-stay prediction in ICU settings. We focus specifically on the family of generalized additive models (GAMs) as powerful interpretable ML models. Our assessment uses the publicly available Medical Information Mart for Intensive Care dataset, and we analyze the models based on (i) predictive performance, (ii) the influence of compact feature sets (i.e., only few features) on predictive performance, and (iii) interpretability and consistency with medical knowledge. Our results show that interpretable models achieve competitive performance, with a minor decrease of 0.2–0.9 percentage points in area under the receiver operating characteristic relative to state-of-the-art black-box models, while preserving complete interpretability. This remains true even for parsimonious models that use only 2.2 % of patient features. Our study highlights the potential of interpretable models to improve decision-making in ICUs by providing medical professionals with easily understandable and verifiable predictions.
在医疗保健领域,尤其是在重症监护室(ICU)内,由于医疗数据的复杂性,医疗专业人员做出明智的决策至关重要。医疗分析旨在通过先进的机器学习(ML)模型(如增强决策树和随机森林)生成准确的预测,从而为这些决策提供支持。虽然这些模型经常能对各种医疗任务做出准确预测,但它们往往缺乏可解释性。为了应对这一挑战,研究人员开发了可解释的 ML 模型,在准确性和可解释性之间取得了平衡。在本研究中,我们评估了可解释模型和黑盒模型在两项医疗预测任务(重症监护室的死亡率和住院时间预测)中的性能差距。我们特别关注作为强大的可解释 ML 模型的广义加法模型(GAM)系列。我们的评估使用了公开的重症监护医疗信息集市数据集,并根据以下几个方面对模型进行了分析:(i) 预测性能;(ii) 紧凑型特征集(即只有少数特征)对预测性能的影响;(iii) 可解释性以及与医学知识的一致性。我们的研究结果表明,可解释模型在保持完全可解释性的同时,还能获得有竞争力的性能,与最先进的黑盒模型相比,接收器操作特征下面积略微下降了 0.2-0.9 个百分点。即使是仅使用 2.2% 患者特征的简约模型,情况也是如此。我们的研究强调了可解释模型的潜力,它能为医疗专业人员提供易于理解和验证的预测,从而改善重症监护室的决策。
{"title":"Leveraging interpretable machine learning in intensive care","authors":"Lasse Bohlen, Julian Rosenberger, Patrick Zschech, Mathias Kraus","doi":"10.1007/s10479-024-06226-8","DOIUrl":"https://doi.org/10.1007/s10479-024-06226-8","url":null,"abstract":"<p>In healthcare, especially within intensive care units (ICU), informed decision-making by medical professionals is crucial due to the complexity of medical data. Healthcare analytics seeks to support these decisions by generating accurate predictions through advanced machine learning (ML) models, such as boosted decision trees and random forests. While these models frequently exhibit accurate predictions across various medical tasks, they often lack interpretability. To address this challenge, researchers have developed interpretable ML models that balance accuracy and interpretability. In this study, we evaluate the performance gap between interpretable and black-box models in two healthcare prediction tasks, mortality and length-of-stay prediction in ICU settings. We focus specifically on the family of generalized additive models (GAMs) as powerful interpretable ML models. Our assessment uses the publicly available Medical Information Mart for Intensive Care dataset, and we analyze the models based on (i) predictive performance, (ii) the influence of compact feature sets (i.e., only few features) on predictive performance, and (iii) interpretability and consistency with medical knowledge. Our results show that interpretable models achieve competitive performance, with a minor decrease of 0.2–0.9 percentage points in area under the receiver operating characteristic relative to state-of-the-art black-box models, while preserving complete interpretability. This remains true even for parsimonious models that use only 2.2 % of patient features. Our study highlights the potential of interpretable models to improve decision-making in ICUs by providing medical professionals with easily understandable and verifiable predictions.</p>","PeriodicalId":8215,"journal":{"name":"Annals of Operations Research","volume":"19 1","pages":""},"PeriodicalIF":4.8,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142251124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}