Pub Date : 2024-08-30DOI: 10.1016/j.dss.2024.114313
Xiaochen Wang , Runtong Zhang , Xiaomin Zhu
Multimorbidity, the presence of two or more chronic conditions within an individual, represents one of the most intricate challenges for global health systems. Traditional single-disease management often fails to address the multifaceted nature of multimorbidity. Network model emerges as a growing field for elucidating the interconnections among multimorbidity. However, the field lacks a standardized method to compute and visually represent of these networks. Given the challenges, this study proposes a three-stage methodology to decipher multimorbidity. First, we integrate the Failure Modes and Effects Analysis (FMEA) method with the multimorbidity encapsulation framework to develop the Multimorbidity Risk Network (MRN). Second, we use complex network techniques to identify high-risk patterns within MRN communities. Finally, we apply machine learning techniques to correlate these communities with the biological attributes of patients that have been marginalized in most studies. Our approach advocates a paradigm shift from the conventional focus on single diseases to a holistic, patient-centric approach, providing decision-makers with integrated information technology artifacts for deciphering the multimorbidity.
{"title":"What can we learn from multimorbidity? A deep dive from its risk patterns to the corresponding patient profiles","authors":"Xiaochen Wang , Runtong Zhang , Xiaomin Zhu","doi":"10.1016/j.dss.2024.114313","DOIUrl":"10.1016/j.dss.2024.114313","url":null,"abstract":"<div><p>Multimorbidity, the presence of two or more chronic conditions within an individual, represents one of the most intricate challenges for global health systems. Traditional single-disease management often fails to address the multifaceted nature of multimorbidity. Network model emerges as a growing field for elucidating the interconnections among multimorbidity. However, the field lacks a standardized method to compute and visually represent of these networks. Given the challenges, this study proposes a three-stage methodology to decipher multimorbidity. First, we integrate the Failure Modes and Effects Analysis (FMEA) method with the multimorbidity encapsulation framework to develop the Multimorbidity Risk Network (MRN). Second, we use complex network techniques to identify high-risk patterns within MRN communities. Finally, we apply machine learning techniques to correlate these communities with the biological attributes of patients that have been marginalized in most studies. Our approach advocates a paradigm shift from the conventional focus on single diseases to a holistic, patient-centric approach, providing decision-makers with integrated information technology artifacts for deciphering the multimorbidity.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"186 ","pages":"Article 114313"},"PeriodicalIF":6.7,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142151015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-30DOI: 10.1016/j.dss.2024.114314
Junbo Zhang , Jiandong Lu , Xiaolei Wang , Luning Liu , Yuqiang Feng
In customer service, emotional expressions by chatbots are considered a promising direction to improve customer experience. However, there is a lack of comprehensive understanding of how and when chatbots' emotional expressions improve customer attitudes. Although chatbots' emotional expressions of care and concern may feel inauthentic to customers in the inferential path, which can negatively affects customer attitudes, we propose that the positive effect of the affective reactions path can result in a positive effect on customer attitude based on the dual-path view of Emotions as Social Information (EASI). The relative strengths of the two EASI paths can be moderated, and we explored the moderating effects of rational thinking styles (information processing in EASI) and beliefs in computer emotion (perceived appropriateness in EASI). According to EASI, situation can affect the meaning of emotions, so we conducted experiments in two situations. With chatbot identity disclosure, we found that the chatbot's emotional expressions reduce customers' perceived authenticity (reflecting the inferential path in EASI) but ultimately improve customer attitudes. Belief in computer emotions and rational thinking style moderated the negative relationship between emotional expressions and authenticity. With chatbot identity non-disclosure, the chatbot's emotional expressions still improve customer attitudes but with no effect on authenticity. Because there is high likelihood of chatbot identities being discovered by customers, this finding of the moderating effect of perceived humanness on authenticity is highly relevant. Our findings make important contributions to research on computer emotion and service authenticity.
{"title":"Emotional expressions of care and concern by customer service chatbots: Improved customer attitudes despite perceived inauthenticity","authors":"Junbo Zhang , Jiandong Lu , Xiaolei Wang , Luning Liu , Yuqiang Feng","doi":"10.1016/j.dss.2024.114314","DOIUrl":"10.1016/j.dss.2024.114314","url":null,"abstract":"<div><p>In customer service, emotional expressions by chatbots are considered a promising direction to improve customer experience. However, there is a lack of comprehensive understanding of how and when chatbots' emotional expressions improve customer attitudes. Although chatbots' emotional expressions of care and concern may feel inauthentic to customers in the inferential path, which can negatively affects customer attitudes, we propose that the positive effect of the affective reactions path can result in a positive effect on customer attitude based on the dual-path view of Emotions as Social Information (EASI). The relative strengths of the two EASI paths can be moderated, and we explored the moderating effects of rational thinking styles (information processing in EASI) and beliefs in computer emotion (perceived appropriateness in EASI). According to EASI, situation can affect the meaning of emotions, so we conducted experiments in two situations. With chatbot identity disclosure, we found that the chatbot's emotional expressions reduce customers' perceived authenticity (reflecting the inferential path in EASI) but ultimately improve customer attitudes. Belief in computer emotions and rational thinking style moderated the negative relationship between emotional expressions and authenticity. With chatbot identity non-disclosure, the chatbot's emotional expressions still improve customer attitudes but with no effect on authenticity. Because there is high likelihood of chatbot identities being discovered by customers, this finding of the moderating effect of perceived humanness on authenticity is highly relevant. Our findings make important contributions to research on computer emotion and service authenticity.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"186 ","pages":"Article 114314"},"PeriodicalIF":6.7,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142150984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-27DOI: 10.1016/j.dss.2024.114310
Jamie Zimmermann , Lance E. Champagne , John M. Dickens , Benjamin T. Hazen
As a part of natural language processing (NLP), the intent of topic modeling is to identify topics in textual corpora with limited human input. Current topic modeling techniques, like Latent Dirichlet Allocation (LDA), are limited in the pre-processing steps and currently require human judgement, increasing analysis time and opportunities for error. The purpose of this research is to allay some of those limitations by introducing new approaches to improve coherence without adding computational complexity and provide an objective method for determining the number of topics within a corpus. First, we identify a requirement for a more robust stop words list and introduce a new dimensionality-reduction heuristic that exploits the number of words within a document to infer importance to word choice. Second, we develop an eigenvalue technique to determine the number of topics within a corpus. Third, we combine all of these techniques into the Zimm Approach, which produces higher quality results than LDA in determining the number of topics within a corpus. The Zimm Approach, when tested against various subsets of the 20newsgroup dataset, produced the correct number of topics in 7 of 9 subsets vs. 0 of 9 using highest coherence value produced by LDA.
{"title":"Approaches to improve preprocessing for Latent Dirichlet Allocation topic modeling","authors":"Jamie Zimmermann , Lance E. Champagne , John M. Dickens , Benjamin T. Hazen","doi":"10.1016/j.dss.2024.114310","DOIUrl":"10.1016/j.dss.2024.114310","url":null,"abstract":"<div><p>As a part of natural language processing (NLP), the intent of topic modeling is to identify topics in textual corpora with limited human input. Current topic modeling techniques, like Latent Dirichlet Allocation (LDA), are limited in the pre-processing steps and currently require human judgement, increasing analysis time and opportunities for error. The purpose of this research is to allay some of those limitations by introducing new approaches to improve coherence without adding computational complexity and provide an objective method for determining the number of topics within a corpus. First, we identify a requirement for a more robust stop words list and introduce a new dimensionality-reduction heuristic that exploits the number of words within a document to infer importance to word choice. Second, we develop an eigenvalue technique to determine the number of topics within a corpus. Third, we combine all of these techniques into the Zimm Approach, which produces higher quality results than LDA in determining the number of topics within a corpus. The Zimm Approach, when tested against various subsets of the 20newsgroup dataset, produced the correct number of topics in 7 of 9 subsets vs. 0 of 9 using highest coherence value produced by LDA.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"185 ","pages":"Article 114310"},"PeriodicalIF":6.7,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142088817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-27DOI: 10.1016/j.dss.2024.114311
Lang Fang, Zhendong Pan, Jiafu Tang
We consider how to make dynamic pricing decision for Chinese Online (COL) at T time-points, an online publisher that allow authors to sell their ongoing book projects. Instead of paying for a book, readers pay for each chapter (pay-per-chapter mode) of the ongoing book project. This mode allows readers to pay for as many chapters as they want without taking the risk that the releasing of new chapters might be delayed or stopped. Despite of the dynamics of chapter-by-chapter released of COL products, the fixed pricing strategy (FPS) does not make fully use of the reading data generated by releasing chapters of the ongoing book. We propose a learning-based dynamic pricing strategy (LDPS) that exploits the newly information to maximize cumulative revenue for the publisher. The LDPS captures the ever changing features of readers. It employs the Thompson sampling method to balance the exploration of investigating different prices sufficiently with the exploitation of settling on the optimal price. Taking COL as a case study and implementing our strategy in the context of the aforementioned real-life data set, we show that LDPS outperform several classical strategies such as Greedy, Prior-Free TS and Prior-Given TS, and average revenue of LDPS is increased by 0.5 % average per time-point compared to the publisher's historical decisions. We also provide some management implications for the COL publisher by analyzing the pricing range of different genres of books and the choice of the exploration threshold parameter.
中文在线(COL)是一家允许作者销售其正在进行的图书项目的在线出版商,我们考虑的是如何在 T 个时间点为中文在线做出动态定价决策。读者不是为一本书付费,而是为正在进行的图书项目的每一章付费(按章付费模式)。这种模式允许读者按章节付费,而不必承担新章节发布可能被推迟或停止的风险。尽管 COL 产品具有逐章发布的动态性,但固定定价策略(FPS)并不能充分利用正在进行的图书章节发布所产生的阅读数据。我们提出了一种基于学习的动态定价策略(LDPS),它能利用新信息为出版商带来最大的累积收益。LDPS 抓住了读者不断变化的特点。它采用汤普森抽样方法,在充分调查不同价格的探索与确定最佳价格的利用之间取得平衡。我们以 COL 为案例,在上述真实数据集的背景下实施了我们的策略,结果表明 LDPS 优于贪婪策略、无优先权 TS 和优先权给定 TS 等几种经典策略,与出版商的历史决策相比,LDPS 的平均收入在每个时间点平均提高了 0.5%。我们还通过分析不同类型图书的定价范围和探索阈值参数的选择,为 COL 出版商提供了一些管理启示。
{"title":"Learning-based dynamic pricing strategy with pay-per-chapter mode for online publisher with case study of COL","authors":"Lang Fang, Zhendong Pan, Jiafu Tang","doi":"10.1016/j.dss.2024.114311","DOIUrl":"10.1016/j.dss.2024.114311","url":null,"abstract":"<div><p>We consider how to make dynamic pricing decision for Chinese Online (COL) at <em>T</em> time-points, an online publisher that allow authors to sell their ongoing book projects. Instead of paying for a book, readers pay for each chapter (pay-per-chapter mode) of the ongoing book project. This mode allows readers to pay for as many chapters as they want without taking the risk that the releasing of new chapters might be delayed or stopped. Despite of the dynamics of chapter-by-chapter released of COL products, the fixed pricing strategy (FPS) does not make fully use of the reading data generated by releasing chapters of the ongoing book. We propose a learning-based dynamic pricing strategy (LDPS) that exploits the newly information to maximize cumulative revenue for the publisher. The LDPS captures the ever changing features of readers. It employs the Thompson sampling method to balance the exploration of investigating different prices sufficiently with the exploitation of settling on the optimal price. Taking COL as a case study and implementing our strategy in the context of the aforementioned real-life data set, we show that LDPS outperform several classical strategies such as Greedy, Prior-Free TS and Prior-Given TS, and average revenue of LDPS is increased by 0.5 % average per time-point compared to the publisher's historical decisions. We also provide some management implications for the COL publisher by analyzing the pricing range of different genres of books and the choice of the exploration threshold parameter.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"186 ","pages":"Article 114311"},"PeriodicalIF":6.7,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142150986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1016/j.dss.2024.114305
Xiaoge Zhang , Indranil Bose
<div><p>The conventional aggregated performance measure (i.e., mean squared error) with respect to the whole dataset would not provide desired safety and quality assurance for each individual prediction made by a machine learning model in risk-sensitive regression problems. In this paper, we propose an informative indicator <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span> to quantify model reliability for individual prediction (MRIP) for the purpose of safeguarding the usage of machine learning (ML) models in mission-critical applications. Specifically, we define the reliability of a ML model with respect to its prediction on each individual input <span><math><mi>x</mi></math></span> as the probability of the observed difference between the prediction of ML model and the actual observation falling within a small interval when the input <span><math><mi>x</mi></math></span> varies within a small range subject to a preset distance constraint, namely <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced><mo>=</mo><mi>P</mi><mfenced><mrow></mrow><mrow><msup><mi>y</mi><mo>∗</mo></msup><mo>−</mo><msup><mover><mi>y</mi><mo>̂</mo></mover><mo>∗</mo></msup></mrow><mrow><mspace></mspace><mo>≤</mo><mi>ε</mi></mrow><mrow><msup><mi>x</mi><mo>∗</mo></msup><mo>∈</mo><mi>B</mi><mfenced><mi>x</mi></mfenced></mrow></mfenced></math></span>, where <span><math><msup><mi>y</mi><mo>∗</mo></msup></math></span> denotes the observed target value for the input <span><math><msup><mi>x</mi><mo>∗</mo></msup><mo>,</mo></math></span> <span><math><msup><mover><mi>y</mi><mo>̂</mo></mover><mo>∗</mo></msup></math></span> denotes the model prediction for the input <span><math><msup><mi>x</mi><mo>∗</mo></msup></math></span>, and <span><math><msup><mi>x</mi><mo>∗</mo></msup></math></span> is an input in the neighborhood of <span><math><mi>x</mi></math></span> subject to the constraint <span><math><mi>B</mi><mfenced><mi>x</mi></mfenced><mo>=</mo><mfenced><mrow><mfenced><msup><mi>x</mi><mo>∗</mo></msup></mfenced><mspace></mspace><mfenced><mrow><msup><mi>x</mi><mo>∗</mo></msup><mo>−</mo><mi>x</mi></mrow></mfenced><mo>≤</mo><mi>δ</mi></mrow></mfenced></math></span>. The developed MRIP indicator <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span> provides a direct, objective, quantitative, and general-purpose measure of “reliability” or the probability of success of the ML model for each individual prediction by fully exploiting the local information associated with the input <span><math><mi>x</mi></math></span> and ML model. Next, to mitigate the intensive computational effort involved in MRIP estimation, we develop a two-stage ML-based framework to directly learn the relationship between <span><math><mi>x</mi></math></span> and its MRIP <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span>, thus enabling to provide the reliability estimate <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span> for any unseen input instantly. Thirdly, we pr
在风险敏感回归问题中,传统的针对整个数据集的汇总性能指标(即均方误差)无法为机器学习模型所做的每个单独预测提供所需的安全和质量保证。在本文中,我们提出了一个信息指标ℛx 来量化单个预测的模型可靠性(MRIP),以保障机器学习(ML)模型在关键任务应用中的使用。具体来说,我们将 ML 模型对每个输入 x 的预测可靠性定义为:当输入 x 在一个小范围内变化时,ML 模型的预测值与实际观测值之间的差值落在一个小区间内的概率,该小区间受预设距离约束、即ℛx=Py∗-ŷ∗≤εx∗∈Bx,其中 y∗ 表示输入 x∗ 的观测目标值、ŷ∗ 表示输入 x∗ 的模型预测值,x∗ 是 x 附近的输入,受 Bx=x∗x∗-x≤δ 约束。所开发的 MRIP 指标ℛx 通过充分利用与输入 x 和 ML 模型相关的本地信息,为每个单独预测的 "可靠性 "或 ML 模型的成功概率提供了直接、客观、定量和通用的衡量标准。其次,为了减轻 MRIP 估计所需的大量计算工作,我们开发了一个基于 ML 的两阶段框架,直接学习 x 与其 MRIP ℛx 之间的关系,从而能够为任何未见输入即时提供可靠性估计ℛx。第三,我们提出了一种基于信息增益的方法,帮助确定ℛx 的阈值,以支持何时接受或放弃依赖 ML 模型预测的决策。在广泛的现实世界数据集上进行的综合计算实验以及与现有方法的定量比较表明,所开发的基于 ML 的 MRIP 估算框架在提高单个预测的可靠性估计方面表现出色,因此,当在风险敏感环境中采用 ML 模型时,MRIP 指标ℛx 提供了一层必不可少的安全网。
{"title":"Reliability estimation for individual predictions in machine learning systems: A model reliability-based approach","authors":"Xiaoge Zhang , Indranil Bose","doi":"10.1016/j.dss.2024.114305","DOIUrl":"10.1016/j.dss.2024.114305","url":null,"abstract":"<div><p>The conventional aggregated performance measure (i.e., mean squared error) with respect to the whole dataset would not provide desired safety and quality assurance for each individual prediction made by a machine learning model in risk-sensitive regression problems. In this paper, we propose an informative indicator <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span> to quantify model reliability for individual prediction (MRIP) for the purpose of safeguarding the usage of machine learning (ML) models in mission-critical applications. Specifically, we define the reliability of a ML model with respect to its prediction on each individual input <span><math><mi>x</mi></math></span> as the probability of the observed difference between the prediction of ML model and the actual observation falling within a small interval when the input <span><math><mi>x</mi></math></span> varies within a small range subject to a preset distance constraint, namely <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced><mo>=</mo><mi>P</mi><mfenced><mrow></mrow><mrow><msup><mi>y</mi><mo>∗</mo></msup><mo>−</mo><msup><mover><mi>y</mi><mo>̂</mo></mover><mo>∗</mo></msup></mrow><mrow><mspace></mspace><mo>≤</mo><mi>ε</mi></mrow><mrow><msup><mi>x</mi><mo>∗</mo></msup><mo>∈</mo><mi>B</mi><mfenced><mi>x</mi></mfenced></mrow></mfenced></math></span>, where <span><math><msup><mi>y</mi><mo>∗</mo></msup></math></span> denotes the observed target value for the input <span><math><msup><mi>x</mi><mo>∗</mo></msup><mo>,</mo></math></span> <span><math><msup><mover><mi>y</mi><mo>̂</mo></mover><mo>∗</mo></msup></math></span> denotes the model prediction for the input <span><math><msup><mi>x</mi><mo>∗</mo></msup></math></span>, and <span><math><msup><mi>x</mi><mo>∗</mo></msup></math></span> is an input in the neighborhood of <span><math><mi>x</mi></math></span> subject to the constraint <span><math><mi>B</mi><mfenced><mi>x</mi></mfenced><mo>=</mo><mfenced><mrow><mfenced><msup><mi>x</mi><mo>∗</mo></msup></mfenced><mspace></mspace><mfenced><mrow><msup><mi>x</mi><mo>∗</mo></msup><mo>−</mo><mi>x</mi></mrow></mfenced><mo>≤</mo><mi>δ</mi></mrow></mfenced></math></span>. The developed MRIP indicator <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span> provides a direct, objective, quantitative, and general-purpose measure of “reliability” or the probability of success of the ML model for each individual prediction by fully exploiting the local information associated with the input <span><math><mi>x</mi></math></span> and ML model. Next, to mitigate the intensive computational effort involved in MRIP estimation, we develop a two-stage ML-based framework to directly learn the relationship between <span><math><mi>x</mi></math></span> and its MRIP <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span>, thus enabling to provide the reliability estimate <span><math><mi>ℛ</mi><mfenced><mi>x</mi></mfenced></math></span> for any unseen input instantly. Thirdly, we pr","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"186 ","pages":"Article 114305"},"PeriodicalIF":6.7,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142151016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-22DOI: 10.1016/j.dss.2024.114304
Michael T. Lash
The use of machine learning (ML) models in decision-making contexts, particularly those used in high-stakes decision-making, are fraught with issue and peril since a person – not a machine – must ultimately be held accountable for the consequences of decisions made using such systems. Machine learning explainability (MLX) promises to provide decision-makers with prediction-specific rationale, assuring them that the model-elicited predictions are made for the right reasons and are thus reliable. Few works explicitly consider this key human-in-the-loop (HITL) component, however. In this work we propose HEX, a human-in-the-loop deep reinforcement learning approach to MLX. HEX incorporates 0-distrust projection to synthesize decider-specific explainers that produce explanations strictly in terms of a decider’s preferred explanatory features using any classification model. Our formulation explicitly considers the decision boundary of the ML model in question using a proposed explanatory point mode of explanation, thus ensuring explanations are specific to the ML model in question. We empirically evaluate HEX against other competing methods, finding that HEX is competitive with the state-of-the-art and outperforms other methods in human-in-the-loop scenarios. We conduct a randomized, controlled laboratory experiment utilizing actual explanations elicited from both HEX and competing methods. We causally establish that our method increases decider’s trust and tendency to rely on trusted features.
在决策环境中使用机器学习(ML)模型,尤其是那些用于高风险决策的模型,充满了问题和危险,因为最终必须由人--而不是机器--来对使用此类系统所做决策的后果负责。机器学习的可解释性(MLX)有望为决策者提供预测的具体理由,确保他们相信由模型引发的预测是出于正确的原因,因而是可靠的。然而,很少有作品明确考虑到这一关键的 "人在回路中"(HITL)要素。在这项工作中,我们提出了 HEX,一种针对 MLX 的人在环深度强化学习方法。HEX 结合了 0 不信任投射,可合成针对决策者的解释器,严格按照决策者的首选解释特征,使用任何分类模型生成解释。我们的表述明确考虑了相关 ML 模型的决策边界,使用了建议的解释点解释模式,从而确保解释是针对相关 ML 模型的。我们对 HEX 与其他竞争方法进行了实证评估,发现 HEX 与最先进的方法相比具有竞争力,在人类在环场景中的表现优于其他方法。我们利用 HEX 和其他竞争方法得出的实际解释进行了随机对照实验室实验。我们从因果关系上证实,我们的方法提高了决定者的信任度,并倾向于依赖可信的特征。
{"title":"HEX: Human-in-the-loop explainability via deep reinforcement learning","authors":"Michael T. Lash","doi":"10.1016/j.dss.2024.114304","DOIUrl":"10.1016/j.dss.2024.114304","url":null,"abstract":"<div><div>The use of machine learning (ML) models in decision-making contexts, particularly those used in high-stakes decision-making, are fraught with issue and peril since a person – not a machine – must ultimately be held accountable for the consequences of decisions made using such systems. Machine learning explainability (MLX) promises to provide decision-makers with prediction-specific rationale, assuring them that the model-elicited predictions are made <em>for the right reasons</em> and are thus reliable. Few works explicitly consider this key human-in-the-loop (HITL) component, however. In this work we propose HEX, a human-in-the-loop deep reinforcement learning approach to MLX. HEX incorporates 0-distrust projection to synthesize decider-specific explainers that produce explanations strictly in terms of a decider’s preferred explanatory features using any classification model. Our formulation explicitly considers the decision boundary of the ML model in question using a proposed <em>explanatory point</em> mode of explanation, thus ensuring explanations are specific to the ML model in question. We empirically evaluate HEX against other competing methods, finding that HEX is competitive with the state-of-the-art and outperforms other methods in human-in-the-loop scenarios. We conduct a randomized, controlled laboratory experiment utilizing actual explanations elicited from both HEX and competing methods. We causally establish that our method increases decider’s trust and tendency to rely on trusted features.</div></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"187 ","pages":"Article 114304"},"PeriodicalIF":6.7,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142531486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-21DOI: 10.1016/j.dss.2024.114307
Viswanath Venkatesh
Rooted in the paradigm changes that accompany the metaverse, this essay proposes research directions covering three major and interconnected aspects of the metaverse ecosystem. First, I propose five research directions connected to the design of technological solutions for the metaverse. Second, I propose five research directions tied to the study of the impact of the adoption and use of these developed technological solutions. Third, I propose the five research directions that relate to understanding the impacts of the so-developed and so-adopted technological solutions. Finally, I propose five overarching research directions that cut across the design-adoption-impacts framework. Taken together, these directions provide holistic guidance for the investigation of the metaverse ecosystem and its short-, medium-, and long-term implications for research.
{"title":"Paradigm changing metaverse: Future research directions in design, technology adoption and use, and impacts","authors":"Viswanath Venkatesh","doi":"10.1016/j.dss.2024.114307","DOIUrl":"10.1016/j.dss.2024.114307","url":null,"abstract":"<div><div>Rooted in the paradigm changes that accompany the metaverse, this essay proposes research directions covering three major and interconnected aspects of the metaverse ecosystem. First, I propose five research directions connected to the design of technological solutions for the metaverse. Second, I propose five research directions tied to the study of the impact of the adoption and use of these developed technological solutions. Third, I propose the five research directions that relate to understanding the impacts of the so-developed and so-adopted technological solutions. Finally, I propose five overarching research directions that cut across the design-adoption-impacts framework. Taken together, these directions provide holistic guidance for the investigation of the metaverse ecosystem and its short-, medium-, and long-term implications for research.</div></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"188 ","pages":"Article 114307"},"PeriodicalIF":6.7,"publicationDate":"2024-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142702534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We propose a novel curvature-based indicator constructed on log-price time series that captures an interplay between trend, acceleration, and volatility found relevant to quantify risks and improve trading strategies. We apply it to diagnose explosive bubble-like behaviors in cryptocurrency price time series and provide early warning signals of impending market shifts or increased volatility. This improves significantly on standard statistical tests such as the Generalized Supremum Augmented Dickey–Fuller (GSADF) and the Backward SADF tests. Furthermore, the incorporation of our curvature-based indicator as a feature into the Light Gradient Boosting Machine enhances its predictive capabilities, as measured by classification accuracy and trading performance.
{"title":"Generalized visible curvature: An indicator for bubble identification and price trend prediction in cryptocurrencies","authors":"Qun Zhang , Canxuan Xie , Zhaoju Weng , Didier Sornette , Ke Wu","doi":"10.1016/j.dss.2024.114309","DOIUrl":"10.1016/j.dss.2024.114309","url":null,"abstract":"<div><p>We propose a novel curvature-based indicator constructed on log-price time series that captures an interplay between trend, acceleration, and volatility found relevant to quantify risks and improve trading strategies. We apply it to diagnose explosive bubble-like behaviors in cryptocurrency price time series and provide early warning signals of impending market shifts or increased volatility. This improves significantly on standard statistical tests such as the Generalized Supremum Augmented Dickey–Fuller (GSADF) and the Backward SADF tests. Furthermore, the incorporation of our curvature-based indicator as a feature into the Light Gradient Boosting Machine enhances its predictive capabilities, as measured by classification accuracy and trading performance.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"185 ","pages":"Article 114309"},"PeriodicalIF":6.7,"publicationDate":"2024-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142083568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-20DOI: 10.1016/j.dss.2024.114308
Motahareh Pourbehzadi , Giti Javidi , C. Jordan Howell , Eden Kamar , Ehsan Sheybani
The Common Vulnerability Scoring System (CVSS) is widely used in the cybersecurity industry to assess the severity of vulnerabilities. However, manual assessments and human error can lead to delays and inconsistencies. This study employs situational awareness theory to develop an automated decision support system, integrating perception, comprehension, and projection components to enhance effectiveness. Specifically, an interpretable principal component analysis (iPCA) combined with machine learning is utilized to forecast CVSS scores using text descriptions from the Common Vulnerabilities and Exposures (CVE) database. Different forecasting approaches, including traditional machine learning models, Long-Short Term Memory Neural Networks, and Transformer architectures (ChatGPT) are compared to determine the best performance. The results show that iPCA combined with support vector regression achieves a high performance (R2 = 98%) in predicting CVSS scores using CVE text descriptions. The results indicate that the variability, length, and details in the vulnerability description contribute to the performance of the transformer model. These findings are consistent across vulnerability descriptions from six companies between 2017 and 2019. The study's outcomes have the potential to enhance organizations' security posture, improving situational awareness and enabling better managerial decision-making in cybersecurity.
{"title":"Enhanced (cyber) situational awareness: Using interpretable principal component analysis (iPCA) to automate vulnerability severity scoring","authors":"Motahareh Pourbehzadi , Giti Javidi , C. Jordan Howell , Eden Kamar , Ehsan Sheybani","doi":"10.1016/j.dss.2024.114308","DOIUrl":"10.1016/j.dss.2024.114308","url":null,"abstract":"<div><p>The Common Vulnerability Scoring System (CVSS) is widely used in the cybersecurity industry to assess the severity of vulnerabilities. However, manual assessments and human error can lead to delays and inconsistencies. This study employs situational awareness theory to develop an automated decision support system, integrating perception, comprehension, and projection components to enhance effectiveness. Specifically, an interpretable principal component analysis (iPCA) combined with machine learning is utilized to forecast CVSS scores using text descriptions from the Common Vulnerabilities and Exposures (CVE) database. Different forecasting approaches, including traditional machine learning models, Long-Short Term Memory Neural Networks, and Transformer architectures (ChatGPT) are compared to determine the best performance. The results show that iPCA combined with support vector regression achieves a high performance (R<sup>2</sup> = 98%) in predicting CVSS scores using CVE text descriptions. The results indicate that the variability, length, and details in the vulnerability description contribute to the performance of the transformer model. These findings are consistent across vulnerability descriptions from six companies between 2017 and 2019. The study's outcomes have the potential to enhance organizations' security posture, improving situational awareness and enabling better managerial decision-making in cybersecurity.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"186 ","pages":"Article 114308"},"PeriodicalIF":6.7,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142150985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-12DOI: 10.1016/j.dss.2024.114306
Xian Cao , Timothy B. Folta , Hongfei Li , Ruoqing Zhu
In today's digital economy, virtually everything from products and services to political debates and cultural phenomena can spark WOM on social media. Analyzing online WOM poses at least three challenges. First, online WOM typically consists of unstructured data that can transform into myriad variables, necessitating effective dimension reduction. Second, online WOM is often continuous and dynamic, with the potential for rapid, time-varying changes. Third, significant events may trigger symmetric or asymmetric responses across various entities, resulting in “bursty” and intense WOM from multiple sources. To address these challenges, we introduce a new computationally efficient method—multi-view sequential canonical covariance analysis. This method is designed to solve the myriad online WOM conversational dimensions, detect online WOM dynamic trends, and examine the shared online WOM across different entities. This approach not only enhances the capability to swiftly interpret and respond to online WOM data but also shows potential to significantly improve decision-making processes across various contexts. We illustrate the method's benefits through two empirical examples, demonstrating its potential to provide profound insights into online WOM dynamics and its extensive applicability in both academic research and practical scenarios.
{"title":"Analyzing the online word of mouth dynamics: A novel approach","authors":"Xian Cao , Timothy B. Folta , Hongfei Li , Ruoqing Zhu","doi":"10.1016/j.dss.2024.114306","DOIUrl":"10.1016/j.dss.2024.114306","url":null,"abstract":"<div><p>In today's digital economy, virtually everything from products and services to political debates and cultural phenomena can spark WOM on social media. Analyzing online WOM poses at least three challenges. First, online WOM typically consists of unstructured data that can transform into myriad variables, necessitating effective dimension reduction. Second, online WOM is often continuous and dynamic, with the potential for rapid, time-varying changes. Third, significant events may trigger symmetric or asymmetric responses across various entities, resulting in “bursty” and intense WOM from multiple sources. To address these challenges, we introduce a new computationally efficient method—multi-view sequential canonical covariance analysis. This method is designed to solve the myriad online WOM conversational dimensions, detect online WOM dynamic trends, and examine the shared online WOM across different entities. This approach not only enhances the capability to swiftly interpret and respond to online WOM data but also shows potential to significantly improve decision-making processes across various contexts. We illustrate the method's benefits through two empirical examples, demonstrating its potential to provide profound insights into online WOM dynamics and its extensive applicability in both academic research and practical scenarios.</p></div>","PeriodicalId":55181,"journal":{"name":"Decision Support Systems","volume":"185 ","pages":"Article 114306"},"PeriodicalIF":6.7,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141998137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}