首页 > 最新文献

arXiv - STAT - Applications最新文献

英文 中文
Bayesian Dynamic Generalized Additive Model for Mortality during COVID-19 Pandemic COVID-19 大流行期间死亡率的贝叶斯动态广义加法模型
Pub Date : 2024-09-04 DOI: arxiv-2409.02378
Wei Zhang, Antonietta Mira, Ernst C. Wit
While COVID-19 has resulted in a significant increase in global mortalityrates, the impact of the pandemic on mortality from other causes remainsuncertain. To gain insight into the broader effects of COVID-19 on variouscauses of death, we analyze an Italian dataset that includes monthly mortalitycounts for different causes from January 2015 to December 2020. Our approachinvolves a generalized additive model enhanced with correlated random effects.The generalized additive model component effectively captures non-linearrelationships between various covariates and mortality rates, while the randomeffects are multivariate time series observations recorded in variouslocations, and they embody information on the dependence structure presentamong geographical locations and different causes of mortality. Adopting aBayesian framework, we impose suitable priors on the model parameters. Forefficient posterior computation, we employ variational inference, specificallyfor fixed effect coefficients and random effects, Gaussian variationalapproximation is assumed, which streamlines the analysis process. Theoptimisation is performed using a coordinate ascent variational inferencealgorithm and several computational strategies are implemented along the way toaddress the issues arising from the high dimensional nature of the data,providing accelerated and stabilised parameter estimation and statisticalinference.
尽管 COVID-19 已导致全球死亡率大幅上升,但这一流行病对其他原因造成的死亡率的影响仍不确定。为了深入了解 COVID-19 对各种死因的广泛影响,我们分析了一个意大利数据集,其中包括 2015 年 1 月至 2020 年 12 月期间不同死因的月死亡率。广义加法模型部分有效地捕捉了各种协变量与死亡率之间的非线性关系,而随机效应是在不同地点记录的多变量时间序列观测值,它们体现了地理位置和不同死因之间存在的依赖结构信息。采用贝叶斯框架,我们对模型参数施加了适当的先验。为了有效地进行后验计算,我们采用了变异推断法,特别是针对固定效应系数和随机效应,假设采用高斯变异近似法,从而简化了分析过程。我们使用坐标上升变异推理算法进行优化,并在此过程中实施了多种计算策略,以解决高维数据带来的问题,从而加速并稳定参数估计和统计推断。
{"title":"Bayesian Dynamic Generalized Additive Model for Mortality during COVID-19 Pandemic","authors":"Wei Zhang, Antonietta Mira, Ernst C. Wit","doi":"arxiv-2409.02378","DOIUrl":"https://doi.org/arxiv-2409.02378","url":null,"abstract":"While COVID-19 has resulted in a significant increase in global mortality\u0000rates, the impact of the pandemic on mortality from other causes remains\u0000uncertain. To gain insight into the broader effects of COVID-19 on various\u0000causes of death, we analyze an Italian dataset that includes monthly mortality\u0000counts for different causes from January 2015 to December 2020. Our approach\u0000involves a generalized additive model enhanced with correlated random effects.\u0000The generalized additive model component effectively captures non-linear\u0000relationships between various covariates and mortality rates, while the random\u0000effects are multivariate time series observations recorded in various\u0000locations, and they embody information on the dependence structure present\u0000among geographical locations and different causes of mortality. Adopting a\u0000Bayesian framework, we impose suitable priors on the model parameters. For\u0000efficient posterior computation, we employ variational inference, specifically\u0000for fixed effect coefficients and random effects, Gaussian variational\u0000approximation is assumed, which streamlines the analysis process. The\u0000optimisation is performed using a coordinate ascent variational inference\u0000algorithm and several computational strategies are implemented along the way to\u0000address the issues arising from the high dimensional nature of the data,\u0000providing accelerated and stabilised parameter estimation and statistical\u0000inference.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"36 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142224684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating Treatment Effect Heterogeneity in Psychiatry: A Review and Tutorial with Causal Forests 估算精神病学中的治疗效果异质性:因果森林回顾与教程
Pub Date : 2024-09-03 DOI: arxiv-2409.01578
Erik Sverdrup, Maria Petukhova, Stefan Wager
Flexible machine learning tools are being used increasingly to estimateheterogeneous treatment effects. This paper gives an accessible tutorialdemonstrating the use of the causal forest algorithm, available in the Rpackage grf. We start with a brief non-technical overview of treatment effectestimation methods with a focus on estimation in observational studies,although similar methods can be used in experimental studies. We then discussthe logic of estimating heterogeneous effects using the extension of the randomforest algorithm implemented in grf. Finally, we illustrate causal forest byconducting a secondary analysis on the extent to which individual differencesin resilience to high combat stress can be measured among US Army soldiersdeploying to Afghanistan based on information about these soldiers availableprior to deployment. Throughout we illustrate simple and interpretableexercises for both model selection and evaluation, including targeting operatorcharacteristics curves, Qini curves, area-under-the-curve summaries, and bestlinear projections. A replication script with simulated data is available atgithub.com/grf-labs/grf/tree/master/experiments/ijmpr
灵活的机器学习工具越来越多地被用于估计异质性治疗效果。本文通过通俗易懂的教程演示了因果森林算法的使用,该算法可在 Rpackage grf 中获得。我们首先从非技术角度简要介绍了治疗效果估算方法,重点是观察性研究中的估算,尽管类似方法也可用于实验研究。然后,我们将讨论使用 grf 中实现的随机森林算法扩展来估计异质性效应的逻辑。最后,我们通过对部署到阿富汗的美国陆军士兵进行二次分析来说明因果森林在多大程度上可以根据这些士兵部署前的信息来衡量他们对高战斗压力的适应能力的个体差异。在整个分析过程中,我们展示了用于模型选择和评估的简单且可解释的练习,包括目标操作者特征曲线、基尼曲线、曲线下面积总结以及最佳线性预测。带有模拟数据的复制脚本可在以下网站获取:github.com/grf-labs/grf/tree/master/experiments/ijmpr
{"title":"Estimating Treatment Effect Heterogeneity in Psychiatry: A Review and Tutorial with Causal Forests","authors":"Erik Sverdrup, Maria Petukhova, Stefan Wager","doi":"arxiv-2409.01578","DOIUrl":"https://doi.org/arxiv-2409.01578","url":null,"abstract":"Flexible machine learning tools are being used increasingly to estimate\u0000heterogeneous treatment effects. This paper gives an accessible tutorial\u0000demonstrating the use of the causal forest algorithm, available in the R\u0000package grf. We start with a brief non-technical overview of treatment effect\u0000estimation methods with a focus on estimation in observational studies,\u0000although similar methods can be used in experimental studies. We then discuss\u0000the logic of estimating heterogeneous effects using the extension of the random\u0000forest algorithm implemented in grf. Finally, we illustrate causal forest by\u0000conducting a secondary analysis on the extent to which individual differences\u0000in resilience to high combat stress can be measured among US Army soldiers\u0000deploying to Afghanistan based on information about these soldiers available\u0000prior to deployment. Throughout we illustrate simple and interpretable\u0000exercises for both model selection and evaluation, including targeting operator\u0000characteristics curves, Qini curves, area-under-the-curve summaries, and best\u0000linear projections. A replication script with simulated data is available at\u0000github.com/grf-labs/grf/tree/master/experiments/ijmpr","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"5 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142188153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Conditional multi-step attribution for climate forcings 气候作用力的有条件多步骤归因
Pub Date : 2024-09-02 DOI: arxiv-2409.01396
Christopher R. Wentland, Michael Weylandt, Laura P. Swiler, Thomas S. Ehrmann, Diana Bull
Attribution of climate impacts to a source forcing is critical tounderstanding, communicating, and addressing the effects of human influence onthe climate. While standard attribution methods, such as optimalfingerprinting, have been successfully applied to long-term, widespread effectssuch as global surface temperature warming, they often struggle in lowsignal-to-noise regimes, typical of short-term climate forcings or climatevariables which are loosely related to the forcing. Single-step approaches,which directly relate a source forcing and final impact, are unable to utilizeadditional climate information to improve attribution certainty. To addressthis shortcoming, this paper presents a novel multi-step attribution approachwhich is capable of analyzing multiple variables conditionally. A connectedseries of climate effects are treated as dependent, and relationships found inintermediary steps of a causal pathway are leveraged to better characterize theforcing impact. This enables attribution of the forcing level responsible forthe observed impacts, while equivalent single-step approaches fail. Utilizing ascalar feature describing the forcing impact, simple forcing response models,and a conditional Bayesian formulation, this method can incorporate severalcausal pathways to identify the correct forcing magnitude. As an exemplar of ashort-term, high-variance forcing, we demonstrate this method for the 1991eruption of Mt. Pinatubo. Results indicate that including stratospheric andsurface temperature and radiative flux measurements increases attributioncertainty compared to analyses derived solely from temperature measurements.This framework has potential to improve climate attribution assessments forboth geoengineering projects and long-term climate change, for which standardattribution methods may fail.
将气候影响归因于源强迫对于理解、交流和解决人类对气候的影响至关重要。虽然标准的归因方法,如最优指纹法,已成功应用于长期、广泛的影响,如全球地表温度变暖,但在低信噪比情况下,即典型的短期气候强迫或与强迫关系松散的气候变量中,这些方法往往难以奏效。单步方法直接将源强迫与最终影响联系起来,无法利用额外的气候信息来提高归因的确定性。为了解决这个问题,本文提出了一种新颖的多步骤归因方法,能够对多个变量进行条件分析。一系列相关的气候效应被视为从属效应,在因果路径的中间步骤中发现的关系被用来更好地描述强迫影响。这样就能确定造成观测到的影响的作用力水平,而单步方法则无法做到这一点。利用描述作用力影响的特征、简单的作用力响应模型和条件贝叶斯公式,这种方法可以结合多个因果途径来确定正确的作用力大小。我们以 1991 年皮纳图博火山爆发为例,演示了这种短期高变异强迫方法。结果表明,与仅从温度测量得出的分析结果相比,包括平流层和地表温度及辐射通量测量在内的分析结果增加了气候归因的确定性。
{"title":"Conditional multi-step attribution for climate forcings","authors":"Christopher R. Wentland, Michael Weylandt, Laura P. Swiler, Thomas S. Ehrmann, Diana Bull","doi":"arxiv-2409.01396","DOIUrl":"https://doi.org/arxiv-2409.01396","url":null,"abstract":"Attribution of climate impacts to a source forcing is critical to\u0000understanding, communicating, and addressing the effects of human influence on\u0000the climate. While standard attribution methods, such as optimal\u0000fingerprinting, have been successfully applied to long-term, widespread effects\u0000such as global surface temperature warming, they often struggle in low\u0000signal-to-noise regimes, typical of short-term climate forcings or climate\u0000variables which are loosely related to the forcing. Single-step approaches,\u0000which directly relate a source forcing and final impact, are unable to utilize\u0000additional climate information to improve attribution certainty. To address\u0000this shortcoming, this paper presents a novel multi-step attribution approach\u0000which is capable of analyzing multiple variables conditionally. A connected\u0000series of climate effects are treated as dependent, and relationships found in\u0000intermediary steps of a causal pathway are leveraged to better characterize the\u0000forcing impact. This enables attribution of the forcing level responsible for\u0000the observed impacts, while equivalent single-step approaches fail. Utilizing a\u0000scalar feature describing the forcing impact, simple forcing response models,\u0000and a conditional Bayesian formulation, this method can incorporate several\u0000causal pathways to identify the correct forcing magnitude. As an exemplar of a\u0000short-term, high-variance forcing, we demonstrate this method for the 1991\u0000eruption of Mt. Pinatubo. Results indicate that including stratospheric and\u0000surface temperature and radiative flux measurements increases attribution\u0000certainty compared to analyses derived solely from temperature measurements.\u0000This framework has potential to improve climate attribution assessments for\u0000both geoengineering projects and long-term climate change, for which standard\u0000attribution methods may fail.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"180 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tonal coarticulation revisited: functional covariance analysis to investigate the planning of co-articulated tones by Standard Chinese speakers 再论声调共鸣:用功能协方差分析法研究标准汉语发音人的声调共鸣规划
Pub Date : 2024-09-02 DOI: arxiv-2409.01194
Valentina Masarotto, Yiya Chen
We aim to explain whether a stress memory task has a significant impact ontonal coarticulation. We contribute a novel approach to analyse tonalcoarticulation in phonetics, where several f0 contours are compared withrespect to their vibrations at higher resolution, something that in statisticalterms is called variation of the second order. We identify speech recordingfrequency curves as functional observations and harness inspiration from themathematical fields of functional data analysis and optimal transport. Byleveraging results from these two disciplines, we make one key observation:weidentify the time and frequency covariance functions as crucial features forcapturing the finer effects of tonal coarticulation. This observation leads usto propose a 2 steps approach where the mean functions are modelled viaGeneralized Additive Models, and the residuals of such models are investigatedfor any structure nested at covariance level. If such structure exist, wedescribe the variation manifested by the covariances through covarianceprincipal component analysis. The 2-steps approach allows to uncover anyvariation not explained by generalized additive modelling, as well as fill aknown shortcoming of these models into incorporating complex correlationstructures in the data. The proposed method is illustrated on an articulatorydataset contrasting the pronunciation non-sensical bi-syllabic combinations inthe presence of a short-memory challenge
我们的目的是解释重音记忆任务是否会对音调共鸣产生重大影响。我们为分析语音学中的声调共鸣贡献了一种新方法,即在更高分辨率下将多个 f0 等值线与其振动进行比较,这种方法在统计学术语中被称为二阶变异。我们将语音记录频率曲线确定为功能观测,并从功能数据分析和最优传输的数学领域获得灵感。利用这两个学科的成果,我们得出了一个关键结论:我们确定时间和频率协方差函数是捕捉音调共混效果的关键特征。这一观察结果促使我们提出了一种两步法,即通过广义加法模型对平均函数进行建模,并研究这些模型的残差是否存在任何嵌套在协方差水平上的结构。如果存在这种结构,我们将通过协方差主成分分析来描述协方差所表现出的变化。通过这两步方法,我们可以发现广义加法模型无法解释的任何变异,并弥补这些模型在纳入数据中复杂相关结构方面的已知缺陷。在一个发音数据集上对所提出的方法进行了说明,该数据集对比了在短时记忆挑战下非感性双音节组合的发音情况
{"title":"Tonal coarticulation revisited: functional covariance analysis to investigate the planning of co-articulated tones by Standard Chinese speakers","authors":"Valentina Masarotto, Yiya Chen","doi":"arxiv-2409.01194","DOIUrl":"https://doi.org/arxiv-2409.01194","url":null,"abstract":"We aim to explain whether a stress memory task has a significant impact on\u0000tonal coarticulation. We contribute a novel approach to analyse tonal\u0000coarticulation in phonetics, where several f0 contours are compared with\u0000respect to their vibrations at higher resolution, something that in statistical\u0000terms is called variation of the second order. We identify speech recording\u0000frequency curves as functional observations and harness inspiration from the\u0000mathematical fields of functional data analysis and optimal transport. By\u0000leveraging results from these two disciplines, we make one key observation:we\u0000identify the time and frequency covariance functions as crucial features for\u0000capturing the finer effects of tonal coarticulation. This observation leads us\u0000to propose a 2 steps approach where the mean functions are modelled via\u0000Generalized Additive Models, and the residuals of such models are investigated\u0000for any structure nested at covariance level. If such structure exist, we\u0000describe the variation manifested by the covariances through covariance\u0000principal component analysis. The 2-steps approach allows to uncover any\u0000variation not explained by generalized additive modelling, as well as fill a\u0000known shortcoming of these models into incorporating complex correlation\u0000structures in the data. The proposed method is illustrated on an articulatory\u0000dataset contrasting the pronunciation non-sensical bi-syllabic combinations in\u0000the presence of a short-memory challenge","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"84 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187906","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A method to convert traditional fingerprint ACE / ACE-V outputs ("identification", "inconclusive", "exclusion") to Bayes factors 将传统指纹 ACE / ACE-V 输出("识别"、"不确定"、"排除")转换为贝叶斯因子的方法
Pub Date : 2024-08-31 DOI: arxiv-2409.00451
Geoffrey Stewart Morrison
Fingerprint examiners appear to be reluctant to adopt probabilisticreasoning, statistical models, and empirical validation. The rate of adoptionof the likelihood-ratio framework by fingerprint practitioners appears to benear zero. A factor in the reluctance to adopt the likelihood-ratio frameworkmay be a perception that it would require a radical change in practice. Thepresent paper proposes a small step that would require minimal changes tocurrent practice. It proposes and demonstrates a method to convert traditionalfingerprint-examination outputs ("identification", "inconclusive", "exclusion")to well-calibrated Bayes factors. The method makes use of a beta-binomialmodel, and both uninformative and informative priors.
指纹检验员似乎不愿意采用概率推理、统计模型和经验验证。指纹鉴定人员采用似然比框架的比率似乎为零。不愿采用似然比框架的一个原因可能是认为这需要彻底改变做法。本文提出了一个小步骤,只需对目前的做法做出最小的改变。本文提出并演示了一种方法,可将传统的指纹检验结果("识别"、"不确定"、"排除")转换为校准良好的贝叶斯系数。该方法使用了贝塔-二叉模型以及非信息和信息先验。
{"title":"A method to convert traditional fingerprint ACE / ACE-V outputs (\"identification\", \"inconclusive\", \"exclusion\") to Bayes factors","authors":"Geoffrey Stewart Morrison","doi":"arxiv-2409.00451","DOIUrl":"https://doi.org/arxiv-2409.00451","url":null,"abstract":"Fingerprint examiners appear to be reluctant to adopt probabilistic\u0000reasoning, statistical models, and empirical validation. The rate of adoption\u0000of the likelihood-ratio framework by fingerprint practitioners appears to be\u0000near zero. A factor in the reluctance to adopt the likelihood-ratio framework\u0000may be a perception that it would require a radical change in practice. The\u0000present paper proposes a small step that would require minimal changes to\u0000current practice. It proposes and demonstrates a method to convert traditional\u0000fingerprint-examination outputs (\"identification\", \"inconclusive\", \"exclusion\")\u0000to well-calibrated Bayes factors. The method makes use of a beta-binomial\u0000model, and both uninformative and informative priors.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"2023 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Localizing Single and Multiple Oscillatory Sources: A Frequency Divider Approach 定位单个和多个振荡源:分频器方法
Pub Date : 2024-08-31 DOI: arxiv-2409.00566
Rajasekhar Anguluri, Anamitra Pal
Localizing sources of troublesome oscillations, particularly forcedoscillations (FOs), in power systems has received considerable attention overthe last few years. This is driven in part by the massive deployment of phasormeasurement units (PMUs) that capture these oscillations when they occur; andin part by the increasing incidents of FOs due to malfunctioning components,wind power fluctuations, and/or cyclic loads. Capitalizing on the frequencydivider formula of [1], we develop methods to localize single and multipleoscillatory sources using bus frequency measurements. The method to localize asingle oscillation source does not require knowledge of network parameters.However, the method for localizing FOs caused by multiple sources requires thisknowledge. We explain the reasoning behind this knowledge difference as well asdemonstrate the success of our methods for source localization in multiple testsystems.
在过去几年中,对电力系统中令人头疼的振荡源(尤其是强迫振荡 (FOs))进行定位受到了广泛关注。其部分原因是大规模部署了相位测量单元(PMU),以便在振荡发生时捕捉到这些振荡;另一部分原因是由于组件故障、风电波动和/或周期性负载导致的 FOs 事件日益增多。利用 [1] 的分频公式,我们开发了利用总线频率测量来定位单振荡源和多振荡源的方法。定位单个振荡源的方法不需要网络参数知识,但定位由多个振荡源引起的 FOs 的方法需要这一知识。我们将解释这种知识差异背后的原因,并展示我们在多个测试系统中成功定位振荡源的方法。
{"title":"Localizing Single and Multiple Oscillatory Sources: A Frequency Divider Approach","authors":"Rajasekhar Anguluri, Anamitra Pal","doi":"arxiv-2409.00566","DOIUrl":"https://doi.org/arxiv-2409.00566","url":null,"abstract":"Localizing sources of troublesome oscillations, particularly forced\u0000oscillations (FOs), in power systems has received considerable attention over\u0000the last few years. This is driven in part by the massive deployment of phasor\u0000measurement units (PMUs) that capture these oscillations when they occur; and\u0000in part by the increasing incidents of FOs due to malfunctioning components,\u0000wind power fluctuations, and/or cyclic loads. Capitalizing on the frequency\u0000divider formula of [1], we develop methods to localize single and multiple\u0000oscillatory sources using bus frequency measurements. The method to localize a\u0000single oscillation source does not require knowledge of network parameters.\u0000However, the method for localizing FOs caused by multiple sources requires this\u0000knowledge. We explain the reasoning behind this knowledge difference as well as\u0000demonstrate the success of our methods for source localization in multiple test\u0000systems.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"65 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Statistics of punctuation in experimental literature -- the remarkable case of "Finnegans Wake" by James Joyce 实验文学中的标点符号统计--詹姆斯-乔伊斯《芬尼根的守灵夜》的非凡案例
Pub Date : 2024-08-31 DOI: arxiv-2409.00483
Tomasz Stanisz, Stanisław Drożdż, Jarosław Kwapień
As the recent studies indicate, the structure imposed onto written texts bythe presence of punctuation develops patterns which reveal certaincharacteristics of universality. In particular, based on a large collection ofclassic literary works, it has been evidenced that the distances betweenconsecutive punctuation marks, measured in terms of the number of words, obeythe discrete Weibull distribution - a discrete variant of a distribution oftenused in survival analysis. The present work extends the analysis of punctuationusage patterns to more experimental pieces of world literature. It turns outthat the compliance of the the distances between punctuation marks with thediscrete Weibull distribution typically applies here as well. However, some ofthe works by James Joyce are distinct in this regard - in the sense that thetails of the relevant distributions are significantly thicker and,consequently, the corresponding hazard functions are decreasing functions notobserved in typical literary texts in prose. "Finnegans Wake" - the same one towhich science owes the word "quarks" for the most fundamental constituents ofmatter - is particularly striking in this context. At the same time, in all thestudied texts, the sentence lengths - representing the distances betweensentence-ending punctuation marks - reveal more freedom and are not constrainedby the discrete Weibull distribution. This freedom in some cases translatesinto long-range nonlinear correlations, which manifest themselves inmultifractality. Again, a text particularly spectacular in terms ofmultifractality is "Finnegans Wake".
最近的研究表明,标点符号的存在给书面文本带来的结构模式显示出某些普遍性特征。特别是,在大量经典文学作品的基础上,研究证明,以字数衡量的连续标点符号之间的距离服从离散的威布尔分布--这是生存分析中常用的分布的离散变体。本研究将标点符号使用模式分析扩展到更多的世界文学实验作品中。事实证明,标点符号之间的距离符合离散韦布尔分布的情况在这里也同样适用。然而,詹姆斯-乔伊斯的一些作品在这方面却与众不同--从这个意义上说,相关分布的尾部明显变粗,因此,相应的危险函数是递减函数,这在典型的散文文学文本中是看不到的。"在这种情况下,《芬尼根的守灵夜》--科学界将物质的最基本成分 "夸克 "一词归功于这本书--尤其引人注目。同时,在所有研究文本中,句子长度--代表句末标点符号之间的距离--显示出更大的自由度,不受离散韦布尔分布的限制。在某些情况下,这种自由度转化为长程非线性相关性,表现为多折射性。同样,《芬尼根的守灵夜》也是在多重折射性方面尤为突出的文本。
{"title":"Statistics of punctuation in experimental literature -- the remarkable case of \"Finnegans Wake\" by James Joyce","authors":"Tomasz Stanisz, Stanisław Drożdż, Jarosław Kwapień","doi":"arxiv-2409.00483","DOIUrl":"https://doi.org/arxiv-2409.00483","url":null,"abstract":"As the recent studies indicate, the structure imposed onto written texts by\u0000the presence of punctuation develops patterns which reveal certain\u0000characteristics of universality. In particular, based on a large collection of\u0000classic literary works, it has been evidenced that the distances between\u0000consecutive punctuation marks, measured in terms of the number of words, obey\u0000the discrete Weibull distribution - a discrete variant of a distribution often\u0000used in survival analysis. The present work extends the analysis of punctuation\u0000usage patterns to more experimental pieces of world literature. It turns out\u0000that the compliance of the the distances between punctuation marks with the\u0000discrete Weibull distribution typically applies here as well. However, some of\u0000the works by James Joyce are distinct in this regard - in the sense that the\u0000tails of the relevant distributions are significantly thicker and,\u0000consequently, the corresponding hazard functions are decreasing functions not\u0000observed in typical literary texts in prose. \"Finnegans Wake\" - the same one to\u0000which science owes the word \"quarks\" for the most fundamental constituents of\u0000matter - is particularly striking in this context. At the same time, in all the\u0000studied texts, the sentence lengths - representing the distances between\u0000sentence-ending punctuation marks - reveal more freedom and are not constrained\u0000by the discrete Weibull distribution. This freedom in some cases translates\u0000into long-range nonlinear correlations, which manifest themselves in\u0000multifractality. Again, a text particularly spectacular in terms of\u0000multifractality is \"Finnegans Wake\".","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"44 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Personalized Pricing Decisions Through Adversarial Risk Analysis 通过对抗性风险分析做出个性化定价决策
Pub Date : 2024-08-31 DOI: arxiv-2409.00444
Daniel García Rasines, Roi Naveiro, David Ríos Insua, Simón Rodríguez Santana
Pricing decisions stand out as one of the most critical tasks a companyfaces, particularly in today's digital economy. As with other businessdecision-making problems, pricing unfolds in a highly competitive and uncertainenvironment. Traditional analyses in this area have heavily relied on gametheory and its variants. However, an important drawback of these approaches istheir reliance on common knowledge assumptions, which are hardly tenable incompetitive business domains. This paper introduces an innovative personalizedpricing framework designed to assist decision-makers in undertaking pricingdecisions amidst competition, considering both buyer's and competitors'preferences. Our approach (i) establishes a coherent framework for modelingcompetition mitigating common knowledge assumptions; (ii) proposes a principledmethod to forecast competitors' pricing and customers' purchasing decisions,acknowledging major business uncertainties; and, (iii) encourages structuredthinking about the competitors' problems, thus enriching the solution process.To illustrate these properties, in addition to a general pricing template, weoutline two specifications - one from the retail domain and a more intricateone from the pension fund domain.
定价决策是公司面临的最关键任务之一,尤其是在当今的数字经济时代。与其他商业决策问题一样,定价也是在高度竞争和不确定的环境中展开的。这方面的传统分析在很大程度上依赖于博弈论及其变体。然而,这些方法的一个重要缺点是依赖于常识假设,而这些常识假设在竞争激烈的商业领域很难站得住脚。本文介绍了一种创新的个性化定价框架,旨在帮助决策者在竞争中做出定价决策,同时考虑买方和竞争对手的偏好。我们的方法(i) 建立了一个连贯的竞争建模框架,减少了常识假设;(ii) 提出了一种有原则的方法来预测竞争对手的定价和客户的购买决策,承认了主要的商业不确定性;(iii) 鼓励对竞争对手的问题进行结构化思考,从而丰富了解决方案的过程。为了说明这些特性,除了一般的定价模板外,我们还概述了两个规范--一个来自零售领域,另一个来自养老基金领域。
{"title":"Personalized Pricing Decisions Through Adversarial Risk Analysis","authors":"Daniel García Rasines, Roi Naveiro, David Ríos Insua, Simón Rodríguez Santana","doi":"arxiv-2409.00444","DOIUrl":"https://doi.org/arxiv-2409.00444","url":null,"abstract":"Pricing decisions stand out as one of the most critical tasks a company\u0000faces, particularly in today's digital economy. As with other business\u0000decision-making problems, pricing unfolds in a highly competitive and uncertain\u0000environment. Traditional analyses in this area have heavily relied on game\u0000theory and its variants. However, an important drawback of these approaches is\u0000their reliance on common knowledge assumptions, which are hardly tenable in\u0000competitive business domains. This paper introduces an innovative personalized\u0000pricing framework designed to assist decision-makers in undertaking pricing\u0000decisions amidst competition, considering both buyer's and competitors'\u0000preferences. Our approach (i) establishes a coherent framework for modeling\u0000competition mitigating common knowledge assumptions; (ii) proposes a principled\u0000method to forecast competitors' pricing and customers' purchasing decisions,\u0000acknowledging major business uncertainties; and, (iii) encourages structured\u0000thinking about the competitors' problems, thus enriching the solution process.\u0000To illustrate these properties, in addition to a general pricing template, we\u0000outline two specifications - one from the retail domain and a more intricate\u0000one from the pension fund domain.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distribution-Based Sub-Population Selection (DSPS): A Method for in-Silico Reproduction of Clinical Trials Outcomes 基于分布的亚群选择(DSPS):模拟再现临床试验结果的方法
Pub Date : 2024-08-30 DOI: arxiv-2409.00232
Mohammadreza Ganji, Anas El Fathi Ph. D., Chiara Fabris Ph. D., Dayu Lv Ph. D., Boris Kovatchev Ph. D., Marc Breton Ph. D
Background and Objective: Diabetes presents a significant challenge tohealthcare due to the negative impact of poor blood sugar control on health andassociated complications. Computer simulation platforms, notably exemplified bythe UVA/Padova Type 1 Diabetes simulator, has emerged as a promising tool foradvancing diabetes treatments by simulating patient responses in a virtualenvironment. The UVA Virtual Lab (UVLab) is a new simulation platform to mimicthe metabolic behavior of people with Type 2 diabetes (T2D) with a largepopulation of 6062 virtual subjects. Methods: The work introduces theDistribution-Based Population Selection (DSPS) method, a systematic approach toidentifying virtual subsets that mimic the clinical behavior observed in realtrials. The method transforms the sub-population selection task into a LinearPrograming problem, enabling the identification of the largest representativevirtual cohort. This selection process centers on key clinical outcomes indiabetes research, such as HbA1c and Fasting plasma Glucose (FPG), ensuringthat the statistical properties (moments) of the selected virtualsub-population closely resemble those observed in real-word clinical trial.Results: DSPS method was applied to the insulin degludec (IDeg) arm of a phase3 clinical trial. This method was used to select a sub-population of virtualsubjects that closely mirrored the clinical trial data across multiple keymetrics, including glycemic efficacy, insulin dosages, and cumulativehypoglycemia events over a 26-week period. Conclusion: The DSPS algorithm isable to select virtual sub-population within UVLab to reproduce and predict theoutcomes of a clinical trial. This statistical method can bridge the gapbetween large population simulation platforms and previously conducted clinicaltrials.
背景和目的:由于血糖控制不佳对健康和相关并发症造成的负面影响,糖尿病给医疗保健带来了巨大挑战。计算机模拟平台,特别是 UVA/Padova 1 型糖尿病模拟器,通过在虚拟环境中模拟病人的反应,已成为促进糖尿病治疗的一种有前途的工具。UVA 虚拟实验室(UVLab)是一个新的模拟平台,可模拟 6062 名虚拟受试者的 2 型糖尿病(T2D)患者的代谢行为。方法:这项工作引入了基于分布的人群选择(Distribution-Based Population Selection,DSPS)方法,这是一种识别虚拟子集的系统方法,可模仿真实试验中观察到的临床行为。该方法将子人群选择任务转化为线性编程问题,从而能够识别出最具代表性的虚拟人群。这一选择过程以糖尿病研究的关键临床结果(如 HbA1c 和空腹血浆葡萄糖 (FPG))为中心,确保所选虚拟子群的统计属性(矩)与在真实临床试验中观察到的属性(矩)非常相似:DSPS 方法适用于一项三期临床试验的胰岛素降糖(IDeg)治疗组。结果:将 DSPS 方法应用于胰岛素去势(IDeg)第三阶段临床试验中,结果显示所选虚拟受试者子群在多个关键指标(包括血糖疗效、胰岛素用量和 26 周内的累积血糖事件)上与临床试验数据密切相关。结论DSPS 算法能够在 UVLab 中选择虚拟子群,以重现和预测临床试验的结果。这种统计方法可以弥补大型人群模拟平台与之前进行的临床试验之间的差距。
{"title":"Distribution-Based Sub-Population Selection (DSPS): A Method for in-Silico Reproduction of Clinical Trials Outcomes","authors":"Mohammadreza Ganji, Anas El Fathi Ph. D., Chiara Fabris Ph. D., Dayu Lv Ph. D., Boris Kovatchev Ph. D., Marc Breton Ph. D","doi":"arxiv-2409.00232","DOIUrl":"https://doi.org/arxiv-2409.00232","url":null,"abstract":"Background and Objective: Diabetes presents a significant challenge to\u0000healthcare due to the negative impact of poor blood sugar control on health and\u0000associated complications. Computer simulation platforms, notably exemplified by\u0000the UVA/Padova Type 1 Diabetes simulator, has emerged as a promising tool for\u0000advancing diabetes treatments by simulating patient responses in a virtual\u0000environment. The UVA Virtual Lab (UVLab) is a new simulation platform to mimic\u0000the metabolic behavior of people with Type 2 diabetes (T2D) with a large\u0000population of 6062 virtual subjects. Methods: The work introduces the\u0000Distribution-Based Population Selection (DSPS) method, a systematic approach to\u0000identifying virtual subsets that mimic the clinical behavior observed in real\u0000trials. The method transforms the sub-population selection task into a Linear\u0000Programing problem, enabling the identification of the largest representative\u0000virtual cohort. This selection process centers on key clinical outcomes in\u0000diabetes research, such as HbA1c and Fasting plasma Glucose (FPG), ensuring\u0000that the statistical properties (moments) of the selected virtual\u0000sub-population closely resemble those observed in real-word clinical trial.\u0000Results: DSPS method was applied to the insulin degludec (IDeg) arm of a phase\u00003 clinical trial. This method was used to select a sub-population of virtual\u0000subjects that closely mirrored the clinical trial data across multiple key\u0000metrics, including glycemic efficacy, insulin dosages, and cumulative\u0000hypoglycemia events over a 26-week period. Conclusion: The DSPS algorithm is\u0000able to select virtual sub-population within UVLab to reproduce and predict the\u0000outcomes of a clinical trial. This statistical method can bridge the gap\u0000between large population simulation platforms and previously conducted clinical\u0000trials.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"160 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Framework for Digital Asset Risks with Insurance Applications 具有保险应用价值的数字资产风险框架
Pub Date : 2024-08-30 DOI: arxiv-2408.17227
Zhengming Li, Jianxi Su, Maochao Xu, Jimmy Yuen
The remarkable growth of digital assets, starting from the inception ofBitcoin in 2009 into a 1 trillion market in 2024, underscores the momentumbehind disruptive technologies and the global appetite for digital assets. Thispaper develops a framework to enhance actuaries' understanding of the cyberrisks associated with the developing digital asset ecosystem, as well as theirmeasurement methods in the context of digital asset insurance. By integratingactuarial perspectives, we aim to enhance understanding and modeling of cyberrisks at both the micro and systemic levels. The qualitative examination shedslight on blockchain technology and its associated risks, while our quantitativeframework offers a rigorous approach to modeling cyber risks in digital assetinsurance portfolios. This multifaceted approach serves three primaryobjectives: i) offer a clear and accessible education on the evolving digitalasset ecosystem and the diverse spectrum of cyber risks it entails; ii) developa scientifically rigorous framework for quantifying cyber risks in the digitalasset ecosystem; iii) provide practical applications, including pricingstrategies and tail risk management. Particularly, we developfrequency-severity models based on real loss data for pricing cyber risks indigit assets and utilize Monte Carlo simulation to estimate the tail risks,offering practical insights for risk management strategies. As digital assetscontinue to reshape finance, our work serves as a foundational step towardssafeguarding the integrity and stability of this rapidly evolving landscape.
从 2009 年比特币的诞生到 2024 年 1 万亿的市场规模,数字资产的显著增长凸显了颠覆性技术背后的动力和全球对数字资产的需求。本文制定了一个框架,以加强精算师对与发展中的数字资产生态系统相关的网络风险的理解,以及在数字资产保险中对其进行衡量的方法。通过整合精算视角,我们旨在加强对微观和系统层面网络风险的理解和建模。定性研究揭示了区块链技术及其相关风险,而我们的定量框架则为数字资产保险组合中的网络风险建模提供了严格的方法。这种多层面的方法有三个主要目标:i) 提供清晰易懂的教育,介绍不断发展的数字资产生态系统及其带来的各种网络风险;ii) 制定科学严谨的框架,量化数字资产生态系统中的网络风险;iii) 提供实际应用,包括定价策略和尾端风险管理。特别是,我们开发了基于真实损失数据的频率-严重性模型,用于对数字资产中的网络风险进行定价,并利用蒙特卡罗模拟来估算尾部风险,为风险管理策略提供了实用见解。随着数字资产不断重塑金融业,我们的工作将成为保障这一快速发展领域的完整性和稳定性的基础性步骤。
{"title":"A Framework for Digital Asset Risks with Insurance Applications","authors":"Zhengming Li, Jianxi Su, Maochao Xu, Jimmy Yuen","doi":"arxiv-2408.17227","DOIUrl":"https://doi.org/arxiv-2408.17227","url":null,"abstract":"The remarkable growth of digital assets, starting from the inception of\u0000Bitcoin in 2009 into a 1 trillion market in 2024, underscores the momentum\u0000behind disruptive technologies and the global appetite for digital assets. This\u0000paper develops a framework to enhance actuaries' understanding of the cyber\u0000risks associated with the developing digital asset ecosystem, as well as their\u0000measurement methods in the context of digital asset insurance. By integrating\u0000actuarial perspectives, we aim to enhance understanding and modeling of cyber\u0000risks at both the micro and systemic levels. The qualitative examination sheds\u0000light on blockchain technology and its associated risks, while our quantitative\u0000framework offers a rigorous approach to modeling cyber risks in digital asset\u0000insurance portfolios. This multifaceted approach serves three primary\u0000objectives: i) offer a clear and accessible education on the evolving digital\u0000asset ecosystem and the diverse spectrum of cyber risks it entails; ii) develop\u0000a scientifically rigorous framework for quantifying cyber risks in the digital\u0000asset ecosystem; iii) provide practical applications, including pricing\u0000strategies and tail risk management. Particularly, we develop\u0000frequency-severity models based on real loss data for pricing cyber risks in\u0000digit assets and utilize Monte Carlo simulation to estimate the tail risks,\u0000offering practical insights for risk management strategies. As digital assets\u0000continue to reshape finance, our work serves as a foundational step towards\u0000safeguarding the integrity and stability of this rapidly evolving landscape.","PeriodicalId":501172,"journal":{"name":"arXiv - STAT - Applications","volume":"33 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142187907","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - STAT - Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1