首页 > 最新文献

arXiv - STAT - Other Statistics最新文献

英文 中文
Guidelines and Best Practices to Share Deidentified Data and Code 共享去标识化数据和代码的指导原则和最佳做法
Pub Date : 2024-05-28 DOI: arxiv-2405.18232
Nicholas J. Horton, Sara Stoudt
In 2022, the Journal of Statistics and Data Science Education (JSDSE)instituted augmented requirements for authors to post deidentified data andcode underlying their papers. These changes were prompted by an increased focuson reproducibility and open science (NASEM 2019). A recent review of dataavailability practices noted that "such policies help increase thereproducibility of the published literature, as well as make a larger body ofdata available for reuse and re-analysis" (PLOS ONE, 2024). JSDSE valuesaccessibility as it endeavors to share knowledge that can improve educationalapproaches to teaching statistics and data science. Because institution,environment, and students differ across readers of the journal, it isespecially important to facilitate the transfer of a journal article's findingsto new contexts. This process may require digging into more of the details,including the deidentified data and code. Our goal is to provide our readersand authors with a review of why the requirements for code and data sharingwere instituted, summarize ongoing trends and developments in open science,discuss options for data and code sharing, and share advice for authors.
2022 年,《统计与数据科学教育期刊》(JSDSE)加强了对作者发布其论文所依据的去标识化数据和代码的要求。这些变化是由于人们越来越关注可重复性和开放科学(NASEM,2019 年)。最近对数据可获取性实践的审查指出,"此类政策有助于提高已发表文献的可再现性,并使更多的数据可用于再利用和再分析"(PLOS ONE,2024 年)。JSDSE 重视数据的可获取性,因为它致力于分享知识,从而改进统计学和数据科学的教学方法。由于期刊读者所处的机构、环境和学生各不相同,因此促进期刊文章的研究成果在新环境中的转化尤为重要。这一过程可能需要挖掘更多细节,包括去标识化的数据和代码。我们的目标是为我们的读者和作者回顾为什么要制定代码和数据共享的要求,总结开放科学的趋势和发展,讨论数据和代码共享的选择,并分享给作者的建议。
{"title":"Guidelines and Best Practices to Share Deidentified Data and Code","authors":"Nicholas J. Horton, Sara Stoudt","doi":"arxiv-2405.18232","DOIUrl":"https://doi.org/arxiv-2405.18232","url":null,"abstract":"In 2022, the Journal of Statistics and Data Science Education (JSDSE)\u0000instituted augmented requirements for authors to post deidentified data and\u0000code underlying their papers. These changes were prompted by an increased focus\u0000on reproducibility and open science (NASEM 2019). A recent review of data\u0000availability practices noted that \"such policies help increase the\u0000reproducibility of the published literature, as well as make a larger body of\u0000data available for reuse and re-analysis\" (PLOS ONE, 2024). JSDSE values\u0000accessibility as it endeavors to share knowledge that can improve educational\u0000approaches to teaching statistics and data science. Because institution,\u0000environment, and students differ across readers of the journal, it is\u0000especially important to facilitate the transfer of a journal article's findings\u0000to new contexts. This process may require digging into more of the details,\u0000including the deidentified data and code. Our goal is to provide our readers\u0000and authors with a review of why the requirements for code and data sharing\u0000were instituted, summarize ongoing trends and developments in open science,\u0000discuss options for data and code sharing, and share advice for authors.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141171713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Epistemology behind Covariate Adjustment 变量调整背后的认识论
Pub Date : 2024-05-27 DOI: arxiv-2405.17224
Grayson L. Baird, Stephen L. Bieber
It is often asserted that to control for the effects of confounders, oneshould include the confounding variables of concern in a statistical model as acovariate. Conversely, it is also asserted that control can only be concludedby design, where the results from an analysis can only be interpreted asevidence of an effect because the design controlled for the cause. To suggestotherwise is said to be a fallacy of cum hoc ergo propter hoc. Obviously, thesetwo assertions create a conundrum: How can the effect of confounder becontrolled for with analysis instead of by design without committing cum hocergo propter hoc? The present manuscript answers this conundrum.
人们通常认为,要控制混杂因素的影响,就应在统计模型中将相关的混杂变量作为一个变量。反之,也有人断言,只有通过设计才能得出控制的结论,即由于设计控制了原因,分析结果只能被解释为效果的证据。反之,则是 "既成事实"(cum hoc ergo propter hoc)的谬误。很明显,这两个论断造成了一个难题:如何通过分析而不是设计来控制混杂因素的影响,而又不犯兼有因果关系的谬误?本手稿回答了这一难题。
{"title":"The Epistemology behind Covariate Adjustment","authors":"Grayson L. Baird, Stephen L. Bieber","doi":"arxiv-2405.17224","DOIUrl":"https://doi.org/arxiv-2405.17224","url":null,"abstract":"It is often asserted that to control for the effects of confounders, one\u0000should include the confounding variables of concern in a statistical model as a\u0000covariate. Conversely, it is also asserted that control can only be concluded\u0000by design, where the results from an analysis can only be interpreted as\u0000evidence of an effect because the design controlled for the cause. To suggest\u0000otherwise is said to be a fallacy of cum hoc ergo propter hoc. Obviously, these\u0000two assertions create a conundrum: How can the effect of confounder be\u0000controlled for with analysis instead of by design without committing cum hoc\u0000ergo propter hoc? The present manuscript answers this conundrum.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141173143","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Logic of Counterfactuals and the Epistemology of Causal Inference 反事实逻辑与因果推理认识论
Pub Date : 2024-05-18 DOI: arxiv-2405.11284
Hanti Lin
The 2021 Nobel Prize in Economics recognized a theory of causal inference,which deserves more attention from philosophers. To that end, I develop adialectic that extends the Lewis-Stalnaker debate on a logical principle calledConditional Excluded Middle (CEM). I first play the good cop for CEM, and givea new argument for it: a Quine-Putnam indispensability argument based on theNobel-Prize winning theory. But then I switch sides and play the bad cop: Iundermine that argument with a new theory of causal inference that preservesthe success of the original theory but dispenses with CEM.
2021 年诺贝尔经济学奖认可了一种因果推理理论,该理论值得哲学家们更多关注。为此,我提出了一个辩证法,扩展了刘易斯-斯塔尔纳克关于一个名为 "条件排除中间"(CEM)的逻辑原则的辩论。我首先为 CEM 扮演了一个好警察的角色,并为它提供了一个新的论证:一个基于诺贝尔奖获奖理论的奎因-普特南不可或缺性论证。但随后,我又换了一边,扮演了坏警察的角色:我用一种新的因果推理理论来破坏这一论证,这种理论保留了原有理论的成功之处,但却摒弃了CEM。
{"title":"The Logic of Counterfactuals and the Epistemology of Causal Inference","authors":"Hanti Lin","doi":"arxiv-2405.11284","DOIUrl":"https://doi.org/arxiv-2405.11284","url":null,"abstract":"The 2021 Nobel Prize in Economics recognized a theory of causal inference,\u0000which deserves more attention from philosophers. To that end, I develop a\u0000dialectic that extends the Lewis-Stalnaker debate on a logical principle called\u0000Conditional Excluded Middle (CEM). I first play the good cop for CEM, and give\u0000a new argument for it: a Quine-Putnam indispensability argument based on the\u0000Nobel-Prize winning theory. But then I switch sides and play the bad cop: I\u0000undermine that argument with a new theory of causal inference that preserves\u0000the success of the original theory but dispenses with CEM.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141147813","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Expected Points Above Average: A Novel NBA Player Metric Based on Bayesian Hierarchical Modeling 平均预期得分:基于贝叶斯层次模型的新型 NBA 球员衡量标准
Pub Date : 2024-05-16 DOI: arxiv-2405.10453
Benjamin Williams, Erin M. Schliep, Bailey Fosdick, Ryan Elmore
Team and player evaluation in professional sport is extremely important giventhe financial implications of success/failure. It is especially critical toidentify and retain elite shooters in the National Basketball Association(NBA), one of the premier basketball leagues worldwide because the ultimategoal of the game is to score more points than one's opponent. To this end wepropose two novel basketball metrics: "expected points" for team-basedcomparisons and "expected points above average (EPAA)" as a player-evaluationtool. Both metrics leverage posterior samples from Bayesian hierarchicalmodeling framework to cluster teams and players based on their shootingpropensities and abilities. We illustrate the concepts for the top 100 shottakers over the last decade and offer our metric as an additional metric forevaluating players.
鉴于成败的经济影响,职业体育中的团队和球员评估极为重要。美国国家篮球协会(NBA)是世界上首屈一指的篮球联赛之一,因为比赛的终极目标就是比对手得到更多的分数,因此在该协会中识别和留住精英射手尤为重要。为此,我们提出了两个新颖的篮球指标:用于基于球队的比较的 "预期得分 "和作为球员评估工具的 "平均预期得分(EPAA)"。这两个指标都利用贝叶斯层次模型框架的后验样本,根据投篮命中率和能力对球队和球员进行分组。我们对过去十年中前 100 名投篮命中率的概念进行了说明,并将我们的指标作为评估球员的额外指标。
{"title":"Expected Points Above Average: A Novel NBA Player Metric Based on Bayesian Hierarchical Modeling","authors":"Benjamin Williams, Erin M. Schliep, Bailey Fosdick, Ryan Elmore","doi":"arxiv-2405.10453","DOIUrl":"https://doi.org/arxiv-2405.10453","url":null,"abstract":"Team and player evaluation in professional sport is extremely important given\u0000the financial implications of success/failure. It is especially critical to\u0000identify and retain elite shooters in the National Basketball Association\u0000(NBA), one of the premier basketball leagues worldwide because the ultimate\u0000goal of the game is to score more points than one's opponent. To this end we\u0000propose two novel basketball metrics: \"expected points\" for team-based\u0000comparisons and \"expected points above average (EPAA)\" as a player-evaluation\u0000tool. Both metrics leverage posterior samples from Bayesian hierarchical\u0000modeling framework to cluster teams and players based on their shooting\u0000propensities and abilities. We illustrate the concepts for the top 100 shot\u0000takers over the last decade and offer our metric as an additional metric for\u0000evaluating players.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141147846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identification of Single-Treatment Effects in Factorial Experiments 因子实验中单一处理效应的识别
Pub Date : 2024-05-16 DOI: arxiv-2405.09797
Guilherme Duarte
Despite their cost, randomized controlled trials (RCTs) are widely regardedas gold-standard evidence in disciplines ranging from social science tomedicine. In recent decades, researchers have increasingly sought to reduce theresource burden of repeated RCTs with factorial designs that simultaneouslytest multiple hypotheses, e.g. experiments that evaluate the effects of manymedications or products simultaneously. Here I show that when multipleinterventions are randomized in experiments, the effect any single interventionwould have outside the experimental setting is not identified absent heroicassumptions, even if otherwise perfectly realistic conditions are achieved.This happens because single-treatment effects involve a counterfactual worldwith a single focal intervention, allowing other variables to take theirnatural values (which may be confounded or modified by the focal intervention).In contrast, observational studies and factorial experiments provideinformation about potential-outcome distributions with zero and multipleinterventions, respectively. In this paper, I formalize sufficient conditionsfor the identifiability of those isolated quantities. I show that researcherswho rely on this type of design have to justify either linearity of functionalforms or -- in the nonparametric case -- specify with Directed Acyclic Graphshow variables are related in the real world. Finally, I develop nonparametricsharp bounds -- i.e., maximally informative best-/worst-case estimatesconsistent with limited RCT data -- that show when extrapolations about effectsigns are empirically justified. These new results are illustrated withsimulated data.
尽管成本高昂,随机对照试验(RCT)仍被广泛视为从社会科学到医学等各学科的黄金标准证据。近几十年来,研究人员越来越多地寻求通过因子设计来减轻重复 RCT 的资源负担,因子设计可同时测试多个假设,例如同时评估多种药物或产品效果的实验。这是因为单一治疗效果涉及一个具有单一重点干预措施的反事实世界,允许其他变量取其自然值(这些值可能被重点干预措施混淆或改变)。在本文中,我正式提出了这些孤立量可识别性的充分条件。我表明,依赖这类设计的研究人员必须证明函数形式的线性,或者--在非参数情况下--用有向无环图(Directed Acyclic Graph)说明变量在现实世界中的关系。最后,我提出了非参数锐界--即与有限的 RCT 数据相一致的信息量最大的最佳/最差情况估计值--表明何时对效应符号的推断在经验上是合理的。这些新结果用模拟数据进行了说明。
{"title":"Identification of Single-Treatment Effects in Factorial Experiments","authors":"Guilherme Duarte","doi":"arxiv-2405.09797","DOIUrl":"https://doi.org/arxiv-2405.09797","url":null,"abstract":"Despite their cost, randomized controlled trials (RCTs) are widely regarded\u0000as gold-standard evidence in disciplines ranging from social science to\u0000medicine. In recent decades, researchers have increasingly sought to reduce the\u0000resource burden of repeated RCTs with factorial designs that simultaneously\u0000test multiple hypotheses, e.g. experiments that evaluate the effects of many\u0000medications or products simultaneously. Here I show that when multiple\u0000interventions are randomized in experiments, the effect any single intervention\u0000would have outside the experimental setting is not identified absent heroic\u0000assumptions, even if otherwise perfectly realistic conditions are achieved.\u0000This happens because single-treatment effects involve a counterfactual world\u0000with a single focal intervention, allowing other variables to take their\u0000natural values (which may be confounded or modified by the focal intervention).\u0000In contrast, observational studies and factorial experiments provide\u0000information about potential-outcome distributions with zero and multiple\u0000interventions, respectively. In this paper, I formalize sufficient conditions\u0000for the identifiability of those isolated quantities. I show that researchers\u0000who rely on this type of design have to justify either linearity of functional\u0000forms or -- in the nonparametric case -- specify with Directed Acyclic Graphs\u0000how variables are related in the real world. Finally, I develop nonparametric\u0000sharp bounds -- i.e., maximally informative best-/worst-case estimates\u0000consistent with limited RCT data -- that show when extrapolations about effect\u0000signs are empirically justified. These new results are illustrated with\u0000simulated data.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141062584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sequential Maximal Updated Density Parameter Estimation for Dynamical Systems with Parameter Drift 有参数漂移的动态系统的序列最大更新密度参数估计
Pub Date : 2024-05-14 DOI: arxiv-2405.08307
Carlos del-Castillo-Negrete, Rylan Spence, Troy Butler, Clint Dawson
We present a novel method for generating sequential parameter estimates andquantifying epistemic uncertainty in dynamical systems within a data-consistent(DC) framework. The DC framework differs from traditional Bayesian approachesdue to the incorporation of the push-forward of an initial density, whichperforms selective regularization in parameter directions not informed by thedata in the resulting updated density. This extends a previous study thatincluded the linear Gaussian theory within the DC framework and introduced themaximal updated density (MUD) estimate as an alternative to both least squaresand maximum a posterior (MAP) estimates. In this work, we introduce algorithmsfor operational settings of MUD estimation in real or near-real time wherespatio-temporal datasets arrive in packets to provide updated estimates ofparameters and identify potential parameter drift. Computational diagnosticswithin the DC framework prove critical for evaluating (1) the quality of the DCupdate and MUD estimate and (2) the detection of parameter value drift. Thealgorithms are applied to estimate (1) wind drag parameters in a high-fidelitystorm surge model, (2) thermal diffusivity field for a heat conductivityproblem, and (3) changing infection and incubation rates of an epidemiologicalmodel.
我们提出了一种在数据一致(DC)框架内生成序列参数估计并量化动态系统中认识不确定性的新方法。数据一致性框架不同于传统的贝叶斯方法,因为它结合了初始密度的前推,在参数方向上进行选择性正则化,而在更新后的密度中,数据并未提供相关信息。这项研究扩展了之前的研究,将线性高斯理论纳入了 DC 框架,并引入了最大更新密度(MUD)估计,作为最小二乘法和最大后验(MAP)估计的替代方法。在这项工作中,我们介绍了 MUD 估计的实际或接近实时的操作设置算法,在这种情况下,空间-时间数据包以数据包的形式到达,以提供参数的更新估计并识别潜在的参数漂移。DC 框架内的计算诊断对于评估 (1) DC 更新和 MUD 估计的质量以及 (2) 参数值漂移的检测至关重要。这些算法被应用于估算:(1) 高保真风暴潮模型中的风阻参数;(2) 热传导问题中的热扩散场;(3) 流行病学模型中不断变化的感染率和潜伏率。
{"title":"Sequential Maximal Updated Density Parameter Estimation for Dynamical Systems with Parameter Drift","authors":"Carlos del-Castillo-Negrete, Rylan Spence, Troy Butler, Clint Dawson","doi":"arxiv-2405.08307","DOIUrl":"https://doi.org/arxiv-2405.08307","url":null,"abstract":"We present a novel method for generating sequential parameter estimates and\u0000quantifying epistemic uncertainty in dynamical systems within a data-consistent\u0000(DC) framework. The DC framework differs from traditional Bayesian approaches\u0000due to the incorporation of the push-forward of an initial density, which\u0000performs selective regularization in parameter directions not informed by the\u0000data in the resulting updated density. This extends a previous study that\u0000included the linear Gaussian theory within the DC framework and introduced the\u0000maximal updated density (MUD) estimate as an alternative to both least squares\u0000and maximum a posterior (MAP) estimates. In this work, we introduce algorithms\u0000for operational settings of MUD estimation in real or near-real time where\u0000spatio-temporal datasets arrive in packets to provide updated estimates of\u0000parameters and identify potential parameter drift. Computational diagnostics\u0000within the DC framework prove critical for evaluating (1) the quality of the DC\u0000update and MUD estimate and (2) the detection of parameter value drift. The\u0000algorithms are applied to estimate (1) wind drag parameters in a high-fidelity\u0000storm surge model, (2) thermal diffusivity field for a heat conductivity\u0000problem, and (3) changing infection and incubation rates of an epidemiological\u0000model.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140941397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting Short Response Ratings with Non-Content Related Features: A Hierarchical Modeling Approach 利用与内容无关的特征预测简短回复评分:分层建模方法
Pub Date : 2024-05-14 DOI: arxiv-2405.08574
Aubrey Condor
We explore whether the human ratings of open ended responses can be explainedwith non-content related features, and if such effects vary across differentmathematics-related items. When scoring is rigorously defined and rooted in ameasurement framework, educators intend that the features of a response whichare indicative of the respondent's level of ability are contributing to scores.However, we find that features such as response length, a grammar score of theresponse, and a metric relating to key phrase frequency are significantpredictors for response ratings. Although our findings are not causallyconclusive, they may propel us to be more critical of he way in which we assessopen ended responses, especially in high stakes scenarios. Educators take greatcare to provide unbiased, consistent ratings, but it may be that extraneousfeatures unrelated to those which were intended to be rated are beingevaluated.
我们探讨了人类对开放式作答的评分是否可以用与内容无关的特征来解释,以及这种影响在不同的数学相关项目中是否会有所不同。当评分被严格定义并植根于一个测量框架时,教育者希望能反映答题者能力水平的答题特征能对评分做出贡献。然而,我们发现,答题长度、答题语法得分以及与关键短语频率相关的指标等特征是答题评分的重要预测因素。尽管我们的研究结果并不具有因果关系,但它们可能会促使我们对评估开放式回答的方式更加挑剔,尤其是在高风险的情况下。教育工作者会非常谨慎地提供公正、一致的评分,但也有可能是那些与评分目的无关的无关特征被评估了。
{"title":"Predicting Short Response Ratings with Non-Content Related Features: A Hierarchical Modeling Approach","authors":"Aubrey Condor","doi":"arxiv-2405.08574","DOIUrl":"https://doi.org/arxiv-2405.08574","url":null,"abstract":"We explore whether the human ratings of open ended responses can be explained\u0000with non-content related features, and if such effects vary across different\u0000mathematics-related items. When scoring is rigorously defined and rooted in a\u0000measurement framework, educators intend that the features of a response which\u0000are indicative of the respondent's level of ability are contributing to scores.\u0000However, we find that features such as response length, a grammar score of the\u0000response, and a metric relating to key phrase frequency are significant\u0000predictors for response ratings. Although our findings are not causally\u0000conclusive, they may propel us to be more critical of he way in which we assess\u0000open ended responses, especially in high stakes scenarios. Educators take great\u0000care to provide unbiased, consistent ratings, but it may be that extraneous\u0000features unrelated to those which were intended to be rated are being\u0000evaluated.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140941398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Nested Instrumental Variables Design: Switcher Average Treatment Effect, Identification, Efficient Estimation and Generalizability 嵌套工具变量设计:转换者平均治疗效果、识别、有效估计和普适性
Pub Date : 2024-05-11 DOI: arxiv-2405.07102
Rui Wang, Ying-Qi Zhao, Oliver Dukes, Bo Zhang
Instrumental variables (IV) are a commonly used tool to estimate causaleffects from non-randomized data. A prototype of an IV is a randomized trialwith non-compliance where the randomized treatment assignment serves as an IVfor the non-ignorable treatment received. Under a monotonicity assumption, avalid IV non-parametrically identifies the average treatment effect among anon-identifiable complier subgroup, whose generalizability is often underdebate. In many studies, there could exist multiple versions of an IV, forinstance, different nudges to take the same treatment in different study sitesin a multi-center clinical trial. These different versions of an IV may resultin different compliance rates and offer a unique opportunity to study IVestimates' generalizability. In this article, we introduce a novel nested IVassumption and study identification of the average treatment effect among twolatent subgroups: always-compliers and switchers, who are defined based on thejoint potential treatment received under two versions of a binary IV. We derivethe efficient influence function for the SWitcher Average Treatment Effect(SWATE) and propose efficient estimators. We then propose formal statisticaltests of the generalizability of IV estimates based on comparing theconditional average treatment effect among the always-compliers and that amongthe switchers under the nested IV framework. We apply the proposed frameworkand method to the Prostate, Lung, Colorectal and Ovarian (PLCO) CancerScreening Trial and study the causal effect of colorectal cancer screening andits generalizability.
工具变量(IV)是从非随机数据中估计因果效应的常用工具。IV 的一个原型是具有非遵从性的随机试验,其中随机治疗分配可作为所接受的不可忽略的治疗的 IV。在单调性假设下,有效的 IV 可以非参数地识别不可识别的违规者亚群中的平均治疗效果,其普遍性往往受到争议。在许多研究中,可能存在多个版本的 IV,例如,在一项多中心临床试验中,不同的研究地点对采取相同治疗方法的不同劝告。这些不同版本的静脉注射可能会导致不同的依从率,为研究静脉注射估计值的可推广性提供了一个独特的机会。在本文中,我们引入了一个新颖的嵌套 IV 假设,并研究了在两类人群中平均治疗效果的识别问题:始终遵从者和转换者,这两类人群是根据二元 IV 的两个版本下共同接受的潜在治疗来定义的。我们推导出转换者平均治疗效果(SWATE)的有效影响函数,并提出了有效的估计值。然后,我们在比较嵌套 IV 框架下始终遵守者和转换者之间的条件平均治疗效果的基础上,对 IV 估计值的可推广性提出了正式的统计检验。我们将提出的框架和方法应用于前列腺癌、肺癌、结直肠癌和卵巢癌(PLCO)筛查试验,研究结直肠癌筛查的因果效应及其可推广性。
{"title":"Nested Instrumental Variables Design: Switcher Average Treatment Effect, Identification, Efficient Estimation and Generalizability","authors":"Rui Wang, Ying-Qi Zhao, Oliver Dukes, Bo Zhang","doi":"arxiv-2405.07102","DOIUrl":"https://doi.org/arxiv-2405.07102","url":null,"abstract":"Instrumental variables (IV) are a commonly used tool to estimate causal\u0000effects from non-randomized data. A prototype of an IV is a randomized trial\u0000with non-compliance where the randomized treatment assignment serves as an IV\u0000for the non-ignorable treatment received. Under a monotonicity assumption, a\u0000valid IV non-parametrically identifies the average treatment effect among a\u0000non-identifiable complier subgroup, whose generalizability is often under\u0000debate. In many studies, there could exist multiple versions of an IV, for\u0000instance, different nudges to take the same treatment in different study sites\u0000in a multi-center clinical trial. These different versions of an IV may result\u0000in different compliance rates and offer a unique opportunity to study IV\u0000estimates' generalizability. In this article, we introduce a novel nested IV\u0000assumption and study identification of the average treatment effect among two\u0000latent subgroups: always-compliers and switchers, who are defined based on the\u0000joint potential treatment received under two versions of a binary IV. We derive\u0000the efficient influence function for the SWitcher Average Treatment Effect\u0000(SWATE) and propose efficient estimators. We then propose formal statistical\u0000tests of the generalizability of IV estimates based on comparing the\u0000conditional average treatment effect among the always-compliers and that among\u0000the switchers under the nested IV framework. We apply the proposed framework\u0000and method to the Prostate, Lung, Colorectal and Ovarian (PLCO) Cancer\u0000Screening Trial and study the causal effect of colorectal cancer screening and\u0000its generalizability.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140941495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Strategies for Rare Population Detection and Sampling: A Methodological Approach in Liguria 稀有种群检测和采样策略:利古里亚的方法论途径
Pub Date : 2024-05-02 DOI: arxiv-2405.01342
G. Lancia, E. Riccomagno
Economic policy sciences are constantly investigating the quality ofwell-being of broad sections of the population in order to describe the currentinterdependence between unequal living conditions, low levels of education anda lack of integration into society. Such studies are often carried out in theform of surveys, e.g. as part of the EU-SILC program. If the survey is designedat national or international level, the results of the study are often used asa reference by a broad range of public institutions. However, the samplingstrategy per se may not capture enough information to provide an accuraterepresentation of all population strata. Problems might arise from rare, orhard-to-sample, populations and the conclusion of the study may be compromisedor unrealistic. We propose here a two-phase methodology to identify rare,poorly sampled populations and then resample the hard-to-sample strata. Wefocused our attention on the 2019 EU-SILC section concerning the Italian regionof Liguria. Methods based on dispersion indices or deep learning were used todetect rare populations. A multi-frame survey was proposed as the samplingdesign. The results showed that factors such as citizenship, materialdeprivation and large families are still fundamental characteristics that aredifficult to capture.
经济政策科学一直在调查广大人口的福利质量,以描述当前不平等的生活条件、低教育水平和缺乏社会融合之间的相互依存关系。此类研究通常以调查的形式进行,如作为欧盟--社会生活基础设施项目(EU-SILC)的一部分。如果调查是在国家或国际一级进行的,那么研究结果通常会被广泛的公共机构用作参考。然而,抽样策略本身可能无法获取足够的信息来准确代表所有人口阶层。罕见或难以取样的人群可能会出现问题,研究结论可能会受到影响或不切实际。在此,我们提出了一种分两个阶段进行的方法,以确定稀有、取样不足的种群,然后对难以取样的层进行重新取样。我们将注意力集中在 2019 年欧盟-SILC 有关意大利利古里亚地区的部分。我们使用了基于离散指数或深度学习的方法来检测稀有种群。我们建议采用多框架调查作为抽样设计。结果显示,公民身份、物质匮乏和大家庭等因素仍然是难以捕捉的基本特征。
{"title":"Strategies for Rare Population Detection and Sampling: A Methodological Approach in Liguria","authors":"G. Lancia, E. Riccomagno","doi":"arxiv-2405.01342","DOIUrl":"https://doi.org/arxiv-2405.01342","url":null,"abstract":"Economic policy sciences are constantly investigating the quality of\u0000well-being of broad sections of the population in order to describe the current\u0000interdependence between unequal living conditions, low levels of education and\u0000a lack of integration into society. Such studies are often carried out in the\u0000form of surveys, e.g. as part of the EU-SILC program. If the survey is designed\u0000at national or international level, the results of the study are often used as\u0000a reference by a broad range of public institutions. However, the sampling\u0000strategy per se may not capture enough information to provide an accurate\u0000representation of all population strata. Problems might arise from rare, or\u0000hard-to-sample, populations and the conclusion of the study may be compromised\u0000or unrealistic. We propose here a two-phase methodology to identify rare,\u0000poorly sampled populations and then resample the hard-to-sample strata. We\u0000focused our attention on the 2019 EU-SILC section concerning the Italian region\u0000of Liguria. Methods based on dispersion indices or deep learning were used to\u0000detect rare populations. A multi-frame survey was proposed as the sampling\u0000design. The results showed that factors such as citizenship, material\u0000deprivation and large families are still fundamental characteristics that are\u0000difficult to capture.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140828627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What's So Hard about the Monty Hall Problem? 蒙蒂-霍尔问题有什么难的?
Pub Date : 2024-05-01 DOI: arxiv-2405.00884
Rafael C. Alvarado
The Monty Hall problem is notorious for its deceptive simplicity. Althoughtoday it is widely used as a provocative thought experiment to introduceBayesian thinking to students of probability, in the not so distant past it wasrejected by established mathematicians. This essay provides some historicalbackground to the problem and explains why it is considered socounter-intuitive to many. It is argued that the main barrier to understandingthe problem is the back-grounding of the concept of dependence in probabilitytheory as it is commonly taught. To demonstrate this, a Bayesian solution isprovided and augmented with a probabilistic graphical model (PGM) inspired bythe work of Pearl (1988, 1998). Although the Bayesian approach produces thecorrect answer, without a representation of the dependency structure of eventsimplied by the problem, the salient fact that motivates the problem's solutionremains hidden.
蒙蒂-霍尔问题因其具有欺骗性的简单性而臭名昭著。尽管今天它被广泛用作向概率论学生介绍贝叶斯思想的启发性思想实验,但在不久的过去,它却被成熟的数学家所拒绝。本文介绍了这一问题的一些历史背景,并解释了为什么许多人认为它有悖于直觉。本文认为,理解这个问题的主要障碍在于概率论中依赖性概念的背景,因为概率论通常是这样讲的。为了证明这一点,我们提供了一种贝叶斯解法,并受 Pearl(1988 年,1998 年)的工作启发,使用概率图形模型 (PGM) 对其进行了补充。虽然贝叶斯方法得出了正确的答案,但由于没有问题所暗示的事件依赖结构的表示,促使问题解决的突出事实仍然被隐藏起来。
{"title":"What's So Hard about the Monty Hall Problem?","authors":"Rafael C. Alvarado","doi":"arxiv-2405.00884","DOIUrl":"https://doi.org/arxiv-2405.00884","url":null,"abstract":"The Monty Hall problem is notorious for its deceptive simplicity. Although\u0000today it is widely used as a provocative thought experiment to introduce\u0000Bayesian thinking to students of probability, in the not so distant past it was\u0000rejected by established mathematicians. This essay provides some historical\u0000background to the problem and explains why it is considered so\u0000counter-intuitive to many. It is argued that the main barrier to understanding\u0000the problem is the back-grounding of the concept of dependence in probability\u0000theory as it is commonly taught. To demonstrate this, a Bayesian solution is\u0000provided and augmented with a probabilistic graphical model (PGM) inspired by\u0000the work of Pearl (1988, 1998). Although the Bayesian approach produces the\u0000correct answer, without a representation of the dependency structure of events\u0000implied by the problem, the salient fact that motivates the problem's solution\u0000remains hidden.","PeriodicalId":501323,"journal":{"name":"arXiv - STAT - Other Statistics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140828628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - STAT - Other Statistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1