首页 > 最新文献

Statistics and Public Policy最新文献

英文 中文
Rejoinder to “Understanding our Markov Chain Significance Test” 对“理解我们的马尔可夫链显著性检验”的答复
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2019-01-01 DOI: 10.1080/2330443X.2019.1619427
Wendy K. Tam Cho, Simon Rubinstein-Salzedo
We thank Chikina, Frieze, and Pegden for their reply to our article. We offer just a short clarification rejoinder. In particular, we would like to be clear that we are not challenging the CFP test as a partisan gerrymandering test. We also do not “cast doubt” on the CFP paper. We have clearly stated that “we take no issues with the mathematics behind the CFP theorem or its proof.” In addition, we do not “prefer” one partisan gerrymandering test over another or advocate a single test. We firmly believe that there is plenty of room for multiple partisan gerrymandering tests. In this space, one test need not be “worse” than another. At the same time, it is indisputable that whether the CFP test would constitute a legal test for partisan gerrymandering is a legal question for the courts to decide. Legal questions cannot be decided by mathematicians. Mathematicians may make proposals, but judges decide whether to accept those proposals. Our point is simply that judges must clearly understand the mathematical concepts (even if not the mathematical details) in order to make a reasoned judgment. However, when the science is unclear, we have only miscommunication, from which no one benefits.
我们感谢中国、弗里兹和佩格登对我们文章的回复。我们只是提供一个简短的澄清反驳。特别是,我们想要明确的是,我们不是在挑战CFP测试作为一个党派的不公正划分选区的测试。我们也不“怀疑”CFP报告。我们已经明确表示,“我们对CFP定理背后的数学及其证明没有任何问题。”此外,我们不“偏爱”某一党派的不公正划分选区的测试,也不提倡单一的测试。我们坚信,有足够的空间进行多党不公正的选区划分测试。在这个空间里,一个测试不一定比另一个“更糟糕”。与此同时,CFP测试是否会构成对党派不公正划分选区的法律测试,这是法院必须决定的法律问题,这是无可争辩的。法律问题不能由数学家来决定。数学家可以提出建议,但由评委决定是否接受这些建议。我们的观点很简单,法官必须清楚地理解数学概念(即使不是数学细节),以便做出合理的判断。然而,当科学不明确时,我们只有误解,没有人从中受益。
{"title":"Rejoinder to “Understanding our Markov Chain Significance Test”","authors":"Wendy K. Tam Cho, Simon Rubinstein-Salzedo","doi":"10.1080/2330443X.2019.1619427","DOIUrl":"https://doi.org/10.1080/2330443X.2019.1619427","url":null,"abstract":"We thank Chikina, Frieze, and Pegden for their reply to our article. We offer just a short clarification rejoinder. In particular, we would like to be clear that we are not challenging the CFP test as a partisan gerrymandering test. We also do not “cast doubt” on the CFP paper. We have clearly stated that “we take no issues with the mathematics behind the CFP theorem or its proof.” In addition, we do not “prefer” one partisan gerrymandering test over another or advocate a single test. We firmly believe that there is plenty of room for multiple partisan gerrymandering tests. In this space, one test need not be “worse” than another. At the same time, it is indisputable that whether the CFP test would constitute a legal test for partisan gerrymandering is a legal question for the courts to decide. Legal questions cannot be decided by mathematicians. Mathematicians may make proposals, but judges decide whether to accept those proposals. Our point is simply that judges must clearly understand the mathematical concepts (even if not the mathematical details) in order to make a reasoned judgment. However, when the science is unclear, we have only miscommunication, from which no one benefits.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"6 1","pages":"54 - 54"},"PeriodicalIF":1.6,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2019.1619427","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48280021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
EPA is Mandating the Normal Distribution EPA正在强制执行正态分布
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2019-01-01 DOI: 10.1080/2330443X.2018.1564639
S. Millard
Abstract The United States Environmental Protection Agency (USEPA) is responsible for overseeing the cleanup of sites that fall within the jurisdiction of the Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA; also known as “Superfund”). This process almost always involves a remedial investigation/feasibility (RI/FS) study, including deriving upper confidence, prediction, and/or tolerance limits based on concentrations from a designated “background” area which are subsequently used to determine whether a remediated site has achieved compliance. Past USEPA guidance states outlying observations in the background data should not be removed based solely on statistical tests, but rather on some scientific or quality assurance basis. However, recent USEPA guidance states “extreme” outliers, based on tests that assume a normal (Gaussian) distribution, should always be removed from background data, and because “extreme” is not defined, USEPA has interpreted this to mean all outliers identified by a test should be removed. This article discusses problems with current USEPA guidance and how it contradicts past guidance, and illustrates USEPA’s current policy via a case study of the Portland, Oregon Harbor Superfund site. Additional materials, including R code, data, and documentation of correspondence are available in the online supplement.
摘要美国环境保护局(USEPA)负责监督《综合环境反应、赔偿和责任法》(CERCLA;也称为“超级基金”)管辖范围内的场地清理工作。该过程几乎总是涉及补救调查/可行性(RI/FS)研究,包括根据指定“背景”区域的浓度得出置信度、预测和/或公差上限,随后用于确定补救场地是否达到合规性。美国环保局过去的指导意见指出,背景数据中的异常观测不应仅基于统计测试,而应基于一些科学或质量保证基础。然而,美国环境保护局最近的指导意见指出,基于假设正态(高斯)分布的测试,“极端”异常值应始终从背景数据中删除,因为“极端”没有定义,美国环境管理局将其解释为应删除测试确定的所有异常值。本文讨论了美国环保局目前的指导意见存在的问题,以及它如何与过去的指导意见相矛盾,并通过俄勒冈州波特兰港超级基金网站的案例研究来说明美国环保局当前的政策。在线增刊中提供了其他材料,包括R代码、数据和通信文件。
{"title":"EPA is Mandating the Normal Distribution","authors":"S. Millard","doi":"10.1080/2330443X.2018.1564639","DOIUrl":"https://doi.org/10.1080/2330443X.2018.1564639","url":null,"abstract":"Abstract The United States Environmental Protection Agency (USEPA) is responsible for overseeing the cleanup of sites that fall within the jurisdiction of the Comprehensive Environmental Response, Compensation, and Liability Act (CERCLA; also known as “Superfund”). This process almost always involves a remedial investigation/feasibility (RI/FS) study, including deriving upper confidence, prediction, and/or tolerance limits based on concentrations from a designated “background” area which are subsequently used to determine whether a remediated site has achieved compliance. Past USEPA guidance states outlying observations in the background data should not be removed based solely on statistical tests, but rather on some scientific or quality assurance basis. However, recent USEPA guidance states “extreme” outliers, based on tests that assume a normal (Gaussian) distribution, should always be removed from background data, and because “extreme” is not defined, USEPA has interpreted this to mean all outliers identified by a test should be removed. This article discusses problems with current USEPA guidance and how it contradicts past guidance, and illustrates USEPA’s current policy via a case study of the Portland, Oregon Harbor Superfund site. Additional materials, including R code, data, and documentation of correspondence are available in the online supplement.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"6 1","pages":"36 - 43"},"PeriodicalIF":1.6,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2018.1564639","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47615775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
An Alternative to the Carnegie Classifications: Identifying Similar Doctoral Institutions With Structural Equation Models and Clustering 卡内基分类的另一种选择:用结构方程模型和聚类识别相似的博士机构
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2019-01-01 DOI: 10.1080/2330443x.2019.1666761
P. Harmon, Sarah M McKnight, L. Hildreth, I. Godwin, M. Greenwood
Abstract The Carnegie Classification of Institutions of Higher Education is a commonly used framework for institutional classification that classifies doctoral-granting schools into three groups based on research productivity. Despite its wide use, the Carnegie methodology involves several shortcomings, including a lack of thorough documentation, subjectively placed thresholds between institutions, and a methodology that is not completely reproducible. We describe the methodology of the 2015 and 2018 updates to the classification and propose an alternative method of classification using the same data that relies on structural equation modeling (SEM) of latent factors rather than principal component-based indices of productivity. In contrast to the Carnegie methodology, we use SEM to obtain a single factor score for each school based on latent metrics of research productivity. Classifications are then made using a univariate model-based clustering algorithm as opposed to subjective thresholding, as is done in the Carnegie methodology. Finally, we present a Shiny web application that demonstrates sensitivity of both the Carnegie Classification and SEM-based classification of a selected university and generates a table of peer institutions in line with the stated goals of the Carnegie Classification.
卡内基高等教育机构分类是一种常用的机构分类框架,它根据研究效率将博士学位授予学校分为三类。尽管卡内基方法被广泛使用,但它有几个缺点,包括缺乏彻底的文件,主观地在机构之间设置阈值,以及一种不完全可复制的方法。我们描述了2015年和2018年更新分类的方法,并提出了使用相同数据的另一种分类方法,该方法依赖于潜在因素的结构方程模型(SEM),而不是基于主成分的生产率指数。与卡内基方法相反,我们使用SEM根据研究生产力的潜在指标获得每所学校的单因素得分。然后使用基于单变量模型的聚类算法进行分类,而不是像卡内基方法那样使用主观阈值。最后,我们展示了一个Shiny的web应用程序,该应用程序展示了对选定大学的卡内基分类和基于sem的分类的敏感性,并生成了一个符合卡内基分类既定目标的同行机构表。
{"title":"An Alternative to the Carnegie Classifications: Identifying Similar Doctoral Institutions With Structural Equation Models and Clustering","authors":"P. Harmon, Sarah M McKnight, L. Hildreth, I. Godwin, M. Greenwood","doi":"10.1080/2330443x.2019.1666761","DOIUrl":"https://doi.org/10.1080/2330443x.2019.1666761","url":null,"abstract":"Abstract The Carnegie Classification of Institutions of Higher Education is a commonly used framework for institutional classification that classifies doctoral-granting schools into three groups based on research productivity. Despite its wide use, the Carnegie methodology involves several shortcomings, including a lack of thorough documentation, subjectively placed thresholds between institutions, and a methodology that is not completely reproducible. We describe the methodology of the 2015 and 2018 updates to the classification and propose an alternative method of classification using the same data that relies on structural equation modeling (SEM) of latent factors rather than principal component-based indices of productivity. In contrast to the Carnegie methodology, we use SEM to obtain a single factor score for each school based on latent metrics of research productivity. Classifications are then made using a univariate model-based clustering algorithm as opposed to subjective thresholding, as is done in the Carnegie methodology. Finally, we present a Shiny web application that demonstrates sensitivity of both the Carnegie Classification and SEM-based classification of a selected university and generates a table of peer institutions in line with the stated goals of the Carnegie Classification.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"6 1","pages":"87 - 97"},"PeriodicalIF":1.6,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443x.2019.1666761","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46951949","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Statistics, Probability, and a Failed Conservation Policy 统计、概率和失败的保护策略
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2019-01-01 DOI: 10.1080/2330443X.2019.1637802
Michael D. Collins
Abstract Many sightings of the Ivory-billed Woodpecker (Campephilus principalis) have been reported during the past several decades, but nobody has managed to obtain the clear photo that is regarded as the standard form of evidence for documenting birds. Despite reports of sightings by teams of ornithologists working independently in Arkansas and Florida, doubts cast on the persistence of this iconic species have impeded the establishment of a meaningful conservation program. An analysis of the expected waiting time for obtaining a photo provides insights into why the policy of insisting upon ideal evidence has failed for this species. Concepts in statistics and probability are used to analyze video footage that was obtained during encounters with birds that were identified in the field as Ivory-billed Woodpeckers. One of the videos shows a series of events that are consistent with that species and are believed to be inconsistent with every other species of the region. Another video shows a large bird in flight with the distinctive wing motion of a large woodpecker. Only two large woodpeckers occur in the region, and the flap rate is about ten standard deviations greater than the mean flap rate of the Pileated Woodpecker (Dryocopus pileatus). Supplemental materials for this article are available online.
在过去的几十年里,有许多关于象牙嘴啄木鸟(Campephilus principalis)的报道,但没有人能够获得被认为是记录鸟类的标准证据形式的清晰照片。尽管在阿肯色州和佛罗里达州独立工作的鸟类学家团队报告了目击事件,但对这种标志性物种持续存在的怀疑阻碍了有意义的保护计划的建立。对获得照片的预期等待时间的分析提供了对为什么坚持理想证据的政策在这个物种中失败的见解。统计和概率的概念被用来分析在遇到鸟类时获得的视频片段,这些鸟类在野外被识别为象牙嘴啄木鸟。其中一个视频显示了一系列与该物种一致的事件,据信与该地区其他物种不一致。另一段视频显示,一只大鸟在飞行中,翅膀的运动很像一只大型啄木鸟。该地区仅有2只大型啄木鸟,拍打率比啄木鸟(Dryocopus pileatus)的平均拍打率大10个标准差左右。本文的补充材料可在网上获得。
{"title":"Statistics, Probability, and a Failed Conservation Policy","authors":"Michael D. Collins","doi":"10.1080/2330443X.2019.1637802","DOIUrl":"https://doi.org/10.1080/2330443X.2019.1637802","url":null,"abstract":"Abstract Many sightings of the Ivory-billed Woodpecker (Campephilus principalis) have been reported during the past several decades, but nobody has managed to obtain the clear photo that is regarded as the standard form of evidence for documenting birds. Despite reports of sightings by teams of ornithologists working independently in Arkansas and Florida, doubts cast on the persistence of this iconic species have impeded the establishment of a meaningful conservation program. An analysis of the expected waiting time for obtaining a photo provides insights into why the policy of insisting upon ideal evidence has failed for this species. Concepts in statistics and probability are used to analyze video footage that was obtained during encounters with birds that were identified in the field as Ivory-billed Woodpeckers. One of the videos shows a series of events that are consistent with that species and are believed to be inconsistent with every other species of the region. Another video shows a large bird in flight with the distinctive wing motion of a large woodpecker. Only two large woodpeckers occur in the region, and the flap rate is about ten standard deviations greater than the mean flap rate of the Pileated Woodpecker (Dryocopus pileatus). Supplemental materials for this article are available online.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"6 1","pages":"67 - 79"},"PeriodicalIF":1.6,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2019.1637802","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"60065970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Quantifying Gerrymandering in North Carolina 量化北卡罗来纳州的不公正划分选区
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2018-01-10 DOI: 10.1080/2330443x.2020.1796400
G. Herschlag, H. Kang, Justin Luo, Christy V. Graves, Sachet Bangia, Robert J. Ravier, Jonathan C. Mattingly
ABSTRACT By comparing a specific redistricting plan to an ensemble of plans, we evaluate whether the plan translates individual votes to election outcomes in an unbiased fashion. Explicitly, we evaluate if a given redistricting plan exhibits extreme statistical properties compared to an ensemble of nonpartisan plans satisfying all legal criteria. Thus, we capture how unbiased redistricting plans interpret individual votes via a state’s geo-political landscape. We generate the ensemble of plans through a Markov chain Monte Carlo algorithm coupled with simulated annealing based on a reference distribution that does not include partisan criteria. Using the ensemble and historical voting data, we create a null hypothesis for various election results, free from partisanship, accounting for the state’s geo-politics. We showcase our methods on two recent congressional districting plans of NC, along with a plan drawn by a bipartisan panel of retired judges. We find the enacted plans are extreme outliers whereas the bipartisan judges’ plan does not give rise to extreme partisan outcomes. Equally important, we illuminate anomalous structures in the plans of interest by developing graphical representations which help identify and understand instances of cracking and packing associated with gerrymandering. These methods were successfully used in recent court cases. Supplementary materials for this article are available online.
摘要通过将一个特定的选区重新划分计划与一系列计划进行比较,我们评估该计划是否以公正的方式将个人选票转化为选举结果。明确地说,与满足所有法律标准的无党派计划相比,我们评估给定的重新划分计划是否表现出极端的统计特性。因此,我们可以捕捉到无偏见的选区重新划分计划是如何通过一个州的地缘政治景观来解释个人选票的。我们通过马尔可夫链蒙特卡罗算法,结合基于不包括党派标准的参考分布的模拟退火,生成计划集合。使用集合和历史投票数据,我们为各种选举结果创建了一个零假设,没有党派偏见,说明了该州的地缘政治。我们在北卡罗来纳州最近的两项国会选区划分计划中展示了我们的方法,以及一个由退休法官组成的两党小组制定的计划。我们发现,制定的计划是极端的局外人,而两党法官的计划并没有导致极端的党派结果。同样重要的是,我们通过开发图形表示来阐明感兴趣的计划中的异常结构,这些图形表示有助于识别和理解与选区划分不当相关的开裂和堆积实例。这些方法在最近的法庭案件中得到了成功的应用。本文的补充材料可在线获取。
{"title":"Quantifying Gerrymandering in North Carolina","authors":"G. Herschlag, H. Kang, Justin Luo, Christy V. Graves, Sachet Bangia, Robert J. Ravier, Jonathan C. Mattingly","doi":"10.1080/2330443x.2020.1796400","DOIUrl":"https://doi.org/10.1080/2330443x.2020.1796400","url":null,"abstract":"ABSTRACT By comparing a specific redistricting plan to an ensemble of plans, we evaluate whether the plan translates individual votes to election outcomes in an unbiased fashion. Explicitly, we evaluate if a given redistricting plan exhibits extreme statistical properties compared to an ensemble of nonpartisan plans satisfying all legal criteria. Thus, we capture how unbiased redistricting plans interpret individual votes via a state’s geo-political landscape. We generate the ensemble of plans through a Markov chain Monte Carlo algorithm coupled with simulated annealing based on a reference distribution that does not include partisan criteria. Using the ensemble and historical voting data, we create a null hypothesis for various election results, free from partisanship, accounting for the state’s geo-politics. We showcase our methods on two recent congressional districting plans of NC, along with a plan drawn by a bipartisan panel of retired judges. We find the enacted plans are extreme outliers whereas the bipartisan judges’ plan does not give rise to extreme partisan outcomes. Equally important, we illuminate anomalous structures in the plans of interest by developing graphical representations which help identify and understand instances of cracking and packing associated with gerrymandering. These methods were successfully used in recent court cases. Supplementary materials for this article are available online.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"7 1","pages":"30 - 38"},"PeriodicalIF":1.6,"publicationDate":"2018-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443x.2020.1796400","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42236973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 73
Inference of Long-Term Screening Outcomes for Individuals with Screening Histories 对有筛查史的个体的长期筛查结果的推断
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2018-01-01 DOI: 10.1080/2330443X.2018.1438939
Dongfeng Wu, K. Kafadar, S. Rai
ABSTRACT We develop a probability model for evaluating long-term outcomes due to regular screening that incorporates the effects of prior screening examinations. Previous models assume that individuals have no prior screening examinations at their current ages. Due to current widespread medical emphasis on screening, the consideration of screening histories is essential, particularly in assessing the benefit of future screening examinations given a certain number of previous negative screens. Screening participants are categorized into four mutually exclusive groups: symptom-free-life, no-early-detection, true-early-detection, and overdiagnosis. For each case, we develop models that incorporate a person’s current age, screening history, expected future screening frequency, screening test sensitivity, and other factors, and derive the probabilities of occurrence for the four groups. The probability of overdiagnosis among screen-detected cases is derived and estimated. The model applies to screening for any disease or condition; for concreteness, we focus on female breast cancer and use data from the study conducted by the Health Insurance Plan of Greater New York (HIP) to estimate these probabilities and corresponding credible intervals. The model can provide policy makers with important information regarding ranges of expected lives saved and percentages of true-early-detection and overdiagnosis among the screen-detected cases.
摘要:我们开发了一个概率模型,用于评估定期筛查的长期结果,该模型结合了先前筛查的影响。先前的模型假设个体在当前年龄没有进行过筛查。由于目前医学上普遍强调筛查,考虑筛查史是至关重要的,特别是在评估未来筛查的益处时,因为之前有一定数量的阴性筛查。筛查参与者被分为四个相互排斥的组:无症状生活、无早期检测、真正的早期检测和过度诊断。对于每种情况,我们都会开发模型,将一个人的当前年龄、筛查史、预期未来筛查频率、筛查测试敏感性和其他因素纳入其中,并推导出四组的发生概率。推导并估计了筛查发现病例中过度诊断的概率。该模型适用于任何疾病或状况的筛查;具体而言,我们关注女性乳腺癌癌症,并使用大纽约健康保险计划(HIP)进行的研究数据来估计这些概率和相应的可信区间。该模型可以为政策制定者提供重要信息,包括预期挽救的生命范围以及筛查发现的病例中真正的早期检测和过度诊断的百分比。
{"title":"Inference of Long-Term Screening Outcomes for Individuals with Screening Histories","authors":"Dongfeng Wu, K. Kafadar, S. Rai","doi":"10.1080/2330443X.2018.1438939","DOIUrl":"https://doi.org/10.1080/2330443X.2018.1438939","url":null,"abstract":"ABSTRACT We develop a probability model for evaluating long-term outcomes due to regular screening that incorporates the effects of prior screening examinations. Previous models assume that individuals have no prior screening examinations at their current ages. Due to current widespread medical emphasis on screening, the consideration of screening histories is essential, particularly in assessing the benefit of future screening examinations given a certain number of previous negative screens. Screening participants are categorized into four mutually exclusive groups: symptom-free-life, no-early-detection, true-early-detection, and overdiagnosis. For each case, we develop models that incorporate a person’s current age, screening history, expected future screening frequency, screening test sensitivity, and other factors, and derive the probabilities of occurrence for the four groups. The probability of overdiagnosis among screen-detected cases is derived and estimated. The model applies to screening for any disease or condition; for concreteness, we focus on female breast cancer and use data from the study conducted by the Health Insurance Plan of Greater New York (HIP) to estimate these probabilities and corresponding credible intervals. The model can provide policy makers with important information regarding ranges of expected lives saved and percentages of true-early-detection and overdiagnosis among the screen-detected cases.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":" ","pages":"1 - 10"},"PeriodicalIF":1.6,"publicationDate":"2018-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2018.1438939","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48631957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Value-Added and Student Growth Percentile Models: What Drives Differences in Estimated Classroom Effects? 增值和学生增长百分比模型:是什么导致了估计课堂效果的差异?
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2018-01-01 DOI: 10.1080/2330443X.2018.1438938
Michael Kurtz
ABSTRACT This study shows value-added models (VAM) and student growth percentile (SGP) models fundamentally disagree regarding estimated teacher effectiveness when the classroom distribution of test scores conditional on prior achievement is skewed (i.e., when a teacher serves a disproportionate number of high- or low-growth students). While conceptually similar, the two models differ in estimation method which can lead to sizable differences in estimated teacher effects. Moreover, the magnitude of conditional skewness needed to drive VAM and SGP models apart often by three and up to 6 deciles is within the ranges observed in actual data. The same teacher may appear weak using one model and strong with the other. Using a simulation, I evaluate the relationship under controllable conditions. I then verify that the results persist in observed student–teacher data from North Carolina.
摘要本研究表明,当以先前成绩为条件的考试成绩的课堂分布出现偏差时(即,当教师为不成比例的高增长或低增长学生服务时),增值模型(VAM)和学生增长百分位(SGP)模型在估计教师有效性方面存在根本分歧。虽然这两个模型在概念上相似,但估计方法不同,这可能导致估计的教师效应存在相当大的差异。此外,通常将VAM和SGP模型分开三个十分位数甚至六个十分位数所需的条件偏度大小在实际数据中观察到的范围内。同一位老师在使用一种模式时可能显得软弱,而在使用另一种模式后可能显得坚强。通过模拟,我评估了可控条件下的关系。然后,我在北卡罗来纳州观察到的学生和教师数据中验证了这一结果。
{"title":"Value-Added and Student Growth Percentile Models: What Drives Differences in Estimated Classroom Effects?","authors":"Michael Kurtz","doi":"10.1080/2330443X.2018.1438938","DOIUrl":"https://doi.org/10.1080/2330443X.2018.1438938","url":null,"abstract":"ABSTRACT This study shows value-added models (VAM) and student growth percentile (SGP) models fundamentally disagree regarding estimated teacher effectiveness when the classroom distribution of test scores conditional on prior achievement is skewed (i.e., when a teacher serves a disproportionate number of high- or low-growth students). While conceptually similar, the two models differ in estimation method which can lead to sizable differences in estimated teacher effects. Moreover, the magnitude of conditional skewness needed to drive VAM and SGP models apart often by three and up to 6 deciles is within the ranges observed in actual data. The same teacher may appear weak using one model and strong with the other. Using a simulation, I evaluate the relationship under controllable conditions. I then verify that the results persist in observed student–teacher data from North Carolina.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":" ","pages":"1 - 8"},"PeriodicalIF":1.6,"publicationDate":"2018-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2018.1438938","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47419044","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Does Predictive Policing Lead to Biased Arrests? Results From a Randomized Controlled Trial 预测性警务会导致有偏见的逮捕吗?随机对照试验的结果
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2018-01-01 DOI: 10.1080/2330443X.2018.1438940
P. Brantingham, Matthew A. Valasik, G. Mohler
ABSTRACT Racial bias in predictive policing algorithms has been the focus of a number of recent news articles, statements of concern by several national organizations (e.g., the ACLU and NAACP), and simulation-based research. There is reasonable concern that predictive algorithms encourage directed police patrols to target minority communities with discriminatory consequences for minority individuals. However, to date there have been no empirical studies on the bias of predictive algorithms used for police patrol. Here, we test for such biases using arrest data from the Los Angeles predictive policing experiments. We find that there were no significant differences in the proportion of arrests by racial-ethnic group between control and treatment conditions. We find that the total numbers of arrests at the division level declined or remained unchanged during predictive policing deployments. Arrests were numerically higher at the algorithmically predicted locations. When adjusted for the higher overall crime rate at algorithmically predicted locations, however, arrests were lower or unchanged.
预测性警务算法中的种族偏见一直是最近一些新闻文章的焦点,几个国家组织(例如,美国公民自由联盟和全国有色人种协进会)关注的声明,以及基于模拟的研究。人们有理由担心,预测算法会鼓励警察定向巡逻针对少数族裔社区,对少数族裔个人造成歧视性后果。然而,到目前为止,还没有关于用于警察巡逻的预测算法偏差的实证研究。在这里,我们使用来自洛杉矶预测性警务实验的逮捕数据来测试这种偏见。我们发现,在控制条件和治疗条件之间,种族-民族群体的逮捕比例没有显着差异。我们发现,在预测性警务部署期间,分区一级的逮捕总数下降或保持不变。在算法预测的地点,逮捕人数更高。然而,当根据算法预测的地点较高的总体犯罪率进行调整时,逮捕率较低或保持不变。
{"title":"Does Predictive Policing Lead to Biased Arrests? Results From a Randomized Controlled Trial","authors":"P. Brantingham, Matthew A. Valasik, G. Mohler","doi":"10.1080/2330443X.2018.1438940","DOIUrl":"https://doi.org/10.1080/2330443X.2018.1438940","url":null,"abstract":"ABSTRACT Racial bias in predictive policing algorithms has been the focus of a number of recent news articles, statements of concern by several national organizations (e.g., the ACLU and NAACP), and simulation-based research. There is reasonable concern that predictive algorithms encourage directed police patrols to target minority communities with discriminatory consequences for minority individuals. However, to date there have been no empirical studies on the bias of predictive algorithms used for police patrol. Here, we test for such biases using arrest data from the Los Angeles predictive policing experiments. We find that there were no significant differences in the proportion of arrests by racial-ethnic group between control and treatment conditions. We find that the total numbers of arrests at the division level declined or remained unchanged during predictive policing deployments. Arrests were numerically higher at the algorithmically predicted locations. When adjusted for the higher overall crime rate at algorithmically predicted locations, however, arrests were lower or unchanged.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"5 1","pages":"1 - 6"},"PeriodicalIF":1.6,"publicationDate":"2018-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2018.1438940","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46587771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 81
Examining the Carnegie Classification Methodology for Research Universities 研究型大学的卡内基分类方法研究
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2018-01-01 DOI: 10.1080/2330443X.2018.1442271
R. Kosar, D. W. Scott
ABSTRACT University ranking is a popular yet controversial endeavor. Most rankings are based on both public data, such as student test scores and retention rates, and proprietary data, such as school reputation as perceived by high school counselors and academic peers. The weights applied to these characteristics to compute the rankings are often determined in a subjective fashion. Of significant importance in the academic field, the Carnegie Classification was developed by the Carnegie Foundation for the Advancement of Teaching. It has been updated approximately every 5 years since 1973, most recently in February 2016. Based on bivariate scores, Carnegie assigns one of three classes (R1/R2/R3) to doctorate-granting universities according to their level of research activity. The Carnegie methodology uses only publicly available data and determines weights via principal component analysis. In this article, we review Carnegie’s stated goals and the extent to which their methodology achieves those goals. In particular, we examine Carnegie’s separation of aggregate and per capita (per tenured/tenure-track faculty member) variables and its use of two separate principal component analyses on each; the resulting bivariate scores are very highly correlated. We propose and evaluate two alternatives and provide a graphical tool for evaluating and comparing the three scenarios.
大学排名是一项很受欢迎但也有争议的工作。大多数排名既基于公共数据,如学生考试成绩和留校率,也基于专有数据,如高中辅导员和学术同行所认为的学校声誉。应用于这些特征以计算排名的权重通常是以主观方式确定的。卡内基分类法是由卡内基教学促进基金会制定的,在学术领域具有重要意义。自1973年以来,它大约每5年更新一次,最近一次是在2016年2月。基于双变量分数,卡内基根据研究活动水平将三个类别(R1/R2/R3)中的一个分配给授予博士学位的大学。卡内基的方法只使用公开可用的数据,并通过主成分分析确定权重。在这篇文章中,我们回顾了卡内基的既定目标,以及他们的方法在多大程度上实现了这些目标。特别地,我们研究了卡内基对总变量和人均变量(每个终身教职/终身教职教职员工)的分离,以及对每个变量使用两个单独的主成分分析;所得的双变量分数是高度相关的。我们提出并评估了两种方案,并提供了一个图形工具来评估和比较这三种方案。
{"title":"Examining the Carnegie Classification Methodology for Research Universities","authors":"R. Kosar, D. W. Scott","doi":"10.1080/2330443X.2018.1442271","DOIUrl":"https://doi.org/10.1080/2330443X.2018.1442271","url":null,"abstract":"ABSTRACT University ranking is a popular yet controversial endeavor. Most rankings are based on both public data, such as student test scores and retention rates, and proprietary data, such as school reputation as perceived by high school counselors and academic peers. The weights applied to these characteristics to compute the rankings are often determined in a subjective fashion. Of significant importance in the academic field, the Carnegie Classification was developed by the Carnegie Foundation for the Advancement of Teaching. It has been updated approximately every 5 years since 1973, most recently in February 2016. Based on bivariate scores, Carnegie assigns one of three classes (R1/R2/R3) to doctorate-granting universities according to their level of research activity. The Carnegie methodology uses only publicly available data and determines weights via principal component analysis. In this article, we review Carnegie’s stated goals and the extent to which their methodology achieves those goals. In particular, we examine Carnegie’s separation of aggregate and per capita (per tenured/tenure-track faculty member) variables and its use of two separate principal component analyses on each; the resulting bivariate scores are very highly correlated. We propose and evaluate two alternatives and provide a graphical tool for evaluating and comparing the three scenarios.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"5 1","pages":"1 - 12"},"PeriodicalIF":1.6,"publicationDate":"2018-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2018.1442271","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42101780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
The Role of Prior Information in Inference on the Annualized Rates of Mass Shootings in the United States 先验信息在推断美国每年大规模枪击事件发生率中的作用
IF 1.6 Q2 SOCIAL SCIENCES, MATHEMATICAL METHODS Pub Date : 2018-01-01 DOI: 10.1080/2330443X.2018.1448733
Nathan Sanders, Victor Lei
ABSTRACT While public debate over gun control in the United States has often hinged on individual public mass shooting incidents, legislative action should be informed by knowledge of the long-term evolution of these events. We present a new Bayesian model for the annualized rate of public mass shootings in the United States based on a Gaussian process with a time-varying mean function. While we present specific findings on long- and short-term trends of these shootings in the U.S., our focus is on understanding the role of model design and prior information in policy analysis. Using a Markov chain Monte Carlo inference technique, we explore the posterior consequences of different prior choices and explore correlations between hyperparameters. We demonstrate that the findings about the long-term evolution of the annualized rate of public mass shootings are robust to choices about prior information, while inferences about the timescale and amplitude of short-term variation depend sensitively on the prior. This work addresses the policy implications of implicit and explicit choices of prior information in model design and the utility of full Bayesian inference in evaluating the consequences of those choices.
摘要虽然美国关于枪支管制的公开辩论往往取决于个别公共大规模枪击事件,但立法行动应了解这些事件的长期演变。我们提出了一个新的贝叶斯模型,用于美国公共大规模枪击事件的年化率,该模型基于具有时变均值函数的高斯过程。虽然我们对美国这些枪击事件的长期和短期趋势提出了具体的调查结果,但我们的重点是了解模型设计和先前信息在政策分析中的作用。使用马尔可夫链蒙特卡罗推理技术,我们探索了不同先验选择的后验结果,并探索了超参数之间的相关性。我们证明,关于公共大规模枪击事件年化率长期演变的发现对先验信息的选择是稳健的,而关于短期变化的时间尺度和幅度的推断敏感地依赖于先验信息。这项工作解决了模型设计中先验信息的隐式和显式选择的政策含义,以及全贝叶斯推理在评估这些选择的后果中的效用。
{"title":"The Role of Prior Information in Inference on the Annualized Rates of Mass Shootings in the United States","authors":"Nathan Sanders, Victor Lei","doi":"10.1080/2330443X.2018.1448733","DOIUrl":"https://doi.org/10.1080/2330443X.2018.1448733","url":null,"abstract":"ABSTRACT While public debate over gun control in the United States has often hinged on individual public mass shooting incidents, legislative action should be informed by knowledge of the long-term evolution of these events. We present a new Bayesian model for the annualized rate of public mass shootings in the United States based on a Gaussian process with a time-varying mean function. While we present specific findings on long- and short-term trends of these shootings in the U.S., our focus is on understanding the role of model design and prior information in policy analysis. Using a Markov chain Monte Carlo inference technique, we explore the posterior consequences of different prior choices and explore correlations between hyperparameters. We demonstrate that the findings about the long-term evolution of the annualized rate of public mass shootings are robust to choices about prior information, while inferences about the timescale and amplitude of short-term variation depend sensitively on the prior. This work addresses the policy implications of implicit and explicit choices of prior information in model design and the utility of full Bayesian inference in evaluating the consequences of those choices.","PeriodicalId":43397,"journal":{"name":"Statistics and Public Policy","volume":"5 1","pages":"1 - 8"},"PeriodicalIF":1.6,"publicationDate":"2018-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/2330443X.2018.1448733","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45772292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
Statistics and Public Policy
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1