回归和效应大小

IF 1 4区 医学 Q4 REHABILITATION Journal of Visual Impairment & Blindness Pub Date : 2023-03-01 DOI:10.1177/0145482X231166596
Robert Wall Emerson
{"title":"回归和效应大小","authors":"Robert Wall Emerson","doi":"10.1177/0145482X231166596","DOIUrl":null,"url":null,"abstract":"Back in the July–August issue of 2020, I discussed regression and R values. In this issue, the article titled, “Predictors of Job Satisfaction for People with Visual Impairments,” allows me to continue this discussion. In this article, authors Steverson and Crudden used multiple linear regression to identify predictors of job satisfaction. They also reported effect sizes for their analyses. Previously in this column, I have discussed the importance of reporting effect sizes for different statistical tests (e.g., January– February 2016 and July–August 2019). Although the concept of effect size is the same across all statistical tests (a measure of the size of the effect or difference being studied), how it is calculated is different for each statistical test. In the July–August 2020 Statistical Sidebar, I noted that the statistic η (read as eta squared) can be used as a measure of effect size in regression analyses. Other measures of effect size for regression include R (for the magnitude of the effect of the entire model), f 2 (for the magnitude of the effect of the entire regression model or individual predictors), and rpart (for the magnitude of the effect of individual predictors). The rpart statistic is called the squared semipartial correlation and is the measure used in the article under discussion in this issue. The R statistic, sometimes called the coefficient of determination, is where the main measures of effect size for regression all begin. Many researchers report the R value because it lends itself well to the interpretation of how much variability in the dependent variable is explained by the regression model. However, the use of the f 2 or rpart statistics allows researchers to focus on individual predictor variables. This operation is especially useful in multiple linear regressions where there are several predictor variables in the regression model. For a simple linear regression, where there is one predictor variable and one outcome variable, the R statistic is just the square of the correlation coefficient between the two variables. For more complicated regressions, R is calculated as 1-RSS/TSS where RSS is the sum of squared residuals and TSS is the total sum of squares. Without getting too deep into the details, these two measures relate to how far each datapoint in a dataset lies off of the regression line that minimizes the overall distance from the regression line to all the data points. The R statistic ranges from 0 to 1 where 0.01 is considered a small effect, 0.09 is a medium effect, and 0.25 is a large effect. The f 2 statistic is calculated based on the R statistic through the equation f =Rinc/ 1-Rinc where R 2 inc is the change in the overall R for a regression model when a given predictor variable is added to a group of other predictor variables. A rule of thumb is that for the f 2 statistic, 0.02 is a small effect, 0.15 is a medium effect, and 0.35 is a large effect. The rpart statistic is also calculated by running two regression models and comparing the R statistics for the two models. By adding one predictor variable to the second regression model, the difference in the R statistics for the two models gives the rpart statistic for that predictor variable. In the article under discussion in this issue, values for rpart are given for 21 different variables and the values range from< 0.001 to 0.028. In Table 4, only the largest 5 rpart measures are statistically significant (p < .05). Statistical Sidebar","PeriodicalId":47438,"journal":{"name":"Journal of Visual Impairment & Blindness","volume":"117 1","pages":"191 - 192"},"PeriodicalIF":1.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Regression and Effect Size\",\"authors\":\"Robert Wall Emerson\",\"doi\":\"10.1177/0145482X231166596\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Back in the July–August issue of 2020, I discussed regression and R values. In this issue, the article titled, “Predictors of Job Satisfaction for People with Visual Impairments,” allows me to continue this discussion. In this article, authors Steverson and Crudden used multiple linear regression to identify predictors of job satisfaction. They also reported effect sizes for their analyses. Previously in this column, I have discussed the importance of reporting effect sizes for different statistical tests (e.g., January– February 2016 and July–August 2019). Although the concept of effect size is the same across all statistical tests (a measure of the size of the effect or difference being studied), how it is calculated is different for each statistical test. In the July–August 2020 Statistical Sidebar, I noted that the statistic η (read as eta squared) can be used as a measure of effect size in regression analyses. Other measures of effect size for regression include R (for the magnitude of the effect of the entire model), f 2 (for the magnitude of the effect of the entire regression model or individual predictors), and rpart (for the magnitude of the effect of individual predictors). The rpart statistic is called the squared semipartial correlation and is the measure used in the article under discussion in this issue. The R statistic, sometimes called the coefficient of determination, is where the main measures of effect size for regression all begin. Many researchers report the R value because it lends itself well to the interpretation of how much variability in the dependent variable is explained by the regression model. However, the use of the f 2 or rpart statistics allows researchers to focus on individual predictor variables. This operation is especially useful in multiple linear regressions where there are several predictor variables in the regression model. For a simple linear regression, where there is one predictor variable and one outcome variable, the R statistic is just the square of the correlation coefficient between the two variables. For more complicated regressions, R is calculated as 1-RSS/TSS where RSS is the sum of squared residuals and TSS is the total sum of squares. Without getting too deep into the details, these two measures relate to how far each datapoint in a dataset lies off of the regression line that minimizes the overall distance from the regression line to all the data points. The R statistic ranges from 0 to 1 where 0.01 is considered a small effect, 0.09 is a medium effect, and 0.25 is a large effect. The f 2 statistic is calculated based on the R statistic through the equation f =Rinc/ 1-Rinc where R 2 inc is the change in the overall R for a regression model when a given predictor variable is added to a group of other predictor variables. A rule of thumb is that for the f 2 statistic, 0.02 is a small effect, 0.15 is a medium effect, and 0.35 is a large effect. The rpart statistic is also calculated by running two regression models and comparing the R statistics for the two models. By adding one predictor variable to the second regression model, the difference in the R statistics for the two models gives the rpart statistic for that predictor variable. In the article under discussion in this issue, values for rpart are given for 21 different variables and the values range from< 0.001 to 0.028. In Table 4, only the largest 5 rpart measures are statistically significant (p < .05). Statistical Sidebar\",\"PeriodicalId\":47438,\"journal\":{\"name\":\"Journal of Visual Impairment & Blindness\",\"volume\":\"117 1\",\"pages\":\"191 - 192\"},\"PeriodicalIF\":1.0000,\"publicationDate\":\"2023-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Visual Impairment & Blindness\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1177/0145482X231166596\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"REHABILITATION\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Visual Impairment & Blindness","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1177/0145482X231166596","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"REHABILITATION","Score":null,"Total":0}
引用次数: 0

摘要

早在2020年7 - 8月刊上,我就讨论了回归和R值。在本期中,一篇名为《视觉障碍人群工作满意度的预测因素》的文章允许我继续讨论这个问题。在这篇文章中,作者Steverson和Crudden使用多元线性回归来确定工作满意度的预测因子。他们还报告了分析的效应值。在本专栏之前,我已经讨论了报告不同统计测试(例如,2016年1月至2月和2019年7月至8月)的效应大小的重要性。虽然效应大小的概念在所有的统计测试中都是一样的(衡量正在研究的效应或差异的大小),但计算方法在每个统计测试中都是不同的。在2020年7月至8月的统计侧栏中,我注意到统计η(读取为eta平方)可以用作回归分析中效应大小的度量。回归效应大小的其他度量包括R(整个模型的影响大小),f2(整个回归模型或单个预测因子的影响大小)和rpart(单个预测因子的影响大小)。rpart统计量称为平方半偏相关,是本文所讨论的度量。R统计量,有时被称为决定系数,是回归效应大小的主要度量开始的地方。许多研究人员报告了R值,因为它很好地解释了回归模型解释了多少因变量的可变性。然而,使用f2或rpart统计可以使研究人员专注于单个预测变量。此操作在多元线性回归中特别有用,其中回归模型中有几个预测变量。对于一个简单的线性回归,其中有一个预测变量和一个结果变量,R统计量只是两个变量之间相关系数的平方。对于更复杂的回归,R计算为1-RSS/TSS,其中RSS为残差平方和,TSS为总平方和。在不深入讨论细节的情况下,这两个度量与数据集中每个数据点离回归线的距离有关,这使得从回归线到所有数据点的总距离最小。R统计量的取值范围为0到1,其中0.01为小影响,0.09为中等影响,0.25为大影响。f2统计量是根据R统计量通过公式f = ring / 1- ring计算的,其中r2 inc是将给定的预测变量添加到一组其他预测变量时回归模型的总体R的变化。经验法则是,对于f2统计量,0.02是小影响,0.15是中等影响,0.35是大影响。rpart统计量也通过运行两个回归模型并比较两个模型的R统计量来计算。通过向第二个回归模型添加一个预测变量,两个模型的R统计量的差异给出了该预测变量的rpart统计量。在本期讨论的文章中,给出了21个不同变量的rpart值,取值范围从< 0.001到0.028。在表4中,只有最大的5个rpart测量值具有统计学意义(p < 0.05)。统计栏
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Regression and Effect Size
Back in the July–August issue of 2020, I discussed regression and R values. In this issue, the article titled, “Predictors of Job Satisfaction for People with Visual Impairments,” allows me to continue this discussion. In this article, authors Steverson and Crudden used multiple linear regression to identify predictors of job satisfaction. They also reported effect sizes for their analyses. Previously in this column, I have discussed the importance of reporting effect sizes for different statistical tests (e.g., January– February 2016 and July–August 2019). Although the concept of effect size is the same across all statistical tests (a measure of the size of the effect or difference being studied), how it is calculated is different for each statistical test. In the July–August 2020 Statistical Sidebar, I noted that the statistic η (read as eta squared) can be used as a measure of effect size in regression analyses. Other measures of effect size for regression include R (for the magnitude of the effect of the entire model), f 2 (for the magnitude of the effect of the entire regression model or individual predictors), and rpart (for the magnitude of the effect of individual predictors). The rpart statistic is called the squared semipartial correlation and is the measure used in the article under discussion in this issue. The R statistic, sometimes called the coefficient of determination, is where the main measures of effect size for regression all begin. Many researchers report the R value because it lends itself well to the interpretation of how much variability in the dependent variable is explained by the regression model. However, the use of the f 2 or rpart statistics allows researchers to focus on individual predictor variables. This operation is especially useful in multiple linear regressions where there are several predictor variables in the regression model. For a simple linear regression, where there is one predictor variable and one outcome variable, the R statistic is just the square of the correlation coefficient between the two variables. For more complicated regressions, R is calculated as 1-RSS/TSS where RSS is the sum of squared residuals and TSS is the total sum of squares. Without getting too deep into the details, these two measures relate to how far each datapoint in a dataset lies off of the regression line that minimizes the overall distance from the regression line to all the data points. The R statistic ranges from 0 to 1 where 0.01 is considered a small effect, 0.09 is a medium effect, and 0.25 is a large effect. The f 2 statistic is calculated based on the R statistic through the equation f =Rinc/ 1-Rinc where R 2 inc is the change in the overall R for a regression model when a given predictor variable is added to a group of other predictor variables. A rule of thumb is that for the f 2 statistic, 0.02 is a small effect, 0.15 is a medium effect, and 0.35 is a large effect. The rpart statistic is also calculated by running two regression models and comparing the R statistics for the two models. By adding one predictor variable to the second regression model, the difference in the R statistics for the two models gives the rpart statistic for that predictor variable. In the article under discussion in this issue, values for rpart are given for 21 different variables and the values range from< 0.001 to 0.028. In Table 4, only the largest 5 rpart measures are statistically significant (p < .05). Statistical Sidebar
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
1.30
自引率
18.20%
发文量
68
期刊介绍: The Journal of Visual Impairment & Blindness is the essential professional resource for information about visual impairment (that is, blindness or low vision). The international peer-reviewed journal of record in the field, it delivers current research and best practice information, commentary from authoritative experts on critical topics, News From the Field, and a calendar of important events. Practitioners and researchers, policymakers and administrators, counselors and advocates rely on JVIB for its delivery of cutting-edge research and the most up-to-date practices in the field of visual impairment and blindness. Available in print and online 24/7, JVIB offers immediate access to information from the leading researchers, teachers of students with visual impairments (often referred to as TVIs), orientation and mobility (O&M) practitioners, vision rehabilitation therapists (often referred to as VRTs), early interventionists, and low vision therapists (often referred to as LVTs) in the field.
期刊最新文献
A graded neonatal mouse model of necrotizing enterocolitis demonstrates that mild enterocolitis is sufficient to activate microglia and increase cerebral cytokine expression. Feasibility and Acceptability of Implementing a Job Search Intervention for Adults With Visual Impairments via Videoconferencing. Components of Valid Learning Media Assessments Using Checklists as a Vocational Rehabilitation Tool for Employed Consumers JVIB Peer Reviewers
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1