数字与传统研究模型的同行评估等级(PAR)的有效性、可重复性和再现性比较。

Q Medicine Australian Orthodontic Journal Pub Date : 2016-11-01
Sridhar Pasapula, Martyn Sherriff, Jeremy Breckon, Dirk Bister, Stefan Abela
{"title":"数字与传统研究模型的同行评估等级(PAR)的有效性、可重复性和再现性比较。","authors":"Sridhar Pasapula,&nbsp;Martyn Sherriff,&nbsp;Jeremy Breckon,&nbsp;Dirk Bister,&nbsp;Stefan Abela","doi":"","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>The validity, reliability and inter-method agreement of Peer Assessment Scores (PAR) from acrylic models and their digital analogues were assessed.</p><p><strong>Method: </strong>Ten models of different occlusions were digitised, using a 3 Shape R700 laser scanner (Copenhagen, Denmark). Each set of models was conventionally and digitally PAR-scored twice in random order by 10 examiners. The minimum time between repeat measurements was two weeks. The repeatability was assessed by applying Carstensen's analysis. Inter-method agreement (IEMA) was assessed by Carstensen's limit of agreement (LOA).</p><p><strong>Results: </strong>Intra-examiner repeatability (IER) for the unweighted and weighted data was slightly better for the conventional rather than the digital models. There was a slightly higher negative bias of -1 .62 for the weighted PAR data for the digital models. IEMA for the overall weighted data ranged from -8.70 - 5.45 (95% Confidence Interval, CI). Intra-class Correlation Coefficients lICC) for the weighted data for conventional, individual and average scenarios were 0.955 0.906 - 0.986 CI), 0.998 (0.995 - 0.999 CII. ICC for the weighted digital data, individual and average scenarios were 0.99 (0.97 - 1.00) and 1.00. The percentage reduction required to achieve an optimal occlusion increased by 0.4% for the digital scoring of the weighted data.</p><p><strong>Conclusion: </strong>Digital PAR scores obtained from scanned plastic models were valid and reliable and, in this context, the digital semi-automated method can be used interchangeably with the conventional method of PAR scoring.</p>","PeriodicalId":55417,"journal":{"name":"Australian Orthodontic Journal","volume":"32 2","pages":"184-192"},"PeriodicalIF":0.0000,"publicationDate":"2016-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Comparison of validity, repeatability and reproducibility of the Peer Assessment Rating (PAR) between digital and conventional study models.\",\"authors\":\"Sridhar Pasapula,&nbsp;Martyn Sherriff,&nbsp;Jeremy Breckon,&nbsp;Dirk Bister,&nbsp;Stefan Abela\",\"doi\":\"\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Introduction: </strong>The validity, reliability and inter-method agreement of Peer Assessment Scores (PAR) from acrylic models and their digital analogues were assessed.</p><p><strong>Method: </strong>Ten models of different occlusions were digitised, using a 3 Shape R700 laser scanner (Copenhagen, Denmark). Each set of models was conventionally and digitally PAR-scored twice in random order by 10 examiners. The minimum time between repeat measurements was two weeks. The repeatability was assessed by applying Carstensen's analysis. Inter-method agreement (IEMA) was assessed by Carstensen's limit of agreement (LOA).</p><p><strong>Results: </strong>Intra-examiner repeatability (IER) for the unweighted and weighted data was slightly better for the conventional rather than the digital models. There was a slightly higher negative bias of -1 .62 for the weighted PAR data for the digital models. IEMA for the overall weighted data ranged from -8.70 - 5.45 (95% Confidence Interval, CI). Intra-class Correlation Coefficients lICC) for the weighted data for conventional, individual and average scenarios were 0.955 0.906 - 0.986 CI), 0.998 (0.995 - 0.999 CII. ICC for the weighted digital data, individual and average scenarios were 0.99 (0.97 - 1.00) and 1.00. The percentage reduction required to achieve an optimal occlusion increased by 0.4% for the digital scoring of the weighted data.</p><p><strong>Conclusion: </strong>Digital PAR scores obtained from scanned plastic models were valid and reliable and, in this context, the digital semi-automated method can be used interchangeably with the conventional method of PAR scoring.</p>\",\"PeriodicalId\":55417,\"journal\":{\"name\":\"Australian Orthodontic Journal\",\"volume\":\"32 2\",\"pages\":\"184-192\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2016-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Australian Orthodontic Journal\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q\",\"JCRName\":\"Medicine\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Australian Orthodontic Journal","FirstCategoryId":"1085","ListUrlMain":"","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q","JCRName":"Medicine","Score":null,"Total":0}
引用次数: 0

摘要

前言:对丙烯酸模型及其数字类似物的同行评估分数(PAR)的效度、信度和方法间一致性进行了评估。方法:采用3 Shape R700激光扫描仪(丹麦哥本哈根)对10个不同咬合模型进行数字化。每组模型由10名审查员按随机顺序进行常规和数字par评分两次。重复测量的最小间隔时间为两周。应用Carstensen的分析来评估可重复性。采用Carstensen协议极限(LOA)评价方法间协议(IEMA)。结果:常规模型的未加权和加权数据的检查员内部重复性(IER)略好于数字模型。数字模型加权PAR数据的负偏倚略高,为- 1.62。整体加权数据的IEMA范围为-8.70 - 5.45(95%置信区间,CI)。常规情景、个体情景和平均情景加权数据的类内相关系数分别为0.955 (0.906 ~ 0.986 CI)、0.998 (0.995 ~ 0.999 CII)。加权数字数据、个人和平均情景的ICC分别为0.99(0.97 - 1.00)和1.00。对于加权数据的数字评分,实现最佳遮挡所需的百分比减少增加了0.4%。结论:通过扫描塑料模型获得的数字PAR评分是有效可靠的,在这种情况下,数字半自动化方法可以与传统的PAR评分方法互换使用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Comparison of validity, repeatability and reproducibility of the Peer Assessment Rating (PAR) between digital and conventional study models.

Introduction: The validity, reliability and inter-method agreement of Peer Assessment Scores (PAR) from acrylic models and their digital analogues were assessed.

Method: Ten models of different occlusions were digitised, using a 3 Shape R700 laser scanner (Copenhagen, Denmark). Each set of models was conventionally and digitally PAR-scored twice in random order by 10 examiners. The minimum time between repeat measurements was two weeks. The repeatability was assessed by applying Carstensen's analysis. Inter-method agreement (IEMA) was assessed by Carstensen's limit of agreement (LOA).

Results: Intra-examiner repeatability (IER) for the unweighted and weighted data was slightly better for the conventional rather than the digital models. There was a slightly higher negative bias of -1 .62 for the weighted PAR data for the digital models. IEMA for the overall weighted data ranged from -8.70 - 5.45 (95% Confidence Interval, CI). Intra-class Correlation Coefficients lICC) for the weighted data for conventional, individual and average scenarios were 0.955 0.906 - 0.986 CI), 0.998 (0.995 - 0.999 CII. ICC for the weighted digital data, individual and average scenarios were 0.99 (0.97 - 1.00) and 1.00. The percentage reduction required to achieve an optimal occlusion increased by 0.4% for the digital scoring of the weighted data.

Conclusion: Digital PAR scores obtained from scanned plastic models were valid and reliable and, in this context, the digital semi-automated method can be used interchangeably with the conventional method of PAR scoring.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Australian Orthodontic Journal
Australian Orthodontic Journal DENTISTRY, ORAL SURGERY & MEDICINE-
CiteScore
0.48
自引率
0.00%
发文量
0
审稿时长
>12 weeks
期刊最新文献
Minimizing publication bias. Looking backwards into the Future of Orthodontics. Orthodontic Education. Australian Orthodontists: are they "with it". The role of the foundation in Australian Orthodontic education.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1