首页 > 最新文献

Research integrity and peer review最新文献

英文 中文
Individual versus general structured feedback to improve agreement in grant peer review: a randomized controlled trial. 个人与一般结构化反馈以提高拨款同行评审的一致性:一项随机对照试验。
Q1 ETHICS Pub Date : 2021-09-30 DOI: 10.1186/s41073-021-00115-5
Jan-Ole Hesselberg, Knut Inge Fostervold, Pål Ulleberg, Ida Svege

Background: Vast sums are distributed based on grant peer review, but studies show that interrater reliability is often low. In this study, we tested the effect of receiving two short individual feedback reports compared to one short general feedback report on the agreement between reviewers.

Methods: A total of 42 reviewers at the Norwegian Foundation Dam were randomly assigned to receive either a general feedback report or an individual feedback report. The general feedback group received one report before the start of the reviews that contained general information about the previous call in which the reviewers participated. In the individual feedback group, the reviewers received two reports, one before the review period (based on the previous call) and one during the period (based on the current call). In the individual feedback group, the reviewers were presented with detailed information on their scoring compared with the review committee as a whole, both before and during the review period. The main outcomes were the proportion of agreement in the eligibility assessment and the average difference in scores between pairs of reviewers assessing the same proposal. The outcomes were measured in 2017 and after the feedback was provided in 2018.

Results: A total of 2398 paired reviews were included in the analysis. There was a significant difference between the two groups in the proportion of absolute agreement on whether the proposal was eligible for the funding programme, with the general feedback group demonstrating a higher rate of agreement. There was no difference between the two groups in terms of the average score difference. However, the agreement regarding the proposal score remained critically low for both groups.

Conclusions: We did not observe changes in proposal score agreement between 2017 and 2018 in reviewers receiving different feedback. The low levels of agreement remain a major concern in grant peer review, and research to identify contributing factors as well as the development and testing of interventions to increase agreement rates are still needed.

Trial registration: The study was preregistered at OSF.io/n4fq3 .

背景:大量的金额是根据拨款同行评审进行分配的,但研究表明,评审者之间的可靠性通常很低。在这项研究中,我们测试了两份简短的个人反馈报告与一份简短的一般反馈报告对评审员之间一致性的影响。方法:挪威基础大坝共有42名评审员被随机分配接受一般反馈报告或个人反馈报告。一般性反馈小组在审查开始前收到一份报告,其中载有关于审查人员参加的上一次电话会议的一般性信息。在个人反馈组中,评审员收到了两份报告,一份是在评审期前(基于上一次电话),另一份是在此期间(基于当前电话)。在个人反馈组中,在审查之前和审查期间,向审查人员提供了与整个审查委员会相比的评分详细信息。主要结果是在资格评估中达成一致的比例,以及评估同一提案的两对评审员之间的平均得分差异。在2017年和2018年提供反馈后对结果进行了测量。结果:共有2398条配对评论被纳入分析。两组在提案是否符合资助方案的绝对一致比例方面存在显著差异,一般反馈组的一致率更高。两组之间的平均得分差异没有差异。然而,对于这两组人来说,关于提案得分的一致性仍然极低。结论:在2017年至2018年间,我们没有观察到收到不同反馈的评审员的提案得分一致性发生变化。低水平的协议仍然是赠款同行审查中的一个主要问题,仍然需要进行研究以确定促成因素,以及制定和测试提高协议率的干预措施。试验注册:该研究在OSF.io/n4fq3预先注册。
{"title":"Individual versus general structured feedback to improve agreement in grant peer review: a randomized controlled trial.","authors":"Jan-Ole Hesselberg,&nbsp;Knut Inge Fostervold,&nbsp;Pål Ulleberg,&nbsp;Ida Svege","doi":"10.1186/s41073-021-00115-5","DOIUrl":"10.1186/s41073-021-00115-5","url":null,"abstract":"<p><strong>Background: </strong>Vast sums are distributed based on grant peer review, but studies show that interrater reliability is often low. In this study, we tested the effect of receiving two short individual feedback reports compared to one short general feedback report on the agreement between reviewers.</p><p><strong>Methods: </strong>A total of 42 reviewers at the Norwegian Foundation Dam were randomly assigned to receive either a general feedback report or an individual feedback report. The general feedback group received one report before the start of the reviews that contained general information about the previous call in which the reviewers participated. In the individual feedback group, the reviewers received two reports, one before the review period (based on the previous call) and one during the period (based on the current call). In the individual feedback group, the reviewers were presented with detailed information on their scoring compared with the review committee as a whole, both before and during the review period. The main outcomes were the proportion of agreement in the eligibility assessment and the average difference in scores between pairs of reviewers assessing the same proposal. The outcomes were measured in 2017 and after the feedback was provided in 2018.</p><p><strong>Results: </strong>A total of 2398 paired reviews were included in the analysis. There was a significant difference between the two groups in the proportion of absolute agreement on whether the proposal was eligible for the funding programme, with the general feedback group demonstrating a higher rate of agreement. There was no difference between the two groups in terms of the average score difference. However, the agreement regarding the proposal score remained critically low for both groups.</p><p><strong>Conclusions: </strong>We did not observe changes in proposal score agreement between 2017 and 2018 in reviewers receiving different feedback. The low levels of agreement remain a major concern in grant peer review, and research to identify contributing factors as well as the development and testing of interventions to increase agreement rates are still needed.</p><p><strong>Trial registration: </strong>The study was preregistered at OSF.io/n4fq3 .</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"12"},"PeriodicalIF":0.0,"publicationDate":"2021-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8485516/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39474032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Strengthening the incentives for responsible research practices in Australian health and medical research funding. 加强在澳大利亚卫生和医学研究供资方面鼓励负责任的研究做法。
Q1 ETHICS Pub Date : 2021-08-02 DOI: 10.1186/s41073-021-00113-7
Joanna Diong, Cynthia M Kroeger, Katherine J Reynolds, Adrian Barnett, Lisa A Bero

Background: Australian health and medical research funders support substantial research efforts, and incentives within grant funding schemes influence researcher behaviour. We aimed to determine to what extent Australian health and medical funders incentivise responsible research practices.

Methods: We conducted an audit of instructions from research grant and fellowship schemes. Eight national research grants and fellowships were purposively sampled to select schemes that awarded the largest amount of funds. The funding scheme instructions were assessed against 9 criteria to determine to what extent they incentivised these responsible research and reporting practices: (1) publicly register study protocols before starting data collection, (2) register analysis protocols before starting data analysis, (3) make study data openly available, (4) make analysis code openly available, (5) make research materials openly available, (6) discourage use of publication metrics, (7) conduct quality research (e.g. adhere to reporting guidelines), (8) collaborate with a statistician, and (9) adhere to other responsible research practices. Each criterion was answered using one of the following responses: "Instructed", "Encouraged", or "No mention".

Results: Across the 8 schemes from 5 funders, applicants were instructed or encouraged to address a median of 4 (range 0 to 5) of the 9 criteria. Three criteria received no mention in any scheme (register analysis protocols, make analysis code open, collaborate with a statistician). Importantly, most incentives did not seem strong as applicants were only instructed to register study protocols, discourage use of publication metrics and conduct quality research. Other criteria were encouraged but were not required.

Conclusions: Funders could strengthen the incentives for responsible research practices by requiring grant and fellowship applicants to implement these practices in their proposals. Administering institutions could be required to implement these practices to be eligible for funding. Strongly rewarding researchers for implementing robust research practices could lead to sustained improvements in the quality of health and medical research.

背景:澳大利亚健康和医学研究资助者支持大量的研究工作,资助计划中的激励措施会影响研究人员的行为。我们旨在确定澳大利亚卫生和医疗资助者在多大程度上激励负责任的研究实践。方法:我们对研究资助和研究金计划的指示进行了审计。有目的地对八项国家研究补助金和研究金进行了抽样,以选择获得最多资金的计划。根据9个标准对资助计划说明进行了评估,以确定它们在多大程度上激励了这些负责任的研究和报告实践:(1)在开始数据收集之前公开注册研究协议,(2)在开始分析之前注册分析协议,(3)公开研究数据,(4)公开分析代码,(5)公开研究材料,(6)不鼓励使用发表指标,(7)进行高质量的研究(例如遵守报告指南),(8)与统计学家合作,以及(9)遵守其他负责任的研究实践。每个标准都使用以下回答之一回答:“指示”、“鼓励”或“不提及”。结果:在来自5名资助者的8个计划中,申请人被指示或鼓励满足9个标准中的4个(范围0至5)。在任何方案中都没有提到三个标准(注册分析协议、开放分析代码、与统计学家合作)。重要的是,大多数激励措施似乎并不强烈,因为申请人只被要求注册研究方案,不鼓励使用发表指标,并进行高质量的研究。鼓励但不要求采用其他标准。结论:资助者可以通过要求拨款和研究金申请人在其提案中实施这些做法来加强对负责任研究实践的激励。管理机构可能被要求实施这些做法,才有资格获得资助。大力奖励实施稳健研究实践的研究人员,可以持续提高健康和医学研究的质量。
{"title":"Strengthening the incentives for responsible research practices in Australian health and medical research funding.","authors":"Joanna Diong,&nbsp;Cynthia M Kroeger,&nbsp;Katherine J Reynolds,&nbsp;Adrian Barnett,&nbsp;Lisa A Bero","doi":"10.1186/s41073-021-00113-7","DOIUrl":"10.1186/s41073-021-00113-7","url":null,"abstract":"<p><strong>Background: </strong>Australian health and medical research funders support substantial research efforts, and incentives within grant funding schemes influence researcher behaviour. We aimed to determine to what extent Australian health and medical funders incentivise responsible research practices.</p><p><strong>Methods: </strong>We conducted an audit of instructions from research grant and fellowship schemes. Eight national research grants and fellowships were purposively sampled to select schemes that awarded the largest amount of funds. The funding scheme instructions were assessed against 9 criteria to determine to what extent they incentivised these responsible research and reporting practices: (1) publicly register study protocols before starting data collection, (2) register analysis protocols before starting data analysis, (3) make study data openly available, (4) make analysis code openly available, (5) make research materials openly available, (6) discourage use of publication metrics, (7) conduct quality research (e.g. adhere to reporting guidelines), (8) collaborate with a statistician, and (9) adhere to other responsible research practices. Each criterion was answered using one of the following responses: \"Instructed\", \"Encouraged\", or \"No mention\".</p><p><strong>Results: </strong>Across the 8 schemes from 5 funders, applicants were instructed or encouraged to address a median of 4 (range 0 to 5) of the 9 criteria. Three criteria received no mention in any scheme (register analysis protocols, make analysis code open, collaborate with a statistician). Importantly, most incentives did not seem strong as applicants were only instructed to register study protocols, discourage use of publication metrics and conduct quality research. Other criteria were encouraged but were not required.</p><p><strong>Conclusions: </strong>Funders could strengthen the incentives for responsible research practices by requiring grant and fellowship applicants to implement these practices in their proposals. Administering institutions could be required to implement these practices to be eligible for funding. Strongly rewarding researchers for implementing robust research practices could lead to sustained improvements in the quality of health and medical research.</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"11"},"PeriodicalIF":0.0,"publicationDate":"2021-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8328133/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39277405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Correction to: Cross-sectional study of medical advertisements in a national general medical journal: evidence, cost, and safe use of advertised versus comparative drugs. 更正:国家普通医学杂志上医疗广告的横断面研究:广告与比较药物的证据、成本和安全使用。
Q1 ETHICS Pub Date : 2021-06-11 DOI: 10.1186/s41073-021-00114-6
Kim Boesen, Anders Lykkemark Simonsen, Karsten Juhl Jørgensen, Peter C Gøtzsche
{"title":"Correction to: Cross-sectional study of medical advertisements in a national general medical journal: evidence, cost, and safe use of advertised versus comparative drugs.","authors":"Kim Boesen,&nbsp;Anders Lykkemark Simonsen,&nbsp;Karsten Juhl Jørgensen,&nbsp;Peter C Gøtzsche","doi":"10.1186/s41073-021-00114-6","DOIUrl":"https://doi.org/10.1186/s41073-021-00114-6","url":null,"abstract":"","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"10"},"PeriodicalIF":0.0,"publicationDate":"2021-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41073-021-00114-6","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39086140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating implementation of the Transparency and Openness Promotion (TOP) guidelines: the TRUST process for rating journal policies, procedures, and practices. 评估 "促进透明与公开(TOP)"指南的实施情况:对期刊政策、程序和实践进行评级的 TRUST 程序。
Q1 ETHICS Pub Date : 2021-06-02 DOI: 10.1186/s41073-021-00112-8
Evan Mayo-Wilson, Sean Grant, Lauren Supplee, Sina Kianersi, Afsah Amin, Alex DeHaven, David Mellor

Background: The Transparency and Openness Promotion (TOP) Guidelines describe modular standards that journals can adopt to promote open science. The TOP Factor is a metric to describe the extent to which journals have adopted the TOP Guidelines in their policies. Systematic methods and rating instruments are needed to calculate the TOP Factor. Moreover, implementation of these open science policies depends on journal procedures and practices, for which TOP provides no standards or rating instruments.

Methods: We describe a process for assessing journal policies, procedures, and practices according to the TOP Guidelines. We developed this process as part of the Transparency of Research Underpinning Social Intervention Tiers (TRUST) Initiative to advance open science in the social intervention research ecosystem. We also provide new instruments for rating journal instructions to authors (policies), manuscript submission systems (procedures), and published articles (practices) according to standards in the TOP Guidelines. In addition, we describe how to determine the TOP Factor score for a journal, calculate reliability of journal ratings, and assess coherence among a journal's policies, procedures, and practices. As a demonstration of this process, we describe a protocol for studying approximately 345 influential journals that have published research used to inform evidence-based policy.

Discussion: The TRUST Process includes systematic methods and rating instruments for assessing and facilitating implementation of the TOP Guidelines by journals across disciplines. Our study of journals publishing influential social intervention research will provide a comprehensive account of whether these journals have policies, procedures, and practices that are consistent with standards for open science and thereby facilitate the publication of trustworthy findings to inform evidence-based policy. Through this demonstration, we expect to identify ways to refine the TOP Guidelines and the TOP Factor. Refinements could include: improving templates for adoption in journal instructions to authors, manuscript submission systems, and published articles; revising explanatory guidance intended to enhance the use, understanding, and dissemination of the TOP Guidelines; and clarifying the distinctions among different levels of implementation. Research materials are available on the Open Science Framework: https://osf.io/txyr3/ .

背景:透明度和公开性促进(TOP)指南》描述了期刊为促进开放科学而可以采用的模块标准。TOP Factor 是一种衡量标准,用来描述期刊在其政策中采用《透明度与公开性促进指南》的程度。计算 TOP 因子需要系统的方法和评级工具。此外,这些开放科学政策的实施取决于期刊的程序和实践,而 TOP 并没有提供这方面的标准或评级工具:方法:我们介绍了根据《顶级期刊指南》评估期刊政策、程序和实践的流程。我们开发了这一流程,作为社会干预层级研究透明度(TRUST)计划的一部分,以推动社会干预研究生态系统中的开放科学。我们还提供了新的工具,用于根据《TOP 指南》中的标准对期刊的作者须知(政策)、投稿系统(程序)和已发表文章(实践)进行评级。此外,我们还介绍了如何确定期刊的 TOP 因子得分,计算期刊评级的可靠性,以及评估期刊政策、程序和实践之间的一致性。作为该流程的演示,我们介绍了对约 345 种有影响力的期刊进行研究的方案,这些期刊发表的研究成果为循证政策提供了依据:TRUST 流程包括系统方法和评级工具,用于评估和促进各学科期刊实施《顶级期刊指南》。我们对发表有影响力的社会干预研究的期刊进行的研究将全面说明这些期刊是否拥有符合开放科学标准的政策、程序和实践,从而促进发表可信的研究成果,为循证政策提供依据。通过此次论证,我们有望找到完善《顶级期刊指南》和《顶级期刊因子》的方法。完善工作可包括:改进模板,以便在期刊的作者须知、投稿系统和发表的文章中采用;修订解释性指南,以加强对《最高学术标准指南》的使用、理解和传播;以及明确不同实施水平之间的区别。研究材料可在开放科学框架网站上查阅:https://osf.io/txyr3/ 。
{"title":"Evaluating implementation of the Transparency and Openness Promotion (TOP) guidelines: the TRUST process for rating journal policies, procedures, and practices.","authors":"Evan Mayo-Wilson, Sean Grant, Lauren Supplee, Sina Kianersi, Afsah Amin, Alex DeHaven, David Mellor","doi":"10.1186/s41073-021-00112-8","DOIUrl":"10.1186/s41073-021-00112-8","url":null,"abstract":"<p><strong>Background: </strong>The Transparency and Openness Promotion (TOP) Guidelines describe modular standards that journals can adopt to promote open science. The TOP Factor is a metric to describe the extent to which journals have adopted the TOP Guidelines in their policies. Systematic methods and rating instruments are needed to calculate the TOP Factor. Moreover, implementation of these open science policies depends on journal procedures and practices, for which TOP provides no standards or rating instruments.</p><p><strong>Methods: </strong>We describe a process for assessing journal policies, procedures, and practices according to the TOP Guidelines. We developed this process as part of the Transparency of Research Underpinning Social Intervention Tiers (TRUST) Initiative to advance open science in the social intervention research ecosystem. We also provide new instruments for rating journal instructions to authors (policies), manuscript submission systems (procedures), and published articles (practices) according to standards in the TOP Guidelines. In addition, we describe how to determine the TOP Factor score for a journal, calculate reliability of journal ratings, and assess coherence among a journal's policies, procedures, and practices. As a demonstration of this process, we describe a protocol for studying approximately 345 influential journals that have published research used to inform evidence-based policy.</p><p><strong>Discussion: </strong>The TRUST Process includes systematic methods and rating instruments for assessing and facilitating implementation of the TOP Guidelines by journals across disciplines. Our study of journals publishing influential social intervention research will provide a comprehensive account of whether these journals have policies, procedures, and practices that are consistent with standards for open science and thereby facilitate the publication of trustworthy findings to inform evidence-based policy. Through this demonstration, we expect to identify ways to refine the TOP Guidelines and the TOP Factor. Refinements could include: improving templates for adoption in journal instructions to authors, manuscript submission systems, and published articles; revising explanatory guidance intended to enhance the use, understanding, and dissemination of the TOP Guidelines; and clarifying the distinctions among different levels of implementation. Research materials are available on the Open Science Framework: https://osf.io/txyr3/ .</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"9"},"PeriodicalIF":0.0,"publicationDate":"2021-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8173977/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39055385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cross-sectional study of medical advertisements in a national general medical journal: evidence, cost, and safe use of advertised versus comparative drugs. 全国性普通医学杂志上医疗广告的横断面研究:广告与比较药物的证据、成本和安全使用。
Q1 ETHICS Pub Date : 2021-05-10 DOI: 10.1186/s41073-021-00111-9
Kim Boesen, Anders Lykkemark Simonsen, Karsten Juhl Jørgensen, Peter C Gøtzsche

Background: Healthcare professionals are exposed to advertisements for prescription drugs in medical journals. Such advertisements may increase prescriptions of new drugs at the expense of older treatments even when they have no added benefits, are more harmful, and are more expensive. The publication of medical advertisements therefore raises ethical questions related to editorial integrity.

Methods: We conducted a descriptive cross-sectional study of all medical advertisements published in the Journal of the Danish Medical Association in 2015. Drugs advertised 6 times or more were compared with older comparators: (1) comparative evidence of added benefit; (2) Defined Daily Dose cost; (3) regulatory safety announcements; and (4) completed and ongoing post-marketing studies 3 years after advertising.

Results: We found 158 medical advertisements for 35 prescription drugs published in 24 issues during 2015, with a median of 7 advertisements per issue (range 0 to 11). Four drug groups and 5 single drugs were advertised 6 times or more, for a total of 10 indications, and we made 14 comparisons with older treatments. We found: (1) 'no added benefit' in 4 (29%) of 14 comparisons, 'uncertain benefits' in 7 (50%), and 'no evidence' in 3 (21%) comparisons. In no comparison did we find evidence of 'substantial added benefit' for the new drug; (2) advertised drugs were 2 to 196 times (median 6) more expensive per Defined Daily Dose; (3) 11 safety announcements for five advertised drugs were issued compared to one announcement for one comparator drug; (4) 20 post-marketing studies (7 completed, 13 ongoing) were requested for the advertised drugs versus 10 studies (4 completed, 6 ongoing) for the comparator drugs, and 7 studies (2 completed, 5 ongoing) assessed both an advertised and a comparator drug at 3 year follow-up.

Conclusions and relevance: In this cross-sectional study of medical advertisements published in the Journal of the Danish Medical Association during 2015, the most advertised drugs did not have documented substantial added benefits over older treatments, whereas they were substantially more expensive. From January 2021, the Journal of the Danish Medical Association no longer publishes medical advertisements.

背景:医疗保健专业人员接触到医学杂志上的处方药广告。这样的广告可能会增加新药的处方,而牺牲旧的治疗方法,即使它们没有额外的好处,更有害,更昂贵。因此,医疗广告的出版引发了与编辑诚信有关的伦理问题。方法:我们对2015年发表在《丹麦医学会杂志》上的所有医疗广告进行了描述性横断面研究。广告6次或6次以上的药物与较老的比较者进行比较:(1)增加获益的比较证据;(2)限定日剂量费用;(三)监管安全公告;(4)广告后3年完成并正在进行的营销后研究。结果:2015年共24期共发现35种处方药158条医疗广告,平均每期7条(范围0 ~ 11)。4个药物组和5个单一药物广告6次及以上,共10个适应症,我们与老疗法进行了14次比较。我们发现:(1)在14项比较中,有4项(29%)为“无额外益处”,7项(50%)为“不确定益处”,3项(21%)为“无证据”。在没有比较的情况下,我们没有发现新药有“实质性的额外益处”的证据;(2)广告药品每限定日剂量贵2 - 196倍(中位数6);(3) 5种药品发布11个安全公告,1种比较药发布1个安全公告;(4) 20项上市后研究(7项已完成,13项正在进行)用于广告药物,10项研究(4项已完成,6项正在进行)用于比较药物,7项研究(2项已完成,5项正在进行)在3年随访期间评估了广告药物和比较药物。结论和相关性:在2015年发表在《丹麦医学协会杂志》(Journal of the Danish medical Association)上的医疗广告的横断面研究中,广告最多的药物并没有证明比旧疗法有实质性的额外益处,相反,它们的价格要贵得多。从2021年1月起,《丹麦医学会杂志》不再刊登医疗广告。
{"title":"Cross-sectional study of medical advertisements in a national general medical journal: evidence, cost, and safe use of advertised versus comparative drugs.","authors":"Kim Boesen,&nbsp;Anders Lykkemark Simonsen,&nbsp;Karsten Juhl Jørgensen,&nbsp;Peter C Gøtzsche","doi":"10.1186/s41073-021-00111-9","DOIUrl":"https://doi.org/10.1186/s41073-021-00111-9","url":null,"abstract":"<p><strong>Background: </strong>Healthcare professionals are exposed to advertisements for prescription drugs in medical journals. Such advertisements may increase prescriptions of new drugs at the expense of older treatments even when they have no added benefits, are more harmful, and are more expensive. The publication of medical advertisements therefore raises ethical questions related to editorial integrity.</p><p><strong>Methods: </strong>We conducted a descriptive cross-sectional study of all medical advertisements published in the Journal of the Danish Medical Association in 2015. Drugs advertised 6 times or more were compared with older comparators: (1) comparative evidence of added benefit; (2) Defined Daily Dose cost; (3) regulatory safety announcements; and (4) completed and ongoing post-marketing studies 3 years after advertising.</p><p><strong>Results: </strong>We found 158 medical advertisements for 35 prescription drugs published in 24 issues during 2015, with a median of 7 advertisements per issue (range 0 to 11). Four drug groups and 5 single drugs were advertised 6 times or more, for a total of 10 indications, and we made 14 comparisons with older treatments. We found: (1) 'no added benefit' in 4 (29%) of 14 comparisons, 'uncertain benefits' in 7 (50%), and 'no evidence' in 3 (21%) comparisons. In no comparison did we find evidence of 'substantial added benefit' for the new drug; (2) advertised drugs were 2 to 196 times (median 6) more expensive per Defined Daily Dose; (3) 11 safety announcements for five advertised drugs were issued compared to one announcement for one comparator drug; (4) 20 post-marketing studies (7 completed, 13 ongoing) were requested for the advertised drugs versus 10 studies (4 completed, 6 ongoing) for the comparator drugs, and 7 studies (2 completed, 5 ongoing) assessed both an advertised and a comparator drug at 3 year follow-up.</p><p><strong>Conclusions and relevance: </strong>In this cross-sectional study of medical advertisements published in the Journal of the Danish Medical Association during 2015, the most advertised drugs did not have documented substantial added benefits over older treatments, whereas they were substantially more expensive. From January 2021, the Journal of the Danish Medical Association no longer publishes medical advertisements.</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"8"},"PeriodicalIF":0.0,"publicationDate":"2021-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41073-021-00111-9","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38968548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Explaining variance in perceived research misbehavior: results from a survey among academic researchers in Amsterdam. 解释被感知的研究不当行为的差异:来自阿姆斯特丹学术研究人员的调查结果。
Q1 ETHICS Pub Date : 2021-05-03 DOI: 10.1186/s41073-021-00110-w
Tamarinde Haven, Joeri Tijdink, Brian Martinson, Lex Bouter, Frans Oort

Background: Concerns about research misbehavior in academic science have sparked interest in the factors that may explain research misbehavior. Often three clusters of factors are distinguished: individual factors, climate factors and publication factors. Our research question was: to what extent can individual, climate and publication factors explain the variance in frequently perceived research misbehaviors?

Methods: From May 2017 until July 2017, we conducted a survey study among academic researchers in Amsterdam. The survey included three measurement instruments that we previously reported individual results of and here we integrate these findings.

Results: One thousand two hundred ninety-eight researchers completed the survey (response rate: 17%). Results showed that individual, climate and publication factors combined explained 34% of variance in perceived frequency of research misbehavior. Individual factors explained 7%, climate factors explained 22% and publication factors 16%.

Conclusions: Our results suggest that the perceptions of the research climate play a substantial role in explaining variance in research misbehavior. This suggests that efforts to improve departmental norms might have a salutary effect on behavior.

背景:对学术科学中研究不端行为的关注引发了人们对可能解释研究不端行为的因素的兴趣。通常可以区分三组因素:个人因素、气候因素和出版因素。我们的研究问题是:个人、气候和出版因素在多大程度上可以解释经常被感知的研究不当行为的差异?方法:2017年5月至2017年7月,我们对阿姆斯特丹的学术研究人员进行了调查研究。该调查包括三种测量工具,我们之前报告了各自的结果,在这里我们整合了这些发现。结果:共有1298名研究人员完成调查,回复率为17%。结果显示,个人、气候和出版因素共同解释了34%的研究不当行为感知频率差异。个人因素解释了7%,气候因素解释了22%,出版因素解释了16%。结论:我们的研究结果表明,对研究气候的感知在解释研究不当行为的差异方面发挥了重要作用。这表明,努力改善部门规范可能会对行为产生有益的影响。
{"title":"Explaining variance in perceived research misbehavior: results from a survey among academic researchers in Amsterdam.","authors":"Tamarinde Haven,&nbsp;Joeri Tijdink,&nbsp;Brian Martinson,&nbsp;Lex Bouter,&nbsp;Frans Oort","doi":"10.1186/s41073-021-00110-w","DOIUrl":"https://doi.org/10.1186/s41073-021-00110-w","url":null,"abstract":"<p><strong>Background: </strong>Concerns about research misbehavior in academic science have sparked interest in the factors that may explain research misbehavior. Often three clusters of factors are distinguished: individual factors, climate factors and publication factors. Our research question was: to what extent can individual, climate and publication factors explain the variance in frequently perceived research misbehaviors?</p><p><strong>Methods: </strong>From May 2017 until July 2017, we conducted a survey study among academic researchers in Amsterdam. The survey included three measurement instruments that we previously reported individual results of and here we integrate these findings.</p><p><strong>Results: </strong>One thousand two hundred ninety-eight researchers completed the survey (response rate: 17%). Results showed that individual, climate and publication factors combined explained 34% of variance in perceived frequency of research misbehavior. Individual factors explained 7%, climate factors explained 22% and publication factors 16%.</p><p><strong>Conclusions: </strong>Our results suggest that the perceptions of the research climate play a substantial role in explaining variance in research misbehavior. This suggests that efforts to improve departmental norms might have a salutary effect on behavior.</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"7"},"PeriodicalIF":0.0,"publicationDate":"2021-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41073-021-00110-w","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38944409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Cooperation & Liaison between Universities & Editors (CLUE): recommendations on best practice. 大学与编辑之间的合作与联络(CLUE):最佳做法建议。
IF 7.2 Q1 ETHICS Pub Date : 2021-04-15 DOI: 10.1186/s41073-021-00109-3
Elizabeth Wager, Sabine Kleinert

Background: Inaccurate, false or incomplete research publications may mislead readers including researchers and decision-makers. It is therefore important that such problems are identified and rectified promptly. This usually involves collaboration between the research institutions and academic journals involved, but these interactions can be problematic.

Methods: These recommendations were developed following discussions at World Conferences on Research Integrity in 2013 and 2017, and at a specially convened 3-day workshop in 2016 involving participants from 7 countries with expertise in publication ethics and research integrity. The recommendations aim to address issues surrounding cooperation and liaison between institutions (e.g. universities) and journals about possible and actual problems with the integrity of reported research arising before and after publication.

Results: The main recommendations are that research institutions should: 1) develop mechanisms for assessing the integrity of reported research (if concerns are raised) that are distinct from processes to determine whether individual researchers have committed misconduct; 2) release relevant sections of reports of research integrity or misconduct investigations to all journals that have published research that was investigated; 3) take responsibility for research performed under their auspices regardless of whether the researcher still works at that institution or how long ago the work was done; 4) work with funders to ensure essential research data is retained for at least 10 years. Journals should: 1) respond to institutions about research integrity cases in a timely manner; 2) have criteria for determining whether, and what type of, information and evidence relating to the integrity of research reports should be passed on to institutions; 3) pass on research integrity concerns to institutions, regardless of whether they intend to accept the work for publication; 4) retain peer review records for at least 10 years to enable the investigation of peer review manipulation or other inappropriate behaviour by authors or reviewers.

Conclusions: Various difficulties can prevent effective cooperation between academic journals and research institutions about research integrity concerns and hinder the correction of the research record if problems are discovered. While the issues and their solutions may vary across different settings, we encourage research institutions, journals and funders to consider how they might improve future collaboration and cooperation on research integrity cases.

背景:不准确、虚假或不完整的研究出版物可能会误导读者,包括研究人员和决策者。因此,及时发现并纠正此类问题非常重要。这通常涉及相关研究机构和学术期刊之间的合作,但这些互动可能存在问题:这些建议是在 2013 年和 2017 年世界研究诚信大会以及 2016 年专门召开的为期 3 天的研讨会上讨论后提出的,与会者来自 7 个国家,均具有出版伦理和研究诚信方面的专业知识。这些建议旨在解决机构(如大学)与期刊之间就所报告研究在发表前后可能出现和实际出现的诚信问题进行合作与联络的问题:主要建议是研究机构应1) 建立评估所报告研究诚信问题的机制(如有疑虑),该机制有别于确定研究人员是否有不当行为的程序;2) 向所有发表过受调查研究的期刊发布研究诚信或不当行为调查报告的相关部分;3) 对在其主持下开展的研究负责,无论研究人员是否仍在该机构工作,也无论该工作是在多久之前完成的;4) 与资助者合作,确保重要的研究数据至少保留 10 年。期刊应1) 及时向机构回复研究诚信案例;2) 制定标准,确定是否应向机构传递与研究报告诚信相关的信息和证据,以及应传递哪类信息和证据;3) 向机构传递研究诚信问题,无论机构是否打算接受相关作品发表;4) 保留同行评议记录至少 10 年,以便调查同行评议操纵或作者或评议人的其他不当行为:各种困难可能会阻碍学术期刊和研究机构就研究诚信问题开展有效合作,并在发现问题时阻碍对研究记录的更正。虽然不同环境下的问题及其解决方案可能各不相同,但我们鼓励研究机构、期刊和资助者考虑如何改善未来在研究诚信问题上的协作与合作。
{"title":"Cooperation & Liaison between Universities & Editors (CLUE): recommendations on best practice.","authors":"Elizabeth Wager, Sabine Kleinert","doi":"10.1186/s41073-021-00109-3","DOIUrl":"10.1186/s41073-021-00109-3","url":null,"abstract":"<p><strong>Background: </strong>Inaccurate, false or incomplete research publications may mislead readers including researchers and decision-makers. It is therefore important that such problems are identified and rectified promptly. This usually involves collaboration between the research institutions and academic journals involved, but these interactions can be problematic.</p><p><strong>Methods: </strong>These recommendations were developed following discussions at World Conferences on Research Integrity in 2013 and 2017, and at a specially convened 3-day workshop in 2016 involving participants from 7 countries with expertise in publication ethics and research integrity. The recommendations aim to address issues surrounding cooperation and liaison between institutions (e.g. universities) and journals about possible and actual problems with the integrity of reported research arising before and after publication.</p><p><strong>Results: </strong>The main recommendations are that research institutions should: 1) develop mechanisms for assessing the integrity of reported research (if concerns are raised) that are distinct from processes to determine whether individual researchers have committed misconduct; 2) release relevant sections of reports of research integrity or misconduct investigations to all journals that have published research that was investigated; 3) take responsibility for research performed under their auspices regardless of whether the researcher still works at that institution or how long ago the work was done; 4) work with funders to ensure essential research data is retained for at least 10 years. Journals should: 1) respond to institutions about research integrity cases in a timely manner; 2) have criteria for determining whether, and what type of, information and evidence relating to the integrity of research reports should be passed on to institutions; 3) pass on research integrity concerns to institutions, regardless of whether they intend to accept the work for publication; 4) retain peer review records for at least 10 years to enable the investigation of peer review manipulation or other inappropriate behaviour by authors or reviewers.</p><p><strong>Conclusions: </strong>Various difficulties can prevent effective cooperation between academic journals and research institutions about research integrity concerns and hinder the correction of the research record if problems are discovered. While the issues and their solutions may vary across different settings, we encourage research institutions, journals and funders to consider how they might improve future collaboration and cooperation on research integrity cases.</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"6"},"PeriodicalIF":7.2,"publicationDate":"2021-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8048029/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25590216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Steps toward preregistration of research on research integrity. 研究诚信研究的预先登记步骤。
Q1 ETHICS Pub Date : 2021-03-01 DOI: 10.1186/s41073-021-00108-4
Klaas Sijtsma, Wilco H M Emons, Nicholas H Steneck, Lex M Bouter

Background: A proposal to encourage the preregistration of research on research integrity was developed and adopted as the Amsterdam Agenda at the 5th World Conference on Research Integrity (Amsterdam, 2017). This paper reports on the degree to which abstracts of the 6th World Conference in Research Integrity (Hong Kong, 2019) reported on preregistered research.

Methods: Conference registration data on participants presenting a paper or a poster at 6th WCRI were made available to the research team. Because the data set was too small for inferential statistics this report is limited to a basic description of results and some recommendations that should be considered when taking further steps to improve preregistration.

Results: 19% of the 308 presenters preregistered their research. Of the 56 usable cases, less than half provided information on the six key elements of the Amsterdam Agenda. Others provided information that invalidated their data, such as an uninformative URL. There was no discernable difference between qualitative and quantitative research.

Conclusions: Some presenters at the WCRI have preregistered their research on research integrity, but further steps are needed to increase frequency and completeness of preregistration. One approach to increase preregistration would be to make it a requirement for research presented at the World Conferences on Research Integrity.

背景:第五届世界科研诚信大会(阿姆斯特丹,2017 年)制定并通过了一项鼓励对科研诚信研究进行预先登记的提案,并将其作为《阿姆斯特丹议程》。本文报告了第六届世界科研诚信大会(香港,2019年)的摘要在多大程度上报告了预先注册的研究:研究小组获得了在第六届世界科研诚信大会上提交论文或海报的与会者的大会注册数据。由于数据集太小,无法进行推理统计,因此本报告仅限于对结果的基本描述,以及在采取进一步措施改进预登记时应考虑的一些建议:在 308 位发言人中,有 19% 的人对其研究进行了预先登记。在 56 个可用的案例中,不到一半提供了有关《阿姆斯特丹议程》六个关键要素的信息。还有一些人提供的信息使他们的数据失效,比如一个信息不全的 URL。定性研究和定量研究之间没有明显的区别:世界研究与创新会议上的一些发言人已经预先登记了他们关于研究诚信的研究,但还需要采取进一步措施,提高预先登记的频率和完整性。增加预登记的一种方法是将其作为在世界研究诚信大会上展示研究成果的一项要求。
{"title":"Steps toward preregistration of research on research integrity.","authors":"Klaas Sijtsma, Wilco H M Emons, Nicholas H Steneck, Lex M Bouter","doi":"10.1186/s41073-021-00108-4","DOIUrl":"10.1186/s41073-021-00108-4","url":null,"abstract":"<p><strong>Background: </strong>A proposal to encourage the preregistration of research on research integrity was developed and adopted as the Amsterdam Agenda at the 5th World Conference on Research Integrity (Amsterdam, 2017). This paper reports on the degree to which abstracts of the 6th World Conference in Research Integrity (Hong Kong, 2019) reported on preregistered research.</p><p><strong>Methods: </strong>Conference registration data on participants presenting a paper or a poster at 6th WCRI were made available to the research team. Because the data set was too small for inferential statistics this report is limited to a basic description of results and some recommendations that should be considered when taking further steps to improve preregistration.</p><p><strong>Results: </strong>19% of the 308 presenters preregistered their research. Of the 56 usable cases, less than half provided information on the six key elements of the Amsterdam Agenda. Others provided information that invalidated their data, such as an uninformative URL. There was no discernable difference between qualitative and quantitative research.</p><p><strong>Conclusions: </strong>Some presenters at the WCRI have preregistered their research on research integrity, but further steps are needed to increase frequency and completeness of preregistration. One approach to increase preregistration would be to make it a requirement for research presented at the World Conferences on Research Integrity.</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"5"},"PeriodicalIF":0.0,"publicationDate":"2021-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7923522/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25425863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Re-evaluation of solutions to the problem of unprofessionalism in peer review. 重新评估同行评议中不专业问题的解决方案。
Q1 ETHICS Pub Date : 2021-02-16 DOI: 10.1186/s41073-020-00107-x
Travis G Gerwing, Alyssa M Allen Gerwing, Chi-Yeung Choi, Stephanie Avery-Gomm, Jeff C Clements, Joshua A Rash

Our recent paper ( https://doi.org/10.1186/s41073-020-00096-x ) reported that 43% of reviewer comment sets (n=1491) shared with authors contained at least one unprofessional comment or an incomplete, inaccurate of unsubstantiated critique (IIUC). Publication of this work sparked an online (i.e., Twitter, Instagram, Facebook, and Reddit) conversation surrounding professionalism in peer review. We collected and analyzed these social media comments as they offered real-time responses to our work and provided insight into the views held by commenters and potential peer-reviewers that would be difficult to quantify using existing empirical tools (96 comments from July 24th to September 3rd, 2020). Overall, 75% of comments were positive, of which 59% were supportive and 16% shared similar personal experiences. However, a subset of negative comments emerged (22% of comments were negative and 6% were an unsubstantiated critique of the methodology), that provided potential insight into the reasons underlying unprofessional comments were made during the peer-review process. These comments were classified into three main themes: (1) forced niceness will adversely impact the peer-review process and allow for publication of poor-quality science (5% of online comments); (2) dismissing comments as not offensive to another person because they were not deemed personally offensive to the reader (6%); and (3) authors brought unprofessional comments upon themselves as they submitted substandard work (5%). Here, we argue against these themes as justifications for directing unprofessional comments towards authors during the peer review process. We argue that it is possible to be both critical and professional, and that no author deserves to be the recipient of demeaning ad hominem attacks regardless of supposed provocation. Suggesting otherwise only serves to propagate a toxic culture within peer review. While we previously postulated that establishing a peer-reviewer code of conduct could help improve the peer-review system, we now posit that priority should be given to repairing the negative cultural zeitgeist that exists in peer-review.

我们最近的一篇论文(https://doi.org/10.1186/s41073-020-00096-x)报道,43%的审稿人评论集(n=1491)与作者共享,至少包含一条不专业的评论或不完整、不准确或未经证实的评论(IIUC)。这项研究的发表引发了一场围绕同行评议专业性的在线讨论(即Twitter、Instagram、Facebook和Reddit)。我们收集并分析了这些社交媒体评论,因为它们对我们的工作提供了实时回应,并提供了评论者和潜在同行审稿人持有的观点的见解,这些观点很难用现有的经验工具量化(2020年7月24日至9月3日的96条评论)。总的来说,75%的评论是积极的,其中59%是支持的,16%的人分享了类似的个人经历。然而,出现了一部分负面评论(22%的评论是负面的,6%是对方法的未经证实的批评),这为在同行评审过程中做出不专业评论的潜在原因提供了潜在的见解。这些评论被分为三个主要主题:(1)强迫的善意将对同行评议过程产生不利影响,并允许发表低质量的科学(占在线评论的5%);(2)因为不认为评论冒犯了读者个人,所以将其视为不冒犯他人而不予理会(6%);(3)作者因提交不合格作品而给自己带来不专业的评论(5%)。在这里,我们反对将这些主题作为在同行评审过程中对作者进行不专业评论的理由。我们认为,批评和专业是可能的,没有一个作者应该受到贬低人身攻击,而不管假设的挑衅。相反的建议只会在同行评议中传播一种有害的文化。虽然我们之前假设建立同行评议行为准则可以帮助改善同行评议制度,但我们现在假设应该优先修复同行评议中存在的负面文化时代精神。
{"title":"Re-evaluation of solutions to the problem of unprofessionalism in peer review.","authors":"Travis G Gerwing,&nbsp;Alyssa M Allen Gerwing,&nbsp;Chi-Yeung Choi,&nbsp;Stephanie Avery-Gomm,&nbsp;Jeff C Clements,&nbsp;Joshua A Rash","doi":"10.1186/s41073-020-00107-x","DOIUrl":"https://doi.org/10.1186/s41073-020-00107-x","url":null,"abstract":"<p><p>Our recent paper ( https://doi.org/10.1186/s41073-020-00096-x ) reported that 43% of reviewer comment sets (n=1491) shared with authors contained at least one unprofessional comment or an incomplete, inaccurate of unsubstantiated critique (IIUC). Publication of this work sparked an online (i.e., Twitter, Instagram, Facebook, and Reddit) conversation surrounding professionalism in peer review. We collected and analyzed these social media comments as they offered real-time responses to our work and provided insight into the views held by commenters and potential peer-reviewers that would be difficult to quantify using existing empirical tools (96 comments from July 24th to September 3rd, 2020). Overall, 75% of comments were positive, of which 59% were supportive and 16% shared similar personal experiences. However, a subset of negative comments emerged (22% of comments were negative and 6% were an unsubstantiated critique of the methodology), that provided potential insight into the reasons underlying unprofessional comments were made during the peer-review process. These comments were classified into three main themes: (1) forced niceness will adversely impact the peer-review process and allow for publication of poor-quality science (5% of online comments); (2) dismissing comments as not offensive to another person because they were not deemed personally offensive to the reader (6%); and (3) authors brought unprofessional comments upon themselves as they submitted substandard work (5%). Here, we argue against these themes as justifications for directing unprofessional comments towards authors during the peer review process. We argue that it is possible to be both critical and professional, and that no author deserves to be the recipient of demeaning ad hominem attacks regardless of supposed provocation. Suggesting otherwise only serves to propagate a toxic culture within peer review. While we previously postulated that establishing a peer-reviewer code of conduct could help improve the peer-review system, we now posit that priority should be given to repairing the negative cultural zeitgeist that exists in peer-review.</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"4"},"PeriodicalIF":0.0,"publicationDate":"2021-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41073-020-00107-x","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25375244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Estimating the prevalence of text overlap in biomedical conference abstracts. 估计生物医学会议摘要中文本重叠的普遍性。
Q1 ETHICS Pub Date : 2021-02-01 DOI: 10.1186/s41073-020-00106-y
Nick Kinney, Araba Wubah, Miguel Roig, Harold R Garner

Background: Scientists communicate progress and exchange information via publication and presentation at scientific meetings. We previously showed that text similarity analysis applied to Medline can identify and quantify plagiarism and duplicate publications in peer-reviewed biomedical journals. In the present study, we applied the same analysis to a large sample of conference abstracts.

Methods: We downloaded 144,149 abstracts from 207 national and international meetings of 63 biomedical conferences. Pairwise comparisons were made using eTBLAST: a text similarity engine. A domain expert then reviewed random samples of highly similar abstracts (1500 total) to estimate the extent of text overlap and possible plagiarism.

Results: Our main findings indicate that the vast majority of textual overlap occurred within the same meeting (2%) and between meetings of the same conference (3%), both of which were significantly higher than instances of plagiarism, which occurred in less than .5% of abstracts.

Conclusions: This analysis indicates that textual overlap in abstracts of papers presented at scientific meetings is one-tenth that of peer-reviewed publications, yet the plagiarism rate is approximately the same as previously measured in peer-reviewed publications. This latter finding underscores a need for monitoring scientific meeting submissions - as is now done when submitting manuscripts to peer-reviewed journals - to improve the integrity of scientific communications.

背景:科学家通过出版物和在科学会议上的展示来交流进展和信息。我们之前的研究表明,应用于Medline的文本相似度分析可以识别和量化同行评议的生物医学期刊中的剽窃和重复出版物。在本研究中,我们将相同的分析应用于大量会议摘要样本。方法:我们下载了63个生物医学会议的207个国内和国际会议的144,149篇摘要。两两比较使用eTBLAST:一个文本相似度引擎。然后,一位领域专家回顾了高度相似的摘要的随机样本(总共1500篇),以估计文本重叠的程度和可能的抄袭。结果:我们的主要发现表明,绝大多数文本重叠发生在同一会议(2%)和同一会议的会议之间(3%),这两种情况都明显高于剽窃的情况,抄袭发生率不到0.5%。结论:该分析表明,在科学会议上发表的论文摘要中的文本重叠是同行评议出版物的十分之一,但剽窃率与之前在同行评议出版物中测量的大致相同。后一项发现强调了监督科学会议提交的必要性——就像现在在向同行评议期刊提交手稿时所做的那样——以提高科学传播的完整性。
{"title":"Estimating the prevalence of text overlap in biomedical conference abstracts.","authors":"Nick Kinney,&nbsp;Araba Wubah,&nbsp;Miguel Roig,&nbsp;Harold R Garner","doi":"10.1186/s41073-020-00106-y","DOIUrl":"https://doi.org/10.1186/s41073-020-00106-y","url":null,"abstract":"<p><strong>Background: </strong>Scientists communicate progress and exchange information via publication and presentation at scientific meetings. We previously showed that text similarity analysis applied to Medline can identify and quantify plagiarism and duplicate publications in peer-reviewed biomedical journals. In the present study, we applied the same analysis to a large sample of conference abstracts.</p><p><strong>Methods: </strong>We downloaded 144,149 abstracts from 207 national and international meetings of 63 biomedical conferences. Pairwise comparisons were made using eTBLAST: a text similarity engine. A domain expert then reviewed random samples of highly similar abstracts (1500 total) to estimate the extent of text overlap and possible plagiarism.</p><p><strong>Results: </strong>Our main findings indicate that the vast majority of textual overlap occurred within the same meeting (2%) and between meetings of the same conference (3%), both of which were significantly higher than instances of plagiarism, which occurred in less than .5% of abstracts.</p><p><strong>Conclusions: </strong>This analysis indicates that textual overlap in abstracts of papers presented at scientific meetings is one-tenth that of peer-reviewed publications, yet the plagiarism rate is approximately the same as previously measured in peer-reviewed publications. This latter finding underscores a need for monitoring scientific meeting submissions - as is now done when submitting manuscripts to peer-reviewed journals - to improve the integrity of scientific communications.</p>","PeriodicalId":74682,"journal":{"name":"Research integrity and peer review","volume":"6 1","pages":"2"},"PeriodicalIF":0.0,"publicationDate":"2021-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1186/s41073-020-00106-y","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25313727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Research integrity and peer review
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1