Pub Date : 2024-09-18DOI: 10.1186/s41235-024-00595-1
Fabian Hutmacher, Markus Appel, Benjamin Schätzlein, Christoph Mengelkamp
Misinformation can profoundly impact an individual's attitudes-sometimes even after the misinformation has been corrected. In two preregistered experiments (N1 = 355, N2 = 725), we investigated whether individual differences in the ability and motivation to process information thoroughly influence the impact of misinformation in a news media context. More specifically, we tested whether fluid intelligence and need for cognition predicted the degree to which individuals who were exposed to misinformation changed their attitudes after receiving a correction message. We found consistent evidence that higher fluid intelligence is associated with a more pronounced correction effect, while need for cognition did not have a significant effect. This suggests that integrating a correction message with a previously encountered piece of misinformation can be challenging and that correction messages consequently need to be communicated in a way that is accessible to a broad audience.
{"title":"Fluid intelligence but not need for cognition is associated with attitude change in response to the correction of misinformation.","authors":"Fabian Hutmacher, Markus Appel, Benjamin Schätzlein, Christoph Mengelkamp","doi":"10.1186/s41235-024-00595-1","DOIUrl":"https://doi.org/10.1186/s41235-024-00595-1","url":null,"abstract":"<p><p>Misinformation can profoundly impact an individual's attitudes-sometimes even after the misinformation has been corrected. In two preregistered experiments (N<sub>1</sub> = 355, N<sub>2</sub> = 725), we investigated whether individual differences in the ability and motivation to process information thoroughly influence the impact of misinformation in a news media context. More specifically, we tested whether fluid intelligence and need for cognition predicted the degree to which individuals who were exposed to misinformation changed their attitudes after receiving a correction message. We found consistent evidence that higher fluid intelligence is associated with a more pronounced correction effect, while need for cognition did not have a significant effect. This suggests that integrating a correction message with a previously encountered piece of misinformation can be challenging and that correction messages consequently need to be communicated in a way that is accessible to a broad audience.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"64"},"PeriodicalIF":3.4,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11411052/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-13DOI: 10.1186/s41235-024-00589-z
Emma Smillie, Natalie Mestry, Dan Clark, Neil Harrison, Nick Donnelly
Two experiments explored the search for pairs of faces in a disjunctive dual-target face search (DDTFS) task for unfamiliar face targets. The distinctiveness of the target was manipulated such that both faces were typical or distinctive or contained one typical and one distinctive target. Targets were searched for in arrays of eight faces. In Experiment 1, participants completed a DDTFS block with targets learnt over the block of trials. In Experiment 2, the dual-target block was preceded by two training blocks of single-target trials. Participants also completed the upright and inverted long-form Cambridge Face Memory Test (CFMT+). The results showed that searching for two typical faces leads to one target being prioritised at the expense of the other. The ability to search for non-prioritised typical faces was associated with scores on the CFMT+. This association disappeared when faces were learnt before completing DDTFS. We interpret the findings in terms of the impact of typicality on face learning, individual differences in the ability to learn faces, and the involvement of capacity-limited working memory in the search for unfamiliar faces. The findings have implications for security-related situations where agents must search for multiple unfamiliar faces having been shown their images.
{"title":"The role of facial distinctiveness in the prioritisation of targets in disjunctive dual-target face search.","authors":"Emma Smillie, Natalie Mestry, Dan Clark, Neil Harrison, Nick Donnelly","doi":"10.1186/s41235-024-00589-z","DOIUrl":"https://doi.org/10.1186/s41235-024-00589-z","url":null,"abstract":"<p><p>Two experiments explored the search for pairs of faces in a disjunctive dual-target face search (DDTFS) task for unfamiliar face targets. The distinctiveness of the target was manipulated such that both faces were typical or distinctive or contained one typical and one distinctive target. Targets were searched for in arrays of eight faces. In Experiment 1, participants completed a DDTFS block with targets learnt over the block of trials. In Experiment 2, the dual-target block was preceded by two training blocks of single-target trials. Participants also completed the upright and inverted long-form Cambridge Face Memory Test (CFMT+). The results showed that searching for two typical faces leads to one target being prioritised at the expense of the other. The ability to search for non-prioritised typical faces was associated with scores on the CFMT+. This association disappeared when faces were learnt before completing DDTFS. We interpret the findings in terms of the impact of typicality on face learning, individual differences in the ability to learn faces, and the involvement of capacity-limited working memory in the search for unfamiliar faces. The findings have implications for security-related situations where agents must search for multiple unfamiliar faces having been shown their images.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"62"},"PeriodicalIF":3.4,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11399498/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1186/s41235-024-00591-5
Monika Lohani, Joel M Cooper, Amy S McDonnell, Gus G Erickson, Trent G Simmons, Amanda E Carriero, Kaedyn W Crabtree, David L Strayer
The reliability of cognitive demand measures in controlled laboratory settings is well-documented; however, limited research has directly established their stability under real-life and high-stakes conditions, such as operating automated technology on actual highways. Partially automated vehicles have advanced to become an everyday mode of transportation, and research on driving these advanced vehicles requires reliable tools for evaluating the cognitive demand on motorists to sustain optimal engagement in the driving process. This study examined the reliability of five cognitive demand measures, while participants operated partially automated vehicles on real roads across four occasions. Seventy-one participants (aged 18-64 years) drove on actual highways while their heart rate, heart rate variability, electroencephalogram (EEG) alpha power, and behavioral performance on the Detection Response Task were measured simultaneously. Findings revealed that EEG alpha power had excellent test-retest reliability, heart rate and its variability were good, and Detection Response Task reaction time and hit-rate had moderate reliabilities. Thus, the current study addresses concerns regarding the reliability of these measures in assessing cognitive demand in real-world automation research, as acceptable test-retest reliabilities were found across all measures for drivers across occasions. Despite the high reliability of each measure, low intercorrelations among measures were observed, and internal consistency was better when cognitive demand was estimated as a multi-factorial construct. This suggests that they tap into different aspects of cognitive demand while operating automation in real life. The findings highlight that a combination of psychophysiological and behavioral methods can reliably capture multi-faceted cognitive demand in real-world automation research.
{"title":"Reliable but multi-dimensional cognitive demand in operating partially automated vehicles: implications for real-world automation research.","authors":"Monika Lohani, Joel M Cooper, Amy S McDonnell, Gus G Erickson, Trent G Simmons, Amanda E Carriero, Kaedyn W Crabtree, David L Strayer","doi":"10.1186/s41235-024-00591-5","DOIUrl":"https://doi.org/10.1186/s41235-024-00591-5","url":null,"abstract":"<p><p>The reliability of cognitive demand measures in controlled laboratory settings is well-documented; however, limited research has directly established their stability under real-life and high-stakes conditions, such as operating automated technology on actual highways. Partially automated vehicles have advanced to become an everyday mode of transportation, and research on driving these advanced vehicles requires reliable tools for evaluating the cognitive demand on motorists to sustain optimal engagement in the driving process. This study examined the reliability of five cognitive demand measures, while participants operated partially automated vehicles on real roads across four occasions. Seventy-one participants (aged 18-64 years) drove on actual highways while their heart rate, heart rate variability, electroencephalogram (EEG) alpha power, and behavioral performance on the Detection Response Task were measured simultaneously. Findings revealed that EEG alpha power had excellent test-retest reliability, heart rate and its variability were good, and Detection Response Task reaction time and hit-rate had moderate reliabilities. Thus, the current study addresses concerns regarding the reliability of these measures in assessing cognitive demand in real-world automation research, as acceptable test-retest reliabilities were found across all measures for drivers across occasions. Despite the high reliability of each measure, low intercorrelations among measures were observed, and internal consistency was better when cognitive demand was estimated as a multi-factorial construct. This suggests that they tap into different aspects of cognitive demand while operating automation in real life. The findings highlight that a combination of psychophysiological and behavioral methods can reliably capture multi-faceted cognitive demand in real-world automation research.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"60"},"PeriodicalIF":3.4,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11387569/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298483","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1186/s41235-024-00580-8
Marina Pi-Ruano, Alexandra Fort, Pilar Tejero, Christophe Jallais, Javier Roca
Partially autonomous vehicles can help minimize human errors. However, being free from some driving subtasks can result in a low vigilance state, which can affect the driver's attention towards the road. The present study first tested whether drivers of partially autonomous vehicles would benefit from the addition of auditory versions of the messages presented in variable message signs (VMS), particularly, when they find themselves in a monotonous driving situation. A second aim was to test whether the addition of auditory messages would also produce an indirect effect on the driver's vigilance, improving performance on other driving subtasks not related to the message processing. Forty-three volunteers participated in a driving simulator study. They completed two tasks: (a) a VMS task, where they had to regain manual control of the car if the VMS message was critical, and (b) a car-following task, where they had to pay attention to the preceding car to respond to occasional brake events. Behavioral and EEG data were registered. Overall, results indicated that the addition of audio messages helped drivers process VMS information more effectively and maintain a higher level of vigilance throughout the driving time. These findings would provide useful information for the development of partially automated vehicles, as their design must guarantee that the driver remains attentive enough to assume control when necessary.
{"title":"Audiovisual messages may improve the processing of traffic information and driver attention during partially automated driving: An EEG study.","authors":"Marina Pi-Ruano, Alexandra Fort, Pilar Tejero, Christophe Jallais, Javier Roca","doi":"10.1186/s41235-024-00580-8","DOIUrl":"https://doi.org/10.1186/s41235-024-00580-8","url":null,"abstract":"<p><p>Partially autonomous vehicles can help minimize human errors. However, being free from some driving subtasks can result in a low vigilance state, which can affect the driver's attention towards the road. The present study first tested whether drivers of partially autonomous vehicles would benefit from the addition of auditory versions of the messages presented in variable message signs (VMS), particularly, when they find themselves in a monotonous driving situation. A second aim was to test whether the addition of auditory messages would also produce an indirect effect on the driver's vigilance, improving performance on other driving subtasks not related to the message processing. Forty-three volunteers participated in a driving simulator study. They completed two tasks: (a) a VMS task, where they had to regain manual control of the car if the VMS message was critical, and (b) a car-following task, where they had to pay attention to the preceding car to respond to occasional brake events. Behavioral and EEG data were registered. Overall, results indicated that the addition of audio messages helped drivers process VMS information more effectively and maintain a higher level of vigilance throughout the driving time. These findings would provide useful information for the development of partially automated vehicles, as their design must guarantee that the driver remains attentive enough to assume control when necessary.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"61"},"PeriodicalIF":3.4,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11387282/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-02DOI: 10.1186/s41235-024-00584-4
Jerrick Teoh, Joseph M Saito, Yvanna Yeo, Sophia Winter, Keisuke Fukuda
Humans are often tasked to remember new faces so that they can recognize the faces later in time. Previous studies found that memory reports for basic visual features (e.g., colors and shapes) are susceptible to systematic distortions as a result of comparison with new visual input, especially when the input is perceived as similar to the memory. The current study tested whether this similarity-induced memory bias (SIMB) would also occur with more complex face stimuli. The results showed that faces that are just perceptually encoded into visual working memory as well as retrieved from visual long-term memory are also susceptible to SIMB. Furthermore, once induced, SIMB persisted over time across cues through which the face memory was accessed for memory report. These results demonstrate the generalizability of SIMB to more complex and practically relevant stimuli, and thus, suggest potential real-world implications.
{"title":"Perceptual comparisons induce lasting and generalizing changes to face memory reports.","authors":"Jerrick Teoh, Joseph M Saito, Yvanna Yeo, Sophia Winter, Keisuke Fukuda","doi":"10.1186/s41235-024-00584-4","DOIUrl":"10.1186/s41235-024-00584-4","url":null,"abstract":"<p><p>Humans are often tasked to remember new faces so that they can recognize the faces later in time. Previous studies found that memory reports for basic visual features (e.g., colors and shapes) are susceptible to systematic distortions as a result of comparison with new visual input, especially when the input is perceived as similar to the memory. The current study tested whether this similarity-induced memory bias (SIMB) would also occur with more complex face stimuli. The results showed that faces that are just perceptually encoded into visual working memory as well as retrieved from visual long-term memory are also susceptible to SIMB. Furthermore, once induced, SIMB persisted over time across cues through which the face memory was accessed for memory report. These results demonstrate the generalizability of SIMB to more complex and practically relevant stimuli, and thus, suggest potential real-world implications.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"57"},"PeriodicalIF":3.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11366729/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142113260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-02DOI: 10.1186/s41235-024-00576-4
Francesca Patterson, Melina A Kunar
Computer Aided Detection (CAD) has been used to help readers find cancers in mammograms. Although these automated systems have been shown to help cancer detection when accurate, the presence of CAD also leads to an over-reliance effect where miss errors and false alarms increase when the CAD system fails. Previous research investigated CAD systems which overlayed salient exogenous cues onto the image to highlight suspicious areas. These salient cues capture attention which may exacerbate the over-reliance effect. Furthermore, overlaying CAD cues directly on the mammogram occludes sections of breast tissue which may disrupt global statistics useful for cancer detection. In this study we investigated whether an over-reliance effect occurred with a binary CAD system, which instead of overlaying a CAD cue onto the mammogram, reported a message alongside the mammogram indicating the possible presence of a cancer. We manipulated the certainty of the message and whether it was presented only to indicate the presence of a cancer, or whether a message was displayed on every mammogram to state whether a cancer was present or absent. The results showed that although an over-reliance effect still occurred with binary CAD systems miss errors were reduced when the CAD message was more definitive and only presented to alert readers of a possible cancer.
计算机辅助检测(CAD)已被用于帮助读者发现乳房 X 光照片中的癌症。虽然这些自动系统在准确时有助于癌症检测,但 CAD 的存在也会导致过度依赖效应,即当 CAD 系统失效时,漏检错误和误报会增加。以前的研究调查了将突出的外来线索叠加到图像上以突出可疑区域的 CAD 系统。这些突出线索会吸引注意力,从而加剧过度依赖效应。此外,将 CAD 提示直接叠加到乳房 X 光照片上会遮挡部分乳腺组织,这可能会破坏对癌症检测有用的全局统计数据。在这项研究中,我们研究了二元 CAD 系统是否会产生过度依赖效应,该系统不是将 CAD 提示叠加到乳房 X 光照片上,而是在乳房 X 光照片旁边报告一条信息,表明可能存在癌症。我们对信息的确定性进行了调整,并调整了信息的显示方式,是仅显示是否存在癌症,还是在每张乳房 X 射线照片上都显示一条信息,说明是否存在癌症。结果表明,虽然二进制 CAD 系统仍会产生过度依赖效应,但当 CAD 信息更加确定且仅用于提醒读者可能存在癌症时,失误率就会降低。
{"title":"The message matters: changes to binary Computer Aided Detection recommendations affect cancer detection in low prevalence search.","authors":"Francesca Patterson, Melina A Kunar","doi":"10.1186/s41235-024-00576-4","DOIUrl":"10.1186/s41235-024-00576-4","url":null,"abstract":"<p><p>Computer Aided Detection (CAD) has been used to help readers find cancers in mammograms. Although these automated systems have been shown to help cancer detection when accurate, the presence of CAD also leads to an over-reliance effect where miss errors and false alarms increase when the CAD system fails. Previous research investigated CAD systems which overlayed salient exogenous cues onto the image to highlight suspicious areas. These salient cues capture attention which may exacerbate the over-reliance effect. Furthermore, overlaying CAD cues directly on the mammogram occludes sections of breast tissue which may disrupt global statistics useful for cancer detection. In this study we investigated whether an over-reliance effect occurred with a binary CAD system, which instead of overlaying a CAD cue onto the mammogram, reported a message alongside the mammogram indicating the possible presence of a cancer. We manipulated the certainty of the message and whether it was presented only to indicate the presence of a cancer, or whether a message was displayed on every mammogram to state whether a cancer was present or absent. The results showed that although an over-reliance effect still occurred with binary CAD systems miss errors were reduced when the CAD message was more definitive and only presented to alert readers of a possible cancer.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"59"},"PeriodicalIF":3.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11366737/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142113261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-02DOI: 10.1186/s41235-024-00583-5
Jonathon Love, Quentin F Gronau, Gemma Palmer, Ami Eidels, Scott D Brown
With the growing role of artificial intelligence (AI) in our lives, attention is increasingly turning to the way that humans and AI work together. A key aspect of human-AI collaboration is how people integrate judgements or recommendations from machine agents, when they differ from their own judgements. We investigated trust in human-machine teaming using a perceptual judgement task based on the judge-advisor system. Participants ( ) estimated a perceptual quantity, then received a recommendation from a machine agent. The participants then made a second response which combined their first estimate and the machine's recommendation. The degree to which participants shifted their second response in the direction of the recommendations provided a measure of their trust in the machine agent. We analysed the role of advice distance in people's willingness to change their judgements. When a recommendation falls a long way from their initial judgement, do people come to doubt their own judgement, trusting the recommendation more, or do they doubt the machine agent, trusting the recommendation less? We found that although some participants exhibited these behaviours, the most common response was neither of these tendencies, and a simple model based on averaging accounted best for participants' trust behaviour. We discuss implications for theories of trust, and human-machine teaming.
{"title":"In human-machine trust, humans rely on a simple averaging strategy.","authors":"Jonathon Love, Quentin F Gronau, Gemma Palmer, Ami Eidels, Scott D Brown","doi":"10.1186/s41235-024-00583-5","DOIUrl":"10.1186/s41235-024-00583-5","url":null,"abstract":"<p><p>With the growing role of artificial intelligence (AI) in our lives, attention is increasingly turning to the way that humans and AI work together. A key aspect of human-AI collaboration is how people integrate judgements or recommendations from machine agents, when they differ from their own judgements. We investigated trust in human-machine teaming using a perceptual judgement task based on the judge-advisor system. Participants ( <math><mrow><mi>n</mi> <mo>=</mo> <mn>89</mn></mrow> </math> ) estimated a perceptual quantity, then received a recommendation from a machine agent. The participants then made a second response which combined their first estimate and the machine's recommendation. The degree to which participants shifted their second response in the direction of the recommendations provided a measure of their trust in the machine agent. We analysed the role of advice distance in people's willingness to change their judgements. When a recommendation falls a long way from their initial judgement, do people come to doubt their own judgement, trusting the recommendation more, or do they doubt the machine agent, trusting the recommendation less? We found that although some participants exhibited these behaviours, the most common response was neither of these tendencies, and a simple model based on averaging accounted best for participants' trust behaviour. We discuss implications for theories of trust, and human-machine teaming.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"58"},"PeriodicalIF":3.4,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11366733/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142113259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-26DOI: 10.1186/s41235-024-00586-2
Laura E Matzen, Zoe N Gastelum, Breannan C Howell, Kristin M Divis, Mallory C Stites
This study addressed the cognitive impacts of providing correct and incorrect machine learning (ML) outputs in support of an object detection task. The study consisted of five experiments that manipulated the accuracy and importance of mock ML outputs. In each of the experiments, participants were given the T and L task with T-shaped targets and L-shaped distractors. They were tasked with categorizing each image as target present or target absent. In Experiment 1, they performed this task without the aid of ML outputs. In Experiments 2-5, they were shown images with bounding boxes, representing the output of an ML model. The outputs could be correct (hits and correct rejections), or they could be erroneous (false alarms and misses). Experiment 2 manipulated the overall accuracy of these mock ML outputs. Experiment 3 manipulated the proportion of different types of errors. Experiments 4 and 5 manipulated the importance of specific types of stimuli or model errors, as well as the framing of the task in terms of human or model performance. These experiments showed that model misses were consistently harder for participants to detect than model false alarms. In general, as the model's performance increased, human performance increased as well, but in many cases the participants were more likely to overlook model errors when the model had high accuracy overall. Warning participants to be on the lookout for specific types of model errors had very little impact on their performance. Overall, our results emphasize the importance of considering human cognition when determining what level of model performance and types of model errors are acceptable for a given task.
本研究探讨了提供正确和错误的机器学习(ML)输出以支持物体检测任务对认知的影响。研究包括五项实验,对模拟 ML 输出的准确性和重要性进行了操作。在每个实验中,参与者都会接到 T 形目标和 L 形干扰物的 T 和 L 任务。他们的任务是将每幅图像归类为目标存在或目标不存在。在实验 1 中,他们在没有 ML 输出辅助的情况下完成了这项任务。在实验 2-5 中,他们看到的图像带有边界框,代表了 ML 模型的输出结果。输出结果可能是正确的(命中和正确拒绝),也可能是错误的(误报和漏报)。实验 2 操作了这些模拟 ML 输出的总体准确性。实验 3 控制了不同类型错误的比例。实验 4 和 5 对特定类型的刺激或模型错误的重要性,以及从人类或模型性能的角度对任务的框架进行了操作。这些实验表明,对于参与者来说,模型失误始终比模型误报更难发现。一般来说,随着模型性能的提高,人类的性能也会随之提高,但在许多情况下,当模型整体准确率较高时,参与者更容易忽略模型错误。警告参与者注意特定类型的模型错误对他们的表现影响很小。总之,我们的研究结果强调了在确定特定任务可接受的模型性能水平和模型错误类型时考虑人类认知的重要性。
{"title":"Effects of machine learning errors on human decision-making: manipulations of model accuracy, error types, and error importance.","authors":"Laura E Matzen, Zoe N Gastelum, Breannan C Howell, Kristin M Divis, Mallory C Stites","doi":"10.1186/s41235-024-00586-2","DOIUrl":"10.1186/s41235-024-00586-2","url":null,"abstract":"<p><p>This study addressed the cognitive impacts of providing correct and incorrect machine learning (ML) outputs in support of an object detection task. The study consisted of five experiments that manipulated the accuracy and importance of mock ML outputs. In each of the experiments, participants were given the T and L task with T-shaped targets and L-shaped distractors. They were tasked with categorizing each image as target present or target absent. In Experiment 1, they performed this task without the aid of ML outputs. In Experiments 2-5, they were shown images with bounding boxes, representing the output of an ML model. The outputs could be correct (hits and correct rejections), or they could be erroneous (false alarms and misses). Experiment 2 manipulated the overall accuracy of these mock ML outputs. Experiment 3 manipulated the proportion of different types of errors. Experiments 4 and 5 manipulated the importance of specific types of stimuli or model errors, as well as the framing of the task in terms of human or model performance. These experiments showed that model misses were consistently harder for participants to detect than model false alarms. In general, as the model's performance increased, human performance increased as well, but in many cases the participants were more likely to overlook model errors when the model had high accuracy overall. Warning participants to be on the lookout for specific types of model errors had very little impact on their performance. Overall, our results emphasize the importance of considering human cognition when determining what level of model performance and types of model errors are acceptable for a given task.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"56"},"PeriodicalIF":3.4,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11345344/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142056909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-26DOI: 10.1186/s41235-024-00577-3
Emily N Line, Sara Jaramillo, Micah Goldwater, Zachary Horne
People are inundated with popular press reports about medical research concerning what is healthy, get advice from doctors, and hear personal anecdotes. How do people integrate conflicting anecdotal and statistical information when making medical decisions? In four experiments (N = 4126), we tested how people use conflicting information to judge the efficacy of artificial and real medical treatments. Participants read an anecdote from someone in a clinical trial, or who had undergone a medical treatment previously, for whom the medical treatment was ineffective. We found that reading anecdotes for either artificial or real medical treatments shifted participants' beliefs about the efficacy of a medical treatment. We observed this result even when the anecdote was uninformative, was paired with an icon array, or when participants were provided with thorough medical decision aids about reproductive health procedures. Our findings highlight the pervasive effect of anecdotes on medical decision making.
{"title":"Anecdotes impact medical decisions even when presented with statistical information or decision aids.","authors":"Emily N Line, Sara Jaramillo, Micah Goldwater, Zachary Horne","doi":"10.1186/s41235-024-00577-3","DOIUrl":"10.1186/s41235-024-00577-3","url":null,"abstract":"<p><p>People are inundated with popular press reports about medical research concerning what is healthy, get advice from doctors, and hear personal anecdotes. How do people integrate conflicting anecdotal and statistical information when making medical decisions? In four experiments (N = 4126), we tested how people use conflicting information to judge the efficacy of artificial and real medical treatments. Participants read an anecdote from someone in a clinical trial, or who had undergone a medical treatment previously, for whom the medical treatment was ineffective. We found that reading anecdotes for either artificial or real medical treatments shifted participants' beliefs about the efficacy of a medical treatment. We observed this result even when the anecdote was uninformative, was paired with an icon array, or when participants were provided with thorough medical decision aids about reproductive health procedures. Our findings highlight the pervasive effect of anecdotes on medical decision making.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"51"},"PeriodicalIF":3.4,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11345347/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142056889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-26DOI: 10.1186/s41235-024-00585-3
Paige L Kemp, Vanessa M Loaiza, Colleen M Kelley, Christopher N Wahlheim
The efficacy of fake news corrections in improving memory and belief accuracy may depend on how often adults see false information before it is corrected. Two experiments tested the competing predictions that repeating fake news before corrections will either impair or improve memory and belief accuracy. These experiments also examined whether fake news exposure effects would differ for younger and older adults due to age-related differences in the recollection of contextual details. Younger and older adults read real and fake news headlines that appeared once or thrice. Next, they identified fake news corrections among real news headlines. Later, recognition and cued recall tests assessed memory for real news, fake news, if corrections occurred, and beliefs in retrieved details. Repeating fake news increased detection and remembering of corrections, correct real news retrieval, and erroneous fake news retrieval. No age differences emerged for detection of corrections, but younger adults remembered corrections better than older adults. At test, correct fake news retrieval for earlier-detected corrections was associated with better real news retrieval. This benefit did not differ between age groups in recognition but was greater for younger than older adults in cued recall. When detected corrections were not remembered at test, repeated fake news increased memory errors. Overall, both age groups believed correctly retrieved real news more than erroneously retrieved fake news to a similar degree. These findings suggest that fake news repetition effects on subsequent memory accuracy depended on age differences in recollection-based retrieval of fake news and that it was corrected.
{"title":"Correcting fake news headlines after repeated exposure: memory and belief accuracy in younger and older adults.","authors":"Paige L Kemp, Vanessa M Loaiza, Colleen M Kelley, Christopher N Wahlheim","doi":"10.1186/s41235-024-00585-3","DOIUrl":"10.1186/s41235-024-00585-3","url":null,"abstract":"<p><p>The efficacy of fake news corrections in improving memory and belief accuracy may depend on how often adults see false information before it is corrected. Two experiments tested the competing predictions that repeating fake news before corrections will either impair or improve memory and belief accuracy. These experiments also examined whether fake news exposure effects would differ for younger and older adults due to age-related differences in the recollection of contextual details. Younger and older adults read real and fake news headlines that appeared once or thrice. Next, they identified fake news corrections among real news headlines. Later, recognition and cued recall tests assessed memory for real news, fake news, if corrections occurred, and beliefs in retrieved details. Repeating fake news increased detection and remembering of corrections, correct real news retrieval, and erroneous fake news retrieval. No age differences emerged for detection of corrections, but younger adults remembered corrections better than older adults. At test, correct fake news retrieval for earlier-detected corrections was associated with better real news retrieval. This benefit did not differ between age groups in recognition but was greater for younger than older adults in cued recall. When detected corrections were not remembered at test, repeated fake news increased memory errors. Overall, both age groups believed correctly retrieved real news more than erroneously retrieved fake news to a similar degree. These findings suggest that fake news repetition effects on subsequent memory accuracy depended on age differences in recollection-based retrieval of fake news and that it was corrected.</p>","PeriodicalId":46827,"journal":{"name":"Cognitive Research-Principles and Implications","volume":"9 1","pages":"55"},"PeriodicalIF":3.4,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11345346/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142056890","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}