Pub Date : 2024-10-01Epub Date: 2024-05-06DOI: 10.3758/s13428-024-02430-3
Massimo Grassi, Andrea Felline, Niccolò Orlandi, Mattia Toffanin, Gnana Prakash Goli, Hurcan Andrei Senyuva, Mauro Migliardi, Giulio Contemori
PSYCHOACOUSTICS-WEB is an online tool written in JavaScript and PHP that enables the estimation of auditory sensory thresholds via adaptive threshold tracking. The toolbox implements the transformed up-down methods proposed by Levitt (Journal of the Acoustical Society of America, 49, 467-477, (1971) for a set of classic psychoacoustical tasks: frequency, intensity, and duration discrimination of pure tones; duration discrimination and gap detection of noise; and amplitude modulation detection with noise carriers. The toolbox can be used through a common web browser; it works with both fixed and mobile devices, and requires no programming skills. PSYCHOACOUSTICS-WEB is suitable for laboratory, classroom, and online testing and is designed for two main types of users: an occasional user and, above all, an experimenter using the toolbox for their own research. This latter user can create a personal account, customise existing experiments, and share them in the form of direct links to further users (e.g., the participants of a hypothetical experiment). Finally, because data storage is centralised, the toolbox offers the potential for creating a database of auditory skills.
PSYCHOACOUSTICS-WEB 是一个用 JavaScript 和 PHP 编写的在线工具,可通过自适应阈值跟踪估算听觉阈值。该工具箱实现了莱维特(Journal of the Acoustical Society of America, 49, 467-477, (1971))针对一系列经典心理声学任务提出的上下变换方法:纯音的频率、强度和持续时间分辨;噪声的持续时间分辨和间隙检测;以及带有噪声载体的振幅调制检测。该工具箱可通过普通网络浏览器使用,既可用于固定设备,也可用于移动设备,而且无需编程技能。PSYCHOACOUSTICS-WEB 适用于实验室、课堂和在线测试,主要为两类用户设计:临时用户,尤其是将工具箱用于自身研究的实验者。后一种用户可以创建个人账户,定制现有实验,并以直接链接的形式与其他用户(如假定实验的参与者)共享。最后,由于数据存储是集中式的,工具箱为创建听觉技能数据库提供了可能。
{"title":"PSYCHOACOUSTICS-WEB: A free online tool for the estimation of auditory thresholds.","authors":"Massimo Grassi, Andrea Felline, Niccolò Orlandi, Mattia Toffanin, Gnana Prakash Goli, Hurcan Andrei Senyuva, Mauro Migliardi, Giulio Contemori","doi":"10.3758/s13428-024-02430-3","DOIUrl":"10.3758/s13428-024-02430-3","url":null,"abstract":"<p><p>PSYCHOACOUSTICS-WEB is an online tool written in JavaScript and PHP that enables the estimation of auditory sensory thresholds via adaptive threshold tracking. The toolbox implements the transformed up-down methods proposed by Levitt (Journal of the Acoustical Society of America, 49, 467-477, (1971) for a set of classic psychoacoustical tasks: frequency, intensity, and duration discrimination of pure tones; duration discrimination and gap detection of noise; and amplitude modulation detection with noise carriers. The toolbox can be used through a common web browser; it works with both fixed and mobile devices, and requires no programming skills. PSYCHOACOUSTICS-WEB is suitable for laboratory, classroom, and online testing and is designed for two main types of users: an occasional user and, above all, an experimenter using the toolbox for their own research. This latter user can create a personal account, customise existing experiments, and share them in the form of direct links to further users (e.g., the participants of a hypothetical experiment). Finally, because data storage is centralised, the toolbox offers the potential for creating a database of auditory skills.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11362506/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140848885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-05-21DOI: 10.3758/s13428-024-02421-4
J M Rutkowska, T Ghilardi, S V Vacaru, J E van Schaik, M Meyer, S Hunnius, R Oostenveld
Surface facial electromyography (EMG) is commonly used to detect emotions from subtle facial expressions. Although there are established procedures for collecting EMG data and some aspects of their processing, there is little agreement among researchers about the optimal way to process the EMG signal, so that the study-unrelated variability (noise) is removed, and the emotion-related variability is best detected. The aim of the current paper was to establish an optimal processing pipeline for EMG data for identifying emotional expressions in facial muscles. We identified the most common processing steps from existing literature and created 72 processing pipelines that represented all the different processing choices. We applied these pipelines to a previously published dataset from a facial mimicry experiment, where 100 adult participants observed happy and sad facial expressions, whilst the activity of their facial muscles, zygomaticus major and corrugator supercilii, was recorded with EMG. We used a resampling approach and subsets of the original data to investigate the effect and robustness of different processing choices on the performance of a logistic regression model that predicted the mimicked emotion (happy/sad) from the EMG signal. In addition, we used a random forest model to identify the most important processing steps for the sensitivity of the logistic regression model. Three processing steps were found to be most impactful: baseline correction, standardisation within muscles, and standardisation within subjects. The chosen feature of interest and the signal averaging had little influence on the sensitivity to the effect. We recommend an optimal processing pipeline, share our code and data, and provide a step-by-step walkthrough for researchers.
{"title":"Optimal processing of surface facial EMG to identify emotional expressions: A data-driven approach.","authors":"J M Rutkowska, T Ghilardi, S V Vacaru, J E van Schaik, M Meyer, S Hunnius, R Oostenveld","doi":"10.3758/s13428-024-02421-4","DOIUrl":"10.3758/s13428-024-02421-4","url":null,"abstract":"<p><p>Surface facial electromyography (EMG) is commonly used to detect emotions from subtle facial expressions. Although there are established procedures for collecting EMG data and some aspects of their processing, there is little agreement among researchers about the optimal way to process the EMG signal, so that the study-unrelated variability (noise) is removed, and the emotion-related variability is best detected. The aim of the current paper was to establish an optimal processing pipeline for EMG data for identifying emotional expressions in facial muscles. We identified the most common processing steps from existing literature and created 72 processing pipelines that represented all the different processing choices. We applied these pipelines to a previously published dataset from a facial mimicry experiment, where 100 adult participants observed happy and sad facial expressions, whilst the activity of their facial muscles, zygomaticus major and corrugator supercilii, was recorded with EMG. We used a resampling approach and subsets of the original data to investigate the effect and robustness of different processing choices on the performance of a logistic regression model that predicted the mimicked emotion (happy/sad) from the EMG signal. In addition, we used a random forest model to identify the most important processing steps for the sensitivity of the logistic regression model. Three processing steps were found to be most impactful: baseline correction, standardisation within muscles, and standardisation within subjects. The chosen feature of interest and the signal averaging had little influence on the sensitivity to the effect. We recommend an optimal processing pipeline, share our code and data, and provide a step-by-step walkthrough for researchers.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11362446/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141074521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-06-04DOI: 10.3758/s13428-024-02443-y
Casey Becker, Russell Conduit, Philippe A Chouinard, Robin Laycock
Video recordings accurately capture facial expression movements; however, they are difficult for face perception researchers to standardise and manipulate. For this reason, dynamic morphs of photographs are often used, despite their lack of naturalistic facial motion. This study aimed to investigate how humans perceive emotions from faces using real videos and two different approaches to artificially generating dynamic expressions - dynamic morphs, and AI-synthesised deepfakes. Our participants perceived dynamic morphed expressions as less intense when compared with videos (all emotions) and deepfakes (fearful, happy, sad). Videos and deepfakes were perceived similarly. Additionally, they perceived morphed happiness and sadness, but not morphed anger or fear, as less genuine than other formats. Our findings support previous research indicating that social responses to morphed emotions are not representative of those to video recordings. The findings also suggest that deepfakes may offer a more suitable standardized stimulus type compared to morphs. Additionally, qualitative data were collected from participants and analysed using ChatGPT, a large language model. ChatGPT successfully identified themes in the data consistent with those identified by an independent human researcher. According to this analysis, our participants perceived dynamic morphs as less natural compared with videos and deepfakes. That participants perceived deepfakes and videos similarly suggests that deepfakes effectively replicate natural facial movements, making them a promising alternative for face perception research. The study contributes to the growing body of research exploring the usefulness of generative artificial intelligence for advancing the study of human perception.
{"title":"Can deepfakes be used to study emotion perception? A comparison of dynamic face stimuli.","authors":"Casey Becker, Russell Conduit, Philippe A Chouinard, Robin Laycock","doi":"10.3758/s13428-024-02443-y","DOIUrl":"10.3758/s13428-024-02443-y","url":null,"abstract":"<p><p>Video recordings accurately capture facial expression movements; however, they are difficult for face perception researchers to standardise and manipulate. For this reason, dynamic morphs of photographs are often used, despite their lack of naturalistic facial motion. This study aimed to investigate how humans perceive emotions from faces using real videos and two different approaches to artificially generating dynamic expressions - dynamic morphs, and AI-synthesised deepfakes. Our participants perceived dynamic morphed expressions as less intense when compared with videos (all emotions) and deepfakes (fearful, happy, sad). Videos and deepfakes were perceived similarly. Additionally, they perceived morphed happiness and sadness, but not morphed anger or fear, as less genuine than other formats. Our findings support previous research indicating that social responses to morphed emotions are not representative of those to video recordings. The findings also suggest that deepfakes may offer a more suitable standardized stimulus type compared to morphs. Additionally, qualitative data were collected from participants and analysed using ChatGPT, a large language model. ChatGPT successfully identified themes in the data consistent with those identified by an independent human researcher. According to this analysis, our participants perceived dynamic morphs as less natural compared with videos and deepfakes. That participants perceived deepfakes and videos similarly suggests that deepfakes effectively replicate natural facial movements, making them a promising alternative for face perception research. The study contributes to the growing body of research exploring the usefulness of generative artificial intelligence for advancing the study of human perception.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11362322/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141247131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-06-17DOI: 10.3758/s13428-024-02426-z
Michael P Woller, Craig K Enders
Recently, Asparouhov and Muthén Structural Equation Modeling: A Multidisciplinary Journal, 28, 1-14, (2021a, 2021b) proposed a variant of the Wald test that uses Markov chain Monte Carlo machinery to generate a chi-square test statistic for frequentist inference. Because the test's composition does not rely on analytic expressions for sampling variation and covariation, it potentially provides a way to get honest significance tests in cases where the likelihood-based test statistic's assumptions break down (e.g., in small samples). The goal of this study is to use simulation to compare the new MCM Wald test to its maximum likelihood counterparts, with respect to both their type I error rate and power. Our simulation examined the test statistics across different levels of sample size, effect size, and degrees of freedom (test complexity). An additional goal was to assess the robustness of the MCMC Wald test with nonnormal data. The simulation results uniformly demonstrated that the MCMC Wald test was superior to the maximum likelihood test statistic, especially with small samples (e.g., sample sizes less than 150) and complex models (e.g., models with five or more predictors). This conclusion held for nonnormal data as well. Lastly, we provide a brief application to a real data example.
{"title":"Exploration of the MCMC Wald test with linear regression.","authors":"Michael P Woller, Craig K Enders","doi":"10.3758/s13428-024-02426-z","DOIUrl":"10.3758/s13428-024-02426-z","url":null,"abstract":"<p><p>Recently, Asparouhov and Muthén Structural Equation Modeling: A Multidisciplinary Journal, 28, 1-14, (2021a, 2021b) proposed a variant of the Wald test that uses Markov chain Monte Carlo machinery to generate a chi-square test statistic for frequentist inference. Because the test's composition does not rely on analytic expressions for sampling variation and covariation, it potentially provides a way to get honest significance tests in cases where the likelihood-based test statistic's assumptions break down (e.g., in small samples). The goal of this study is to use simulation to compare the new MCM Wald test to its maximum likelihood counterparts, with respect to both their type I error rate and power. Our simulation examined the test statistics across different levels of sample size, effect size, and degrees of freedom (test complexity). An additional goal was to assess the robustness of the MCMC Wald test with nonnormal data. The simulation results uniformly demonstrated that the MCMC Wald test was superior to the maximum likelihood test statistic, especially with small samples (e.g., sample sizes less than 150) and complex models (e.g., models with five or more predictors). This conclusion held for nonnormal data as well. Lastly, we provide a brief application to a real data example.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141417596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-06-21DOI: 10.3758/s13428-024-02449-6
Camille Pluchot, Hans Adriaensen, Céline Parias, Didier Dubreuil, Cécile Arnould, Elodie Chaillou, Scott A Love
Magnetic resonance imaging (MRI) is a non-invasive technique that requires the participant to be completely motionless. To date, MRI in awake and unrestrained animals has only been achieved with humans and dogs. For other species, alternative techniques such as anesthesia, restraint and/or sedation have been necessary. Anatomical and functional MRI studies with sheep have only been conducted under general anesthesia. This ensures the absence of movement and allows relatively long MRI experiments but it removes the non-invasive nature of the MRI technique (i.e., IV injections, intubation). Anesthesia can also be detrimental to health, disrupt neurovascular coupling, and does not permit the study of higher-level cognition. Here, we present a proof-of-concept that sheep can be trained to perform a series of tasks, enabling them to voluntarily participate in MRI sessions without anesthesia or restraint. We describe a step-by-step training protocol based on positive reinforcement (food and praise) that could be used as a basis for future neuroimaging research in sheep. This protocol details the two successive phases required for sheep to successfully achieve MRI acquisitions of their brain. By providing structural brain MRI images from six out of ten sheep, we demonstrate the feasibility of our training protocol. This innovative training protocol paves the way for the possibility of conducting animal welfare-friendly functional MRI studies with sheep to investigate ovine cognition.
{"title":"Sheep (Ovis aries) training protocol for voluntary awake and unrestrained structural brain MRI acquisitions.","authors":"Camille Pluchot, Hans Adriaensen, Céline Parias, Didier Dubreuil, Cécile Arnould, Elodie Chaillou, Scott A Love","doi":"10.3758/s13428-024-02449-6","DOIUrl":"10.3758/s13428-024-02449-6","url":null,"abstract":"<p><p>Magnetic resonance imaging (MRI) is a non-invasive technique that requires the participant to be completely motionless. To date, MRI in awake and unrestrained animals has only been achieved with humans and dogs. For other species, alternative techniques such as anesthesia, restraint and/or sedation have been necessary. Anatomical and functional MRI studies with sheep have only been conducted under general anesthesia. This ensures the absence of movement and allows relatively long MRI experiments but it removes the non-invasive nature of the MRI technique (i.e., IV injections, intubation). Anesthesia can also be detrimental to health, disrupt neurovascular coupling, and does not permit the study of higher-level cognition. Here, we present a proof-of-concept that sheep can be trained to perform a series of tasks, enabling them to voluntarily participate in MRI sessions without anesthesia or restraint. We describe a step-by-step training protocol based on positive reinforcement (food and praise) that could be used as a basis for future neuroimaging research in sheep. This protocol details the two successive phases required for sheep to successfully achieve MRI acquisitions of their brain. By providing structural brain MRI images from six out of ten sheep, we demonstrate the feasibility of our training protocol. This innovative training protocol paves the way for the possibility of conducting animal welfare-friendly functional MRI studies with sheep to investigate ovine cognition.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11362526/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141436570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-07-08DOI: 10.3758/s13428-024-02459-4
Edwin J Burns
The Diagnostic Statistical Manual of Mental Disorders (DSM-5) recommends diagnosing neurocognitive disorders (i.e., cognitive impairment) when a patient scores beyond - 1 SD below neurotypical norms on two tests. I review how this approach will fail due to cognitive tests' power limitations, validity issues, imperfect reliabilities, and biases, before summarizing their resulting negative consequences. As a proof of concept, I use developmental prosopagnosia, a condition characterized by difficulties recognizing faces, to show the DSM-5 only diagnoses 62-70% (n1 = 61, n2 = 165) versus 100% (n1 = 61) through symptoms alone. Pooling the DSM-5 missed cases confirmed the presence of group-level impairments on objective tests, which were further evidenced through meta-analyses, thus validating their highly atypical symptoms. These findings support a paradigm shift towards bespoke diagnostic approaches for distinct cognitive impairments, including a symptom-based method when validated effective. I reject dogmatic adherence to the DSM-5 approach to neurocognitive disorders, and underscore the importance of a data driven, transdiagnostic approach to understanding patients' subjective cognitive impairments. This will ultimately benefit patients, their families, clinicians, and scientific progress.
{"title":"Improving the DSM-5 approach to cognitive impairment: Developmental prosopagnosia reveals the need for tailored diagnoses.","authors":"Edwin J Burns","doi":"10.3758/s13428-024-02459-4","DOIUrl":"10.3758/s13428-024-02459-4","url":null,"abstract":"<p><p>The Diagnostic Statistical Manual of Mental Disorders (DSM-5) recommends diagnosing neurocognitive disorders (i.e., cognitive impairment) when a patient scores beyond - 1 SD below neurotypical norms on two tests. I review how this approach will fail due to cognitive tests' power limitations, validity issues, imperfect reliabilities, and biases, before summarizing their resulting negative consequences. As a proof of concept, I use developmental prosopagnosia, a condition characterized by difficulties recognizing faces, to show the DSM-5 only diagnoses 62-70% (n1 = 61, n2 = 165) versus 100% (n1 = 61) through symptoms alone. Pooling the DSM-5 missed cases confirmed the presence of group-level impairments on objective tests, which were further evidenced through meta-analyses, thus validating their highly atypical symptoms. These findings support a paradigm shift towards bespoke diagnostic approaches for distinct cognitive impairments, including a symptom-based method when validated effective. I reject dogmatic adherence to the DSM-5 approach to neurocognitive disorders, and underscore the importance of a data driven, transdiagnostic approach to understanding patients' subjective cognitive impairments. This will ultimately benefit patients, their families, clinicians, and scientific progress.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11362378/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141557940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.3758/s13428-024-02509-x
Timo Seitz, Eunike Wetzel, Benjamin E Hilbig, Thorsten Meiser
Faking in self-report personality questionnaires describes a deliberate response distortion aimed at presenting oneself in an overly favorable manner. Unless the influence of faking on item responses is taken into account, faking can harm multiple psychometric properties of a test. In the present article, we account for faking using an extension of the multidimensional nominal response model (MNRM), which is an item response theory (IRT) model that offers a flexible framework for modeling different kinds of response biases. Particularly, we investigated under which circumstances the MNRM can adequately adjust substantive trait scores and latent correlations for the influence of faking and examined the role of variation in the way item content is related to social desirability (i.e., item desirability characteristics) in facilitating the modeling of faking and counteracting its detrimental effects. Using a simulation, we found that the inclusion of a faking dimension in the model can overall improve the recovery of substantive trait person parameters and latent correlations between substantive traits, especially when the impact of faking in the data is high. Item desirability characteristics moderated the effect of modeling faking and were themselves associated with different levels of parameter recovery. In an empirical demonstration with N = 1070 test-takers, we also showed that the faking modeling approach in combination with different item desirability characteristics can prove successful in empirical questionnaire data. We end the article with a discussion of implications for psychological assessment.
自我报告式人格问卷中的作假是指故意歪曲回答,目的是以过于有利的方式表现自己。如果不考虑伪造对项目反应的影响,伪造就会损害测验的多种心理测量属性。多维名义反应模型(MNRM)是一个项目反应理论(IRT)模型,它为不同类型的反应偏差建模提供了一个灵活的框架。特别是,我们研究了多维名义反应模型在什么情况下可以充分调整实质性特质得分和潜在相关性,使其免受作假的影响,并考察了项目内容与社会可取性相关方式的变化(即项目可取性特征)在促进作假建模和抵消其不利影响方面的作用。通过模拟,我们发现在模型中加入伪造维度可以从整体上改善实质性特质人称参数和实质性特质之间潜在相关性的恢复,尤其是当数据中伪造的影响较大时。项目可取性特征调节了建模伪造的效果,其本身也与不同程度的参数恢复有关。在对 N = 1070 名应试者进行的实证演示中,我们还证明了将伪造建模方法与不同的项目可取性特征相结合可以在实证问卷数据中取得成功。文章最后,我们讨论了对心理评估的影响。
{"title":"Using the multidimensional nominal response model to model faking in questionnaire data: The importance of item desirability characteristics.","authors":"Timo Seitz, Eunike Wetzel, Benjamin E Hilbig, Thorsten Meiser","doi":"10.3758/s13428-024-02509-x","DOIUrl":"https://doi.org/10.3758/s13428-024-02509-x","url":null,"abstract":"<p><p>Faking in self-report personality questionnaires describes a deliberate response distortion aimed at presenting oneself in an overly favorable manner. Unless the influence of faking on item responses is taken into account, faking can harm multiple psychometric properties of a test. In the present article, we account for faking using an extension of the multidimensional nominal response model (MNRM), which is an item response theory (IRT) model that offers a flexible framework for modeling different kinds of response biases. Particularly, we investigated under which circumstances the MNRM can adequately adjust substantive trait scores and latent correlations for the influence of faking and examined the role of variation in the way item content is related to social desirability (i.e., item desirability characteristics) in facilitating the modeling of faking and counteracting its detrimental effects. Using a simulation, we found that the inclusion of a faking dimension in the model can overall improve the recovery of substantive trait person parameters and latent correlations between substantive traits, especially when the impact of faking in the data is high. Item desirability characteristics moderated the effect of modeling faking and were themselves associated with different levels of parameter recovery. In an empirical demonstration with N = 1070 test-takers, we also showed that the faking modeling approach in combination with different item desirability characteristics can prove successful in empirical questionnaire data. We end the article with a discussion of implications for psychological assessment.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142279943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.3758/s13428-024-02476-3
Julien P Irmer, Andreas G Klein, Karin Schermelleh-Engel
The model-implied simulation-based power estimation (MSPE) approach is a new general method for power estimation (Irmer et al., 2024). MSPE was developed especially for power estimation of non-linear structural equation models (SEM), but it also can be applied to linear SEM and manifest models using the R package powerNLSEM. After first providing some information about MSPE and the new adaptive algorithm that automatically selects sample sizes for the best prediction of power using simulation, a tutorial on how to conduct the MSPE for quadratic and interaction SEM (QISEM) using the powerNLSEM package is provided. Power estimation is demonstrated for four methods, latent moderated structural equations (LMS), the unconstrained product indicator (UPI), a simple factor score regression (FSR), and a scale regression (SR) approach to QISEM. In two simulation studies, we highlight the performance of the MSPE for all four methods applied to two QISEM with varying complexity and reliability. Further, we justify the settings of the newly developed adaptive search algorithm via performance evaluations using simulation. Overall, the MSPE using the adaptive approach performs well in terms of bias and Type I error rates.
基于模型推导模拟的功率估计(MSPE)方法是一种新的功率估计通用方法(Irmer 等人,2024 年)。MSPE 是专为非线性结构方程模型(SEM)的幂估计而开发的,但也可使用 R 软件包 powerNLSEM 用于线性 SEM 和显式模型。首先介绍了有关 MSPE 和新的自适应算法的一些信息,该算法可通过模拟自动选择样本大小以获得最佳的预测功率,然后介绍了如何使用 powerNLSEM 软件包对二次和交互 SEM (QISEM) 进行 MSPE。我们演示了四种方法的功率估计,即潜在调节结构方程 (LMS)、无约束乘积指标 (UPI)、简单因子得分回归 (FSR) 和 QISEM 的尺度回归 (SR) 方法。在两项模拟研究中,我们强调了 MSPE 对所有四种方法的性能,并将其应用于两个具有不同复杂性和可靠性的 QISEM。此外,我们还通过模拟性能评估来证明新开发的自适应搜索算法的设置是合理的。总体而言,使用自适应方法的 MSPE 在偏差和 I 类错误率方面表现良好。
{"title":"Estimating power in complex nonlinear structural equation modeling including moderation effects: The powerNLSEM R-package.","authors":"Julien P Irmer, Andreas G Klein, Karin Schermelleh-Engel","doi":"10.3758/s13428-024-02476-3","DOIUrl":"https://doi.org/10.3758/s13428-024-02476-3","url":null,"abstract":"<p><p>The model-implied simulation-based power estimation (MSPE) approach is a new general method for power estimation (Irmer et al., 2024). MSPE was developed especially for power estimation of non-linear structural equation models (SEM), but it also can be applied to linear SEM and manifest models using the R package powerNLSEM. After first providing some information about MSPE and the new adaptive algorithm that automatically selects sample sizes for the best prediction of power using simulation, a tutorial on how to conduct the MSPE for quadratic and interaction SEM (QISEM) using the powerNLSEM package is provided. Power estimation is demonstrated for four methods, latent moderated structural equations (LMS), the unconstrained product indicator (UPI), a simple factor score regression (FSR), and a scale regression (SR) approach to QISEM. In two simulation studies, we highlight the performance of the MSPE for all four methods applied to two QISEM with varying complexity and reliability. Further, we justify the settings of the newly developed adaptive search algorithm via performance evaluations using simulation. Overall, the MSPE using the adaptive approach performs well in terms of bias and Type I error rates.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142279942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-20DOI: 10.3758/s13428-024-02493-2
Alejandrina Cristia, Lucas Gautheron, Zixing Zhang, Björn Schuller, Camila Scaff, Caroline Rowland, Okko Räsänen, Loann Peurey, Marvin Lavechin, William Havard, Caitlin M Fausey, Margaret Cychosz, Elika Bergelson, Heather Anderson, Najla Al Futaisi, Melanie Soderstrom
Long-form audio recordings are increasingly used to study individual variation, group differences, and many other topics in theoretical and applied fields of developmental science, particularly for the description of children's language input (typically speech from adults) and children's language output (ranging from babble to sentences). The proprietary LENA software has been available for over a decade, and with it, users have come to rely on derived metrics like adult word count (AWC) and child vocalization counts (CVC), which have also more recently been derived using an open-source alternative, the ACLEW pipeline. Yet, there is relatively little work assessing the reliability of long-form metrics in terms of the stability of individual differences across time. Filling this gap, we analyzed eight spoken-language datasets: four from North American English-learning infants, and one each from British English-, French-, American English-/Spanish-, and Quechua-/Spanish-learning infants. The audio data were analyzed using two types of processing software: LENA and the ACLEW open-source pipeline. When all corpora were included, we found relatively low to moderate reliability (across multiple recordings, intraclass correlation coefficient attributed to the child identity [Child ICC], was < 50% for most metrics). There were few differences between the two pipelines. Exploratory analyses suggested some differences as a function of child age and corpora. These findings suggest that, while reliability is likely sufficient for various group-level analyses, caution is needed when using either LENA or ACLEW tools to study individual variation. We also encourage improvement of extant tools, specifically targeting accurate measurement of individual variation.
长篇录音越来越多地被用于研究个体差异、群体差异以及发育科学理论和应用领域的许多其他课题,特别是用于描述儿童的语言输入(通常是成人的讲话)和儿童的语言输出(从咿呀学语到句子)。专有的 LENA 软件已问世十多年,用户已开始依赖成人词数(AWC)和儿童发声数(CVC)等衍生指标。然而,就个体差异在不同时期的稳定性而言,评估长式指标可靠性的工作相对较少。为了填补这一空白,我们分析了八个口语数据集:四个数据集来自学习北美英语的婴儿,另一个数据集来自学习英国英语、法语、美国英语/西班牙语和克丘亚语/西班牙语的婴儿。音频数据使用两种处理软件进行分析:LENA 和 ACLEW 开源管道。当包含所有语料库时,我们发现了相对较低到中等的可靠性(在多个录音中,归因于儿童身份的类内相关系数 [Child ICC] 为
{"title":"Establishing the reliability of metrics extracted from long-form recordings using LENA and the ACLEW pipeline.","authors":"Alejandrina Cristia, Lucas Gautheron, Zixing Zhang, Björn Schuller, Camila Scaff, Caroline Rowland, Okko Räsänen, Loann Peurey, Marvin Lavechin, William Havard, Caitlin M Fausey, Margaret Cychosz, Elika Bergelson, Heather Anderson, Najla Al Futaisi, Melanie Soderstrom","doi":"10.3758/s13428-024-02493-2","DOIUrl":"https://doi.org/10.3758/s13428-024-02493-2","url":null,"abstract":"<p><p>Long-form audio recordings are increasingly used to study individual variation, group differences, and many other topics in theoretical and applied fields of developmental science, particularly for the description of children's language input (typically speech from adults) and children's language output (ranging from babble to sentences). The proprietary LENA software has been available for over a decade, and with it, users have come to rely on derived metrics like adult word count (AWC) and child vocalization counts (CVC), which have also more recently been derived using an open-source alternative, the ACLEW pipeline. Yet, there is relatively little work assessing the reliability of long-form metrics in terms of the stability of individual differences across time. Filling this gap, we analyzed eight spoken-language datasets: four from North American English-learning infants, and one each from British English-, French-, American English-/Spanish-, and Quechua-/Spanish-learning infants. The audio data were analyzed using two types of processing software: LENA and the ACLEW open-source pipeline. When all corpora were included, we found relatively low to moderate reliability (across multiple recordings, intraclass correlation coefficient attributed to the child identity [Child ICC], was < 50% for most metrics). There were few differences between the two pipelines. Exploratory analyses suggested some differences as a function of child age and corpora. These findings suggest that, while reliability is likely sufficient for various group-level analyses, caution is needed when using either LENA or ACLEW tools to study individual variation. We also encourage improvement of extant tools, specifically targeting accurate measurement of individual variation.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142279941","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-16DOI: 10.3758/s13428-024-02504-2
Charisse B. Pickron, Alexia J. Brown, Caitlin M. Hudac, Lisa S. Scott
Face processing is a central component of human communication and social engagement. The present investigation introduces a set of racially and ethnically inclusive faces created for researchers interested in perceptual and socio-cognitive processes linked to human faces. The Diverse Face Images (DFI) stimulus set includes high-quality still images of female faces that are racially and ethnically representative, include multiple images of direct and indirect gaze for each model and control for low-level perceptual variance between images. The DFI stimuli will support researchers interested in studying face processing throughout the lifespan as well as other questions that require a diversity of faces or gazes. This report includes a detailed description of stimuli development and norming data for each model. Adults completed a questionnaire rating each image in the DFI stimuli set on three major qualities relevant to face processing: (1) strength of race/ethnicity group associations, (2) strength of eye gaze orientation, and (3) strength of emotion expression. These validation data highlight the presence of rater variability within and between individual model images as well as within and between race and ethnicity groups.
{"title":"Diverse Face Images (DFI): Validated for racial representation and eye gaze","authors":"Charisse B. Pickron, Alexia J. Brown, Caitlin M. Hudac, Lisa S. Scott","doi":"10.3758/s13428-024-02504-2","DOIUrl":"https://doi.org/10.3758/s13428-024-02504-2","url":null,"abstract":"<p>Face processing is a central component of human communication and social engagement. The present investigation introduces a set of racially and ethnically inclusive faces created for researchers interested in perceptual and socio-cognitive processes linked to human faces. The Diverse Face Images (DFI) stimulus set includes high-quality still images of female faces that are racially and ethnically representative, include multiple images of direct and indirect gaze for each model and control for low-level perceptual variance between images. The DFI stimuli will support researchers interested in studying face processing throughout the lifespan as well as other questions that require a diversity of faces or gazes. This report includes a detailed description of stimuli development and norming data for each model. Adults completed a questionnaire rating each image in the DFI stimuli set on three major qualities relevant to face processing: (1) strength of race/ethnicity group associations, (2) strength of eye gaze orientation, and (3) strength of emotion expression. These validation data highlight the presence of rater variability within and between individual model images as well as within and between race and ethnicity groups.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":null,"pages":null},"PeriodicalIF":5.4,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142252798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}