Empirically supported measures of suicidal thoughts and behaviors (STBs) are needed to serve as reference outcomes for suicide risk screening tools and to monitor severity and treatment progress in children and adolescents with STBs. The present paper systematically reviewed existing measures of STBs in youth and studies evaluating their psychometric properties and clinical utility. Measures were then evaluated on reliability, validity, and clinical utility. Sixteen articles (20 independent samples) were found with psychometric data with youth samples for eight measures. Interview-based measures were found to have the strongest psychometric support and clinical utility. Significant limitations exist for all self-report measures due to inherent characteristics of these measures that cannot be remedied through additional psychometric study. There is an urgent need for the development and validation of new self-report measures of STBs, particularly for preadolescent children, sexual and gender minority youth, and racial/ethnic minority youth.
{"title":"Measures of Suicidal Thoughts and Behaviors in Children and Adolescents: A Systematic Review and Recommendations for Use in Clinical and Research Settings.","authors":"Richard T Liu, Alexandra H Bettis, Hannah R Lawrence, Rachel F L Walsh, Ana E Sheehan, Olivia H Pollak, Auburn R Stephenson, Marin M Kautz, Rachel M Marlowe","doi":"10.1177/10731911241249438","DOIUrl":"10.1177/10731911241249438","url":null,"abstract":"<p><p>Empirically supported measures of suicidal thoughts and behaviors (STBs) are needed to serve as reference outcomes for suicide risk screening tools and to monitor severity and treatment progress in children and adolescents with STBs. The present paper systematically reviewed existing measures of STBs in youth and studies evaluating their psychometric properties and clinical utility. Measures were then evaluated on reliability, validity, and clinical utility. Sixteen articles (20 independent samples) were found with psychometric data with youth samples for eight measures. Interview-based measures were found to have the strongest psychometric support and clinical utility. Significant limitations exist for all self-report measures due to inherent characteristics of these measures that cannot be remedied through additional psychometric study. There is an urgent need for the development and validation of new self-report measures of STBs, particularly for preadolescent children, sexual and gender minority youth, and racial/ethnic minority youth.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"303-320"},"PeriodicalIF":3.5,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140920968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The Diagnostic and Statistical Manual of Mental Disorders, 5th Edition (DSM-5; American Psychiatric Association, 2013), includes 10 categorical personality disorders (PD) in Section II (Section II PD) and a dimensional alternative model of PD (AMPD) in Section III. We compared the two models in explaining concurrent psychosocial functioning levels in psychiatric outpatients and community residents screened as at risk for PD pathology (N = 600). The AMPD's fully dimensional form showed stronger associations with psychosocial difficulties and explained more of their variance compared with the categorical Section II PD. AMPD Criterion A (personality functioning impairment) and Criterion B (pathological traits) incrementally predicted psychosocial functioning about equally with some unique predictions. Finally, AMPD's six categorical PD diagnoses did not show stronger associations with psychosocial functioning than the corresponding Section II PD diagnoses. Findings directly comparing the two models remain important and timely for informing future conceptualizations of PD in the diagnostic system.
{"title":"Competing Models of Personality Disorder: Relations With Psychosocial Functioning.","authors":"Eunyoe Ro, Hallie Nuzum, Lee Anna Clark","doi":"10.1177/10731911241253409","DOIUrl":"10.1177/10731911241253409","url":null,"abstract":"<p><p>The <i>Diagnostic and Statistical Manual of Mental Disorders, 5th Edition</i> (<i>DSM-5</i>; American Psychiatric Association, 2013), includes 10 categorical personality disorders (PD) in Section II (Section II PD) and a dimensional alternative model of PD (AMPD) in Section III. We compared the two models in explaining concurrent psychosocial functioning levels in psychiatric outpatients and community residents screened as at risk for PD pathology (<i>N</i> = 600). The AMPD's fully dimensional form showed stronger associations with psychosocial difficulties and explained more of their variance compared with the categorical Section II PD. AMPD Criterion A (personality functioning impairment) and Criterion B (pathological traits) incrementally predicted psychosocial functioning about equally with some unique predictions. Finally, AMPD's six categorical PD diagnoses did not show stronger associations with psychosocial functioning than the corresponding Section II PD diagnoses. Findings directly comparing the two models remain important and timely for informing future conceptualizations of PD in the diagnostic system.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"321-334"},"PeriodicalIF":3.5,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141154869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This study assessed psychometric qualities of indirect measures assessing Implicit Theories (ITs) of sexual offending: Implicit Association Task (IAT), Implicit Relational Assessment Procedure (IRAP), and Relational Responding Task (RRT). For comparison reasons, aggressive behavior was also assessed. In a male sample from the general population (N = 109), we assessed each measure's (a) feasibility (mean latency, error rate, passing criteria), (b) internal consistency, (c) convergent and discriminant validity, and (d) incremental and predictive validity. Results indicated that no indirect measure met all criteria. Although the IAT was reasonably feasible and reliable in measuring aggression, ITs could not be reliably assessed. The RRT was feasible and somewhat reliable in assessing ITs, whereas the IRAP showed limited feasibility, high task complexity, low reliability, and the presence of a method factor. No measure had incremental predictive validity over the use of self-report measures, although we note that the power to detect such associations was limited. As none of the indirect measures performed satisfactorily on the measured criteria, the use of these measures in clinical practice seems currently unwarranted to assess ITs.
本研究评估了评估性犯罪内隐理论(ITs)的间接测量方法的心理测量质量:内隐联想任务(IAT)、内隐关系评估程序(IRAP)和关系反应任务(RRT)。为了便于比较,还对攻击行为进行了评估。在一个来自普通人群的男性样本(N = 109)中,我们评估了每项测量的(a)可行性(平均潜伏期、错误率、通过标准)、(b)内部一致性、(c)收敛性和鉴别性有效性,以及(d)增量性和预测性有效性。结果表明,没有一种间接测量方法符合所有标准。虽然 IAT 在测量攻击性方面相当可行和可靠,但 ITs 无法得到可靠的评估。RRT 在评估 ITs 方面具有可行性和一定的可靠性,而 IRAP 则显示出有限的可行性、较高的任务复杂性、较低的可靠性以及方法因素的存在。与使用自我报告的方法相比,没有一种方法具有更高的预测有效性,尽管我们注意到检测这种关联的能力有限。由于没有一种间接测量方法在测量标准上的表现令人满意,因此目前似乎没有必要在临床实践中使用这些方法来评估信息技术。
{"title":"Assessing Implicit Theories in Sexual Offending Using Indirect Measures: Feasibility, Reliability, and Incremental Validity.","authors":"Mirthe G C Noteborn, Jelle J Sijtsema, Jaap J A Denissen, Stefan Bogaerts","doi":"10.1177/10731911241245009","DOIUrl":"10.1177/10731911241245009","url":null,"abstract":"<p><p>This study assessed psychometric qualities of indirect measures assessing Implicit Theories (ITs) of sexual offending: Implicit Association Task (IAT), Implicit Relational Assessment Procedure (IRAP), and Relational Responding Task (RRT). For comparison reasons, aggressive behavior was also assessed. In a male sample from the general population (<i>N</i> = 109), we assessed each measure's (a) feasibility (mean latency, error rate, passing criteria), (b) internal consistency, (c) convergent and discriminant validity, and (d) incremental and predictive validity. Results indicated that no indirect measure met all criteria. Although the IAT was reasonably feasible and reliable in measuring aggression, ITs could not be reliably assessed. The RRT was feasible and somewhat reliable in assessing ITs, whereas the IRAP showed limited feasibility, high task complexity, low reliability, and the presence of a method factor. No measure had incremental predictive validity over the use of self-report measures, although we note that the power to detect such associations was limited. As none of the indirect measures performed satisfactorily on the measured criteria, the use of these measures in clinical practice seems currently unwarranted to assess ITs.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"447-469"},"PeriodicalIF":3.5,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11915770/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140875717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-31DOI: 10.1177/10731911251330092
Stephen Raynes, Karen Dobkins
The current research aimed to provide initial psychometric validation of a new multifaceted mindfulness questionnaire (referred to as the State Four Facet Mindfulness Questionnaire, or the "state-4FMQ" for short) adapted from the commonly used Five Facet Mindfulness Questionnaire (referred to as the "trait-FFMQ"). The research was divided into two pre-registered studies. In both, undergraduates partook in a 20-minute mindfulness meditation (via audio recording), and then answered questions, including the state-4FMQ, pertaining to their experience during the meditation. In Study 2, participants additionally partook in a 20-minute control condition. The state-4FMQ was developed using exploratory factor analysis (EFA; Study 1) and confirmatory factor analysis (CFA; Study 2). In Study 2, a short-form of the state-4FMQ was established, and several additional forms of measurement validity were tested. EFA and CFA results supported a four-factor structure, which was identical to the trait-FFMQ with the exclusion of Nonreactivity. This newly created state-4FMQ, and its short-form, showed good internal consistency as well as convergent, predictive, and construct validity. In addition, it was found that some facets, more than others, predicted momentary well-being. The validity of the state-4FMQ shows that it can be used to measure multiple facets of state mindfulness across a variety of situations.
{"title":"Development and Initial Validation of the State Four Facet Mindfulness Questionnaire.","authors":"Stephen Raynes, Karen Dobkins","doi":"10.1177/10731911251330092","DOIUrl":"10.1177/10731911251330092","url":null,"abstract":"<p><p>The current research aimed to provide initial psychometric validation of a new multifaceted mindfulness questionnaire (referred to as the State Four Facet Mindfulness Questionnaire, or the \"state-4FMQ\" for short) adapted from the commonly used Five Facet Mindfulness Questionnaire (referred to as the \"trait-FFMQ\"). The research was divided into two pre-registered studies. In both, undergraduates partook in a 20-minute mindfulness meditation (via audio recording), and then answered questions, including the state-4FMQ, pertaining to their experience during the meditation. In Study 2, participants additionally partook in a 20-minute control condition. The state-4FMQ was developed using exploratory factor analysis (EFA; Study 1) and confirmatory factor analysis (CFA; Study 2). In Study 2, a short-form of the state-4FMQ was established, and several additional forms of measurement validity were tested. EFA and CFA results supported a four-factor structure, which was identical to the trait-FFMQ with the exclusion of Nonreactivity. This newly created state-4FMQ, and its short-form, showed good internal consistency as well as convergent, predictive, and construct validity. In addition, it was found that some facets, more than others, predicted momentary well-being. The validity of the state-4FMQ shows that it can be used to measure multiple facets of state mindfulness across a variety of situations.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251330092"},"PeriodicalIF":3.5,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143750900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-31DOI: 10.1177/10731911251320597
Ohad Gilbar, Laura Watkins, Ruby Charak
The Cyber Aggression in Relationships Scale (CARS) is a widely used and validated measure of adult cyber intimate partner violence (C-IPV). The current study aimed to create a short screener of CARS (CARS-SS) through testing cross-cultural samples. To do so, we examined the rate of chosen items of the original scale on five pre-existing study samples: we then procured and tested the construct validity and reliability of the CARS-SS among two study samples (United States and Israel). Results of an item factor analysis confirmed one factor for both perpetration and victimization and analyses suggested good internal reliability. Overall, the CARS-SS was associated with a measure of in-person intimate partner violence. Findings suggest the CARS-SS may be a helpful measure of C-IPV among clinical and community samples of young adults.
{"title":"Short Screener of the Cyber Aggression in Relationships Scale: Construct Validity and Reliability Cross-Cultural Samples.","authors":"Ohad Gilbar, Laura Watkins, Ruby Charak","doi":"10.1177/10731911251320597","DOIUrl":"https://doi.org/10.1177/10731911251320597","url":null,"abstract":"<p><p>The Cyber Aggression in Relationships Scale (CARS) is a widely used and validated measure of adult cyber intimate partner violence (C-IPV). The current study aimed to create a short screener of CARS (CARS-SS) through testing cross-cultural samples. To do so, we examined the rate of chosen items of the original scale on five pre-existing study samples: we then procured and tested the construct validity and reliability of the CARS-SS among two study samples (United States and Israel). Results of an item factor analysis confirmed one factor for both perpetration and victimization and analyses suggested good internal reliability. Overall, the CARS-SS was associated with a measure of in-person intimate partner violence. Findings suggest the CARS-SS may be a helpful measure of C-IPV among clinical and community samples of young adults.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251320597"},"PeriodicalIF":3.5,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143750901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-31DOI: 10.1177/10731911251326395
Edric S Widjaja, Emily K Spackman, Timothy F Bainbridge, Steven G Ludeke, Mirko Uljarevic, Kristelle Hudry, Luke D Smillie
Growing research suggests that clinical psychological traits are contiguous with normal personality and can be located within the same psychometric frameworks. In this article, we examined whether autism-related traits (ARTs) can plausibly be located within the Five Factor Model (FFM) of personality. Across two studies (Ns = 408 and 423), participants completed measures of ARTs, broad FFM domains, and narrower FFM facets. We used empirically derived criteria to evaluate whether ARTs overlapped (i.e., shared variance) with the FFM domains to a degree that was comparable to FFM facets. Results suggested that most socially oriented ARTs could be represented as facets of the extraversion domain, whereas behaviorally oriented ARTs were more peripheral to the FFM. Cognitively oriented ARTs were less consistently linked with the FFM. These findings highlight the value of the FFM as an organizing framework for ARTs, marking an important step toward synthesis across the personality and autism literatures.
越来越多的研究表明,临床心理特征与正常人格是相辅相成的,并且可以被定位在相同的心理测量框架内。在本文中,我们研究了自闭症相关特质(ARTs)是否可以合理地归入人格五因素模型(FFM)。在两项研究中(Ns = 408 和 423),参与者完成了 ARTs、广义 FFM 领域和狭义 FFM 面的测量。我们使用经验得出的标准来评估 ART 是否与 FFM 领域重叠(即共享方差),其程度是否与 FFM 方面相当。结果表明,大多数社交导向的 ART 都可以表示为外向性领域的切面,而行为导向的 ART 与 FFM 的切面更为边缘。认知导向的 ART 与 FFM 的联系则不那么一致。这些发现凸显了FFM作为ART组织框架的价值,标志着我们在人格和自闭症研究领域迈出了重要的一步。
{"title":"The Utility of the Five Factor Model of Personality as an Organizing Framework for Autism-Related Traits.","authors":"Edric S Widjaja, Emily K Spackman, Timothy F Bainbridge, Steven G Ludeke, Mirko Uljarevic, Kristelle Hudry, Luke D Smillie","doi":"10.1177/10731911251326395","DOIUrl":"https://doi.org/10.1177/10731911251326395","url":null,"abstract":"<p><p>Growing research suggests that clinical psychological traits are contiguous with normal personality and can be located within the same psychometric frameworks. In this article, we examined whether autism-related traits (ARTs) can plausibly be located within the Five Factor Model (FFM) of personality. Across two studies (<i>N</i>s = 408 and 423), participants completed measures of ARTs, broad FFM <i>domains</i>, and narrower FFM <i>facets</i>. We used empirically derived criteria to evaluate whether ARTs overlapped (i.e., shared variance) with the FFM domains to a degree that was comparable to FFM facets. Results suggested that most socially oriented ARTs could be represented as facets of the extraversion domain, whereas behaviorally oriented ARTs were more peripheral to the FFM. Cognitively oriented ARTs were less consistently linked with the FFM. These findings highlight the value of the FFM as an organizing framework for ARTs, marking an important step toward synthesis across the personality and autism literatures.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251326395"},"PeriodicalIF":3.5,"publicationDate":"2025-03-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143750060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-29DOI: 10.1177/10731911251327255
Martin Hochheimer, Justin C Strickland, Jennifer D Ellis, Jill A Rabinowitz, J Gregory Hobelmann, Andrew S Huhn
This study evaluated the Penn State Worry Questionnaire (PSWQ) as a tool for measuring worry and anxiety levels among individuals entering treatment for substance use disorders (SUDs). The sample included 75,047 individuals admitted to SUD treatment centers, with assessments conducted weekly. Individuals entering SUD treatment exhibited higher baseline levels of worry; however, worry levels declined over the course of treatment. The PSWQ demonstrated good internal consistency, high test-retest reliability, and good discriminant validity when correlated with measures of depression and stress. The factor structure analysis confirmed that the PSWQ measures the same underlying construct of worry in the SUD treatment population, with a single-factor model showing satisfactory fit. This extends the reach of the PSWQ to the SUD treatment population by reaffirming its reliability, validity, and factor structure, with the expectation of higher levels of worry compared to a non-SUD population at the beginning of treatment, which decline over time.
{"title":"Normative Values and Psychometric Properties of the Penn State Worry Questionnaire in Substance Use Disorder Treatment Population.","authors":"Martin Hochheimer, Justin C Strickland, Jennifer D Ellis, Jill A Rabinowitz, J Gregory Hobelmann, Andrew S Huhn","doi":"10.1177/10731911251327255","DOIUrl":"https://doi.org/10.1177/10731911251327255","url":null,"abstract":"<p><p>This study evaluated the Penn State Worry Questionnaire (PSWQ) as a tool for measuring worry and anxiety levels among individuals entering treatment for substance use disorders (SUDs). The sample included 75,047 individuals admitted to SUD treatment centers, with assessments conducted weekly. Individuals entering SUD treatment exhibited higher baseline levels of worry; however, worry levels declined over the course of treatment. The PSWQ demonstrated good internal consistency, high test-retest reliability, and good discriminant validity when correlated with measures of depression and stress. The factor structure analysis confirmed that the PSWQ measures the same underlying construct of worry in the SUD treatment population, with a single-factor model showing satisfactory fit. This extends the reach of the PSWQ to the SUD treatment population by reaffirming its reliability, validity, and factor structure, with the expectation of higher levels of worry compared to a non-SUD population at the beginning of treatment, which decline over time.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251327255"},"PeriodicalIF":3.5,"publicationDate":"2025-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143742080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-29DOI: 10.1177/10731911251321930
Jack T Waddell, Scott E King, William R Corbin
Literature on the location and contextual features of drinking events (i.e., physical context) remains scant and underdeveloped. This study developed and provided preliminary validation of a measure typical physical drinking contexts. Participants (N = 1,642) self-reported their typical physical drinking context (via items generated), their drinking behavior, demographics, and typical social drinking context. Three samples (total N = 1,642) assessed factor structure, measurement invariance, and validity testing. Factor analyses suggested a four-factor structure, indicative of high arousal private (e.g., at a large house party), high arousal public (e.g., at a concert), low arousal private (e.g., at home), and low arousal public (e.g., on a date) contexts. Measurement invariance was established across sex, race/ethnicity, and drinking frequency, and convergent and discriminant validity was evaluated via bivariate correlations with social/solitary drinking frequency. High arousal contexts were associated with heavier/binge drinking, whereas high arousal private contexts and low arousal contexts were associated with greater negative consequences. Relations between high arousal contexts and heavier drinking remained above and beyond the overall drinking frequency and social context items. Findings lay the framework for future validation and longitudinal/diary studies to test how (and for whom) relations between physical drinking contexts and drinking behavior operate.
{"title":"Initial Development and Preliminary Validation of the Physical Drinking Contexts Scale.","authors":"Jack T Waddell, Scott E King, William R Corbin","doi":"10.1177/10731911251321930","DOIUrl":"https://doi.org/10.1177/10731911251321930","url":null,"abstract":"<p><p>Literature on the location and contextual features of drinking events (i.e., physical context) remains scant and underdeveloped. This study developed and provided preliminary validation of a measure typical physical drinking contexts. Participants (<i>N</i> = 1,642) self-reported their typical physical drinking context (via items generated), their drinking behavior, demographics, and typical social drinking context. Three samples (total <i>N</i> = 1,642) assessed factor structure, measurement invariance, and validity testing. Factor analyses suggested a four-factor structure, indicative of high arousal private (e.g., at a large house party), high arousal public (e.g., at a concert), low arousal private (e.g., at home), and low arousal public (e.g., on a date) contexts. Measurement invariance was established across sex, race/ethnicity, and drinking frequency, and convergent and discriminant validity was evaluated via bivariate correlations with social/solitary drinking frequency. High arousal contexts were associated with heavier/binge drinking, whereas high arousal private contexts and low arousal contexts were associated with greater negative consequences. Relations between high arousal contexts and heavier drinking remained above and beyond the overall drinking frequency and social context items. Findings lay the framework for future validation and longitudinal/diary studies to test how (and for whom) relations between physical drinking contexts and drinking behavior operate.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251321930"},"PeriodicalIF":3.5,"publicationDate":"2025-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143742078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-29DOI: 10.1177/10731911251326371
Adam P Natoli
The validity and utility of translated instruments (psychological measures) depend on the quality of their translation, and differences in key linguistic characteristics could introduce bias. Likewise, linguistic differences between instruments designed to measure analogous constructs might contribute to similar instruments possessing dissimilar psychometrics. This article introduces and demonstrates the use of natural language processing (NLP), a subfield of artificial intelligence, to linguistically analyze 13 translations of two psychological measures previously translated into numerous languages. NLP was used to generate estimates reflecting specific linguistic characteristics of test items (emotional tone/intensity, sentiment, valence, arousal, and dominance), which were then compared across translations at both the test- and item-level, as well as between the two instruments. Results revealed that key linguistic characteristics can profoundly vary both within and between tests. Following a discussion of results, the current limitations of this approach are summarized and strategies for advancing this methodology are proposed.
{"title":"Leveraging Artificial Intelligence to Linguistically Compare Test Translations: A Methodological Introduction and Demonstration.","authors":"Adam P Natoli","doi":"10.1177/10731911251326371","DOIUrl":"https://doi.org/10.1177/10731911251326371","url":null,"abstract":"<p><p>The validity and utility of translated instruments (psychological measures) depend on the quality of their translation, and differences in key linguistic characteristics could introduce bias. Likewise, linguistic differences between instruments designed to measure analogous constructs might contribute to similar instruments possessing dissimilar psychometrics. This article introduces and demonstrates the use of natural language processing (NLP), a subfield of artificial intelligence, to linguistically analyze 13 translations of two psychological measures previously translated into numerous languages. NLP was used to generate estimates reflecting specific linguistic characteristics of test items (emotional tone/intensity, sentiment, valence, arousal, and dominance), which were then compared across translations at both the test- and item-level, as well as between the two instruments. Results revealed that key linguistic characteristics can profoundly vary both within and between tests. Following a discussion of results, the current limitations of this approach are summarized and strategies for advancing this methodology are proposed.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251326371"},"PeriodicalIF":3.5,"publicationDate":"2025-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143742079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-29DOI: 10.1177/10731911251328604
Wendy C Higgins, Victoria Savalei, Vince Polito, Robert M Ross
The Reading the Mind in the Eyes Test (RMET) is widely used in clinical and non-clinical research. However, the structural properties of RMET scores have yet to be rigorously examined. We analyzed the structural properties of RMET scores in nine existing datasets comprising non-clinical samples ranging from 558 to 9,267 (median = 1,112) participants each. We used confirmatory factor analysis to assess two theoretically derived factor models, exploratory factor analysis to identify possible alternative factor models, and reliability estimates to assess internal consistency. Neither of the theoretically derived models was a good fit for any of the nine datasets, and we were unable to identify any better fitting multidimensional models. Internal consistency metrics were acceptable in six of the nine datasets, but these metrics are difficult to interpret given the uncertain factor structures. Our findings contribute to a growing body of evidence questioning the reliability and validity of RMET scores.
{"title":"Reading the Mind in the Eyes Test Scores Demonstrate Poor Structural Properties in Nine Large Non-Clinical Samples.","authors":"Wendy C Higgins, Victoria Savalei, Vince Polito, Robert M Ross","doi":"10.1177/10731911251328604","DOIUrl":"https://doi.org/10.1177/10731911251328604","url":null,"abstract":"<p><p>The Reading the Mind in the Eyes Test (RMET) is widely used in clinical and non-clinical research. However, the structural properties of RMET scores have yet to be rigorously examined. We analyzed the structural properties of RMET scores in nine existing datasets comprising non-clinical samples ranging from 558 to 9,267 (median = 1,112) participants each. We used confirmatory factor analysis to assess two theoretically derived factor models, exploratory factor analysis to identify possible alternative factor models, and reliability estimates to assess internal consistency. Neither of the theoretically derived models was a good fit for any of the nine datasets, and we were unable to identify any better fitting multidimensional models. Internal consistency metrics were acceptable in six of the nine datasets, but these metrics are difficult to interpret given the uncertain factor structures. Our findings contribute to a growing body of evidence questioning the reliability and validity of RMET scores.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251328604"},"PeriodicalIF":3.5,"publicationDate":"2025-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143742081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}