Pub Date : 2025-09-25DOI: 10.1177/10731911251376235
Anna L Gilmour, Brian A Feinstein, Mark A Whisman
The Kessler Psychological Distress Scale (K6) is used as a self-report measure of nonspecific psychological distress. Although research documents higher K6 scores among lesbian, gay, bisexual, and transgender (LGBT) individuals relative to cisgender, heterosexual individuals, measurement invariance of the K6 has not been established between these groups. We used multigroup confirmatory factor analysis to examine factorial invariance of the K6 between 1,765 LGBT and 20,632 cisgender, heterosexual individuals who completed the Well-Being and Basic Needs Survey. The K6 exhibited configural, weak/metric, and strong/scalar measurement invariance between groups, suggesting that it operates equivalently for both groups. We then examined differences in latent mean K6 scores between groups and differences in the percentage of individuals in each group who met a threshold for serious psychological distress (scores ≥ 13). The latent K6 mean and the percentage of individuals who met the threshold for serious psychological distress were both significantly higher for LGBT than for cisgender, heterosexual individuals.
{"title":"Measurement Invariance of and Mean Differences on the Kessler Psychological Distress Scale (K6) for LGBT and Cisgender, Heterosexual Individuals.","authors":"Anna L Gilmour, Brian A Feinstein, Mark A Whisman","doi":"10.1177/10731911251376235","DOIUrl":"https://doi.org/10.1177/10731911251376235","url":null,"abstract":"<p><p>The Kessler Psychological Distress Scale (K6) is used as a self-report measure of nonspecific psychological distress. Although research documents higher K6 scores among lesbian, gay, bisexual, and transgender (LGBT) individuals relative to cisgender, heterosexual individuals, measurement invariance of the K6 has not been established between these groups. We used multigroup confirmatory factor analysis to examine factorial invariance of the K6 between 1,765 LGBT and 20,632 cisgender, heterosexual individuals who completed the Well-Being and Basic Needs Survey. The K6 exhibited configural, weak/metric, and strong/scalar measurement invariance between groups, suggesting that it operates equivalently for both groups. We then examined differences in latent mean K6 scores between groups and differences in the percentage of individuals in each group who met a threshold for serious psychological distress (scores ≥ 13). The latent K6 mean and the percentage of individuals who met the threshold for serious psychological distress were both significantly higher for LGBT than for cisgender, heterosexual individuals.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251376235"},"PeriodicalIF":3.4,"publicationDate":"2025-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145136026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-25DOI: 10.1177/10731911251376214
Aaron N McInnes, Christi R P Sullivan, Angus W MacDonald, Alik S Widge
Measuring the function of decision-making systems reliably is a key goal to assess cognitive functions that underlie psychopathology. However, few metrics are demonstrably reliable, clinically relevant, and able to capture complex overlapping cognitive domains while quantifying heterogeneity across individuals. The WebSurf task is a reverse-translational human experiential foraging paradigm that indexes naturalistic and clinically relevant decision-making. To determine its potential clinical utility, we examined the psychometric properties and clinical correlates of behavioral parameters extracted from WebSurf in an initial exploratory experiment (N = 132) and a preregistered validation experiment (N = 109). Behavior was stable over repeated administrations of the task, as were individual differences. The ability to measure decision-making consistently supports WebSurf's potential utility to predict treatment response, monitor clinical change, and define neurocognitive profiles associated with psychopathology. Moreover, specific WebSurf metrics were predicted by psychiatric symptoms in a replicable manner. Mania and externalizing symptom profiles predicted variability in reward pursuit, while externalizing profiles also predicted reward evaluation. These replicable results suggest that WebSurf and similar paradigms offer promising platforms for computational psychological methods, providing reliable, clinically relevant metrics of decision-making that may enhance psychiatric assessment and personalize treatment approaches.
{"title":"Psychometric Validation and Preliminary Clinical Correlation of an Experiential Foraging Task.","authors":"Aaron N McInnes, Christi R P Sullivan, Angus W MacDonald, Alik S Widge","doi":"10.1177/10731911251376214","DOIUrl":"10.1177/10731911251376214","url":null,"abstract":"<p><p>Measuring the function of decision-making systems reliably is a key goal to assess cognitive functions that underlie psychopathology. However, few metrics are demonstrably reliable, clinically relevant, and able to capture complex overlapping cognitive domains while quantifying heterogeneity across individuals. The WebSurf task is a reverse-translational human experiential foraging paradigm that indexes naturalistic and clinically relevant decision-making. To determine its potential clinical utility, we examined the psychometric properties and clinical correlates of behavioral parameters extracted from WebSurf in an initial exploratory experiment (<i>N</i> = 132) and a preregistered validation experiment (N = 109). Behavior was stable over repeated administrations of the task, as were individual differences. The ability to measure decision-making consistently supports WebSurf's potential utility to predict treatment response, monitor clinical change, and define neurocognitive profiles associated with psychopathology. Moreover, specific WebSurf metrics were predicted by psychiatric symptoms in a replicable manner. Mania and externalizing symptom profiles predicted variability in reward pursuit, while externalizing profiles also predicted reward evaluation. These replicable results suggest that WebSurf and similar paradigms offer promising platforms for computational psychological methods, providing reliable, clinically relevant metrics of decision-making that may enhance psychiatric assessment and personalize treatment approaches.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251376214"},"PeriodicalIF":3.4,"publicationDate":"2025-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145136014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-24DOI: 10.1177/10731911251367142
Angelina Pei-Tzu Tsai, Peter F Halpin, Lucy Lurie, Meredith Gruhn, Maya Rosen, Donald H Baucom, Michael B Sarabosing, Sneha Sai Boda, Katie A McLaughlin, Margaret A Sheridan
Research on the developmental consequences of early adversity has grown rapidly, yet measures of childhood adversity have not kept pace with evolving theoretical models. Existing measures often lack comprehensive assessment and psychometric evidence. This study addresses these gaps by developing the Deprivation and Threat-Adult Self-report (DT-AS) measure, a psychometrically sound scale assessing childhood threat and deprivation exposure, evaluated in young adults. Psychometric analysis was performed in waves on a total sample of N = 796 participants. Pilot data (n1 = 210; n2 = 208) were analyzed using Classical Test Theory (CTT) and Item Response Theory (IRT) to refine item selection and optimize response formats. The final sample (n3 = 378) confirmed a correlated factor structure of threat and deprivation with excellent psychometric properties. DT-AS consists of 33 items measuring threat and 30 measuring deprivation, offering a robust tool to examine associations between childhood adversity and psychopathology outcomes in adulthood.
{"title":"The Development and Validation of a Dimensional Childhood Adversity Measure.","authors":"Angelina Pei-Tzu Tsai, Peter F Halpin, Lucy Lurie, Meredith Gruhn, Maya Rosen, Donald H Baucom, Michael B Sarabosing, Sneha Sai Boda, Katie A McLaughlin, Margaret A Sheridan","doi":"10.1177/10731911251367142","DOIUrl":"https://doi.org/10.1177/10731911251367142","url":null,"abstract":"<p><p>Research on the developmental consequences of early adversity has grown rapidly, yet measures of childhood adversity have not kept pace with evolving theoretical models. Existing measures often lack comprehensive assessment and psychometric evidence. This study addresses these gaps by developing the Deprivation and Threat-Adult Self-report (DT-AS) measure, a psychometrically sound scale assessing childhood threat and deprivation exposure, evaluated in young adults. Psychometric analysis was performed in waves on a total sample of <i>N</i> = 796 participants. Pilot data (<i>n</i><sub>1</sub> = 210; <i>n</i><sub>2</sub> = 208) were analyzed using Classical Test Theory (CTT) and Item Response Theory (IRT) to refine item selection and optimize response formats. The final sample (<i>n</i><sub>3</sub> = 378) confirmed a correlated factor structure of threat and deprivation with excellent psychometric properties. DT-AS consists of 33 items measuring threat and 30 measuring deprivation, offering a robust tool to examine associations between childhood adversity and psychopathology outcomes in adulthood.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251367142"},"PeriodicalIF":3.4,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145136024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-24DOI: 10.1177/10731911251369977
Soo Youn Kim, Eric A Youngstrom, Megan Norris, Ann Levine, Eric M Butter, Kevin G Stephenson
There is limited information concerning the presence of empirically derived, person-centered latent cognitive profiles in youth with autism spectrum disorder (ASD) or attention-deficit/hyperactivity disorder (ADHD) and whether these profiles are diagnostically useful. The aim of this study was to identify empirically driven cognitive subgroups in youth with ASD or ADHD and examine predictors of those profiles. A retrospective chart review was conducted with patients seen at a developmental assessment clinic who were identified with ASD or ADHD aged 2 to 16 years (n = 1,679, Mage = 8.4, SDage = 3.1). A Latent Profile Analysis with Stanford-Binet-Fifth Edition composites resulted in 14 profiles, which were roughly parallel to each other across various levels of cognitive functioning. Several profiles were characterized by a relatively large discrepancy between the Nonverbal IQ and Verbal IQ. Younger age and higher IQ were significant predictors of those with scattered profiles, whereas diagnoses (i.e., ASD or ADHD), sex, and emotional-behavioral functioning were not.
{"title":"Identifying Latent Cognitive Profiles in Autism Spectrum Disorder and Attention-Deficit/Hyperactivity Disorder Using the Stanford-Binet Intelligence Scales-5th Edition.","authors":"Soo Youn Kim, Eric A Youngstrom, Megan Norris, Ann Levine, Eric M Butter, Kevin G Stephenson","doi":"10.1177/10731911251369977","DOIUrl":"https://doi.org/10.1177/10731911251369977","url":null,"abstract":"<p><p>There is limited information concerning the presence of empirically derived, person-centered latent cognitive profiles in youth with autism spectrum disorder (ASD) or attention-deficit/hyperactivity disorder (ADHD) and whether these profiles are diagnostically useful. The aim of this study was to identify empirically driven cognitive subgroups in youth with ASD or ADHD and examine predictors of those profiles. A retrospective chart review was conducted with patients seen at a developmental assessment clinic who were identified with ASD or ADHD aged 2 to 16 years (<i>n</i> = 1,679, <i>M</i><sub>age</sub> = 8.4, <i>SD</i><sub>age</sub> = 3.1). A Latent Profile Analysis with Stanford-Binet-Fifth Edition composites resulted in 14 profiles, which were roughly parallel to each other across various levels of cognitive functioning. Several profiles were characterized by a relatively large discrepancy between the Nonverbal IQ and Verbal IQ. Younger age and higher IQ were significant predictors of those with scattered profiles, whereas diagnoses (i.e., ASD or ADHD), sex, and emotional-behavioral functioning were not.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251369977"},"PeriodicalIF":3.4,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145136053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-24DOI: 10.1177/10731911251371395
Marieke E Timmerman, Annelies De Bildt, Julian Urban
Psychological test manuals vary widely in their reporting of the construction and interpretation of standardized scores. Consequently, the critical evaluation of norm quality and meaning is difficult for test users and reviewers. Because a specific standard for reporting on standardized scores is lacking, we developed Guidelines for Reporting on Norm-referenced and Criterion-referenced Scores (GRoNC), following a systematic approach for creating reporting guidelines (EQUATOR). The development took place in two stages: Stage 1, developing a preliminary version of the GRoNC based on a literature review; Stage 2, a Delphi process in two rounds, involving both theoretical experts (n = 11) and test developers (n = 14). The GRoNC includes a series of questions and associated explanations. It supports test developers in developing and reporting upon their standardized scores, and reviewers in evaluating a psychological test on its standardized scores. We provide recommendations on using the GRoNC and conclude by describing our expectations and plans to increase the impact of the GRoNC on reporting practice.
{"title":"The GRoNC: Guidelines for Reporting on Norm-Referenced and Criterion-Referenced Scores.","authors":"Marieke E Timmerman, Annelies De Bildt, Julian Urban","doi":"10.1177/10731911251371395","DOIUrl":"https://doi.org/10.1177/10731911251371395","url":null,"abstract":"<p><p>Psychological test manuals vary widely in their reporting of the construction and interpretation of standardized scores. Consequently, the critical evaluation of norm quality and meaning is difficult for test users and reviewers. Because a specific standard for reporting on standardized scores is lacking, we developed Guidelines for Reporting on Norm-referenced and Criterion-referenced Scores (GRoNC), following a systematic approach for creating reporting guidelines (EQUATOR). The development took place in two stages: Stage 1, developing a preliminary version of the GRoNC based on a literature review; Stage 2, a Delphi process in two rounds, involving both theoretical experts (<i>n =</i> 11) and test developers (<i>n =</i> 14). The GRoNC includes a series of questions and associated explanations. It supports test developers in developing and reporting upon their standardized scores, and reviewers in evaluating a psychological test on its standardized scores. We provide recommendations on using the GRoNC and conclude by describing our expectations and plans to increase the impact of the GRoNC on reporting practice.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251371395"},"PeriodicalIF":3.4,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145136163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-20DOI: 10.1177/10731911251364022
Zhuojun Gu, Katarina Kjell, H Andrew Schwartz, Oscar Kjell
Large language models can transform individuals' mental health descriptions into scores that correlate with rating scales approaching theoretical upper limits. However, such analyses have combined word- and text responses with little known about their differences. We develop response formats ranging from closed-ended to open-ended: (a) select words from lists, write (b) descriptive words, (c) phrases, or (d) texts. Participants answered questions about their depression/worry using the response formats and related rating scales. Language responses were transformed into word embeddings and trained to rating scales. We compare the validity (concurrent, incremental, face, discriminant, and external validity) and reliability (prospective sample and test-retest reliability) of the response formats. Using the Sequential Evaluation with Model Pre-Registration design, machine-learning models were trained on a development dataset (N = 963), and then pre-registered before tested on a prospective sample (N = 145). The pre-registered models demonstrate strong validity and reliability, yielding high accuracy in the prospective sample (r= .60-.79). Additionally, the models demonstrated external validity to self-reported sick-leave/healthcare visits, where the text-format yielded the strongest correlations (being higher/equal to rating scales for 9 of 12 cases). The overall high validity and reliability across formats suggest the possibility of choosing formats according to clinical needs.
{"title":"Natural Language Response Formats for Assessing Depression and Worry With Large Language Models: A Sequential Evaluation With Model Pre-Registration.","authors":"Zhuojun Gu, Katarina Kjell, H Andrew Schwartz, Oscar Kjell","doi":"10.1177/10731911251364022","DOIUrl":"https://doi.org/10.1177/10731911251364022","url":null,"abstract":"<p><p>Large language models can transform individuals' mental health descriptions into scores that correlate with rating scales approaching theoretical upper limits. However, such analyses have combined word- and text responses with little known about their differences. We develop response formats ranging from closed-ended to open-ended: (a) select words from lists, write (b) descriptive words, (c) phrases, or (d) texts. Participants answered questions about their depression/worry using the response formats and related rating scales. Language responses were transformed into word embeddings and trained to rating scales. We compare the validity (concurrent, incremental, face, discriminant, and external validity) and reliability (prospective sample and test-retest reliability) of the response formats. Using the <i>Sequential Evaluation with Model Pre-Registration</i> design, machine-learning models were trained on a development dataset (<i>N</i> = 963), and then <i>pre-registered</i> before tested on a prospective sample (<i>N</i> = 145). The pre-registered models demonstrate strong validity and reliability, yielding high accuracy in the prospective sample (<i>r</i> <i>=</i> .60-.79). Additionally, the models demonstrated external validity to self-reported sick-leave/healthcare visits, where the text-format yielded the strongest correlations (being higher/equal to rating scales for 9 of 12 cases). The overall high validity and reliability across formats suggest the possibility of choosing formats according to clinical needs.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251364022"},"PeriodicalIF":3.4,"publicationDate":"2025-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145091186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-16DOI: 10.1177/10731911251367122
Xingjian Ruan, Femke Vergeer-Hagoort, Margreet Ten Have, Annemarie I Luik, Marlous Tuithof, Ernst Bohlmeijer, Peter M Ten Klooster
The 5-item Brief INSPIRE-O instrument, based on the Connectedness, Hope, Identity, Meaning in Life, and Empowerment framework, is a novel tool to assess personal recovery. Although initially developed for clinical populations, its conceptual alignment with core dimensions of psychological well-being suggests its potential applicability to a broader audience. The current study aimed to examine its validity, reliability, and measurement invariance across people with and without common mental disorders (CMDs). The scale was administered in a Dutch general population sample (n = 5,451). Confirmatory factor analyses supported a unidimensional structure with robust factor loadings and scalar invariance across individuals with and without CMDs in the past year. In addition, the Brief INPSIRE-O showed acceptable reliability (ω = .71-.78) and the expected pattern of correlations with other health indicators supported its construct validity. In conclusion, the Brief INSPIRE-O appears to be a psychometrically sound measure of positive psychological functioning that can be validly used and compared across people with and without CMDs.
{"title":"Personal Recovery in the General Population: Comparison of Psychometric Properties of the Brief INSPIRE-O in Those With and Without Common Mental Disorders.","authors":"Xingjian Ruan, Femke Vergeer-Hagoort, Margreet Ten Have, Annemarie I Luik, Marlous Tuithof, Ernst Bohlmeijer, Peter M Ten Klooster","doi":"10.1177/10731911251367122","DOIUrl":"https://doi.org/10.1177/10731911251367122","url":null,"abstract":"<p><p>The 5-item Brief INSPIRE-O instrument, based on the Connectedness, Hope, Identity, Meaning in Life, and Empowerment framework, is a novel tool to assess personal recovery. Although initially developed for clinical populations, its conceptual alignment with core dimensions of psychological well-being suggests its potential applicability to a broader audience. The current study aimed to examine its validity, reliability, and measurement invariance across people with and without common mental disorders (CMDs). The scale was administered in a Dutch general population sample (<i>n</i> = 5,451). Confirmatory factor analyses supported a unidimensional structure with robust factor loadings and scalar invariance across individuals with and without CMDs in the past year. In addition, the Brief INPSIRE-O showed acceptable reliability (ω = .71-.78) and the expected pattern of correlations with other health indicators supported its construct validity. In conclusion, the Brief INSPIRE-O appears to be a psychometrically sound measure of positive psychological functioning that can be validly used and compared across people with and without CMDs.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251367122"},"PeriodicalIF":3.4,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145074244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-16DOI: 10.1177/10731911251365744
Hui Dong, Anselm B M Fuermaier, Janneke Koerts, Gerdina H M Pijnenborg, Nana Guo, Ragnar Schwierczok, Norbert Scherbaum, Bernhard W Müller
Impulsivity in adult attention-deficit/hyperactivity disorder (ADHD) represents a multidimensional construct rather than a unitary trait. This study examined a proposed three-factor model of impulsivity comprising (a) self-reported impulsive behavior (Barratt Impulsiveness Scale), (b) commission errors, and (c) reaction time measures from neuropsychological tests in 654 adults undergoing routine clinical assessment of adult ADHD. Using confirmatory factor analyses on split subsamples, we found consistent support for the proposed three-factor structure, whereas the network analysis favored a two-group conceptualization that separates performance-based from self-report-based measures. Self-reported impulsivity demonstrated the highest severity levels, followed by commission errors, with reaction times being least affected. Demographic and clinical characteristics significantly predicted self-reports and commission error measures but not reaction times. The results emphasize the importance of interpreting self-reports independently of performance-based tests. The coherence between commission errors and reaction time variables across tasks of related constructs suggests that administering multiple tasks may yield redundant information in the clinical assessment of impulsivity.
{"title":"Indicators of Impulsivity in Routine Clinical Assessment of Adult ADHD.","authors":"Hui Dong, Anselm B M Fuermaier, Janneke Koerts, Gerdina H M Pijnenborg, Nana Guo, Ragnar Schwierczok, Norbert Scherbaum, Bernhard W Müller","doi":"10.1177/10731911251365744","DOIUrl":"https://doi.org/10.1177/10731911251365744","url":null,"abstract":"<p><p>Impulsivity in adult attention-deficit/hyperactivity disorder (ADHD) represents a multidimensional construct rather than a unitary trait. This study examined a proposed three-factor model of impulsivity comprising (a) self-reported impulsive behavior (Barratt Impulsiveness Scale), (b) commission errors, and (c) reaction time measures from neuropsychological tests in 654 adults undergoing routine clinical assessment of adult ADHD. Using confirmatory factor analyses on split subsamples, we found consistent support for the proposed three-factor structure, whereas the network analysis favored a two-group conceptualization that separates performance-based from self-report-based measures. Self-reported impulsivity demonstrated the highest severity levels, followed by commission errors, with reaction times being least affected. Demographic and clinical characteristics significantly predicted self-reports and commission error measures but not reaction times. The results emphasize the importance of interpreting self-reports independently of performance-based tests. The coherence between commission errors and reaction time variables across tasks of related constructs suggests that administering multiple tasks may yield redundant information in the clinical assessment of impulsivity.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251365744"},"PeriodicalIF":3.4,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145074206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-10DOI: 10.1177/10731911251361230
Eric Sucitra, Riangga Novrianto, Yolanda T Pasaribu, Tania M Lincoln, Edo S Jaya
The Patient Health Questionnaire-9 (PHQ-9) is a screening tool for assessing depressive symptomatology that has received widespread use. However, there is a scarcity of research on whether the instrument measures the same construct between high-income (HIC) and low- and middle-income countries (LMICs). Online surveys were utilized to assess samples across Indonesia, Germany, and the USA (N = 2350). Measurement Invariance (MI) was computed using multi-group confirmatory factor analyses. We found the general factor model to have a good fit and configural, metric, scalar, and residual MI across three countries. There were no significant differences in mean scores (Indonesia, M = 1.87, SD = 0.56; Germany, M = 1.90, SD = 0.65; USA, M = 1.90, SD = 0.75). These results highlight that depressive symptomatology is universal across distinct geographical regions, regardless of the population's income levels. Hence, this study further emphasizes the urgency of developing universal, accessible assessment and treatment for depression.
{"title":"An Analysis of Measurement Invariance of the Patient Health Questionnaire-9 Between Indonesia, Germany, and the USA.","authors":"Eric Sucitra, Riangga Novrianto, Yolanda T Pasaribu, Tania M Lincoln, Edo S Jaya","doi":"10.1177/10731911251361230","DOIUrl":"https://doi.org/10.1177/10731911251361230","url":null,"abstract":"<p><p>The Patient Health Questionnaire-9 (PHQ-9) is a screening tool for assessing depressive symptomatology that has received widespread use. However, there is a scarcity of research on whether the instrument measures the same construct between high-income (HIC) and low- and middle-income countries (LMICs). Online surveys were utilized to assess samples across Indonesia, Germany, and the USA (<i>N</i> = 2350). Measurement Invariance (MI) was computed using multi-group confirmatory factor analyses. We found the general factor model to have a good fit and configural, metric, scalar, and residual MI across three countries. There were no significant differences in mean scores (Indonesia, <i>M</i> = 1.87, <i>SD</i> = 0.56; Germany, <i>M</i> = 1.90, <i>SD</i> = 0.65; USA, <i>M</i> = 1.90, <i>SD</i> = 0.75). These results highlight that depressive symptomatology is universal across distinct geographical regions, regardless of the population's income levels. Hence, this study further emphasizes the urgency of developing universal, accessible assessment and treatment for depression.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251361230"},"PeriodicalIF":3.4,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145028827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-09-10DOI: 10.1177/10731911251357466
Jinsoo Choi, Nanhee Kim, Bo Zhang, Sang Woo Park, Seonghee Cho, Young Woo Sohn, Christopher J Soto, Oliver P John
The Big Five Inventory-2 (BFI-2) has enjoyed global popularity due to its good balance of content coverage and brevity and has been officially translated into 12 languages in addition to the original English version. The current study aimed to further enhance the cultural accessibility of the BFI-2 by translating it into the Korean language and comprehensively validating the Korean version in two South Korean samples: working adults and college students. Across the two samples, the Korean BFI-2 demonstrated good reliability (e.g., test-retest reliability, Cronbach's alpha), construct validity (e.g., convergent/discriminant validity), and criterion-related validity with a wide range of outcome measures. Additionally, we compared the psychometric properties of the Korean BFI-2 to those of the original English and Chinese versions to further establish its comparability with other language versions. Overall, our results demonstrate that the Korean BFI-2 is a reliable and valid personality measure that can be confidently used by other researchers. Implications, limitations, and future directions were discussed.
《五大目录-2》(Big Five Inventory-2,简称BFI-2)在内容覆盖面和简练之间取得了良好的平衡,在全球广受欢迎,除英文原版外,还被正式翻译成12种语言。本研究旨在进一步提高BFI-2的文化可及性,将其翻译成韩语,并在两个韩国样本中进行全面验证:工作成年人和大学生。在两个样本中,韩国BFI-2表现出良好的信度(例如,测试-重测信度,Cronbach's alpha),结构效度(例如,收敛/判别效度),以及与广泛结果测量相关的标准效度。此外,我们还比较了韩文BFI-2与中英文原版的心理测量特征,以进一步确定其与其他语言版本的可比性。总体而言,我们的研究结果表明,韩国人的BFI-2是一个可靠和有效的人格测量,可以被其他研究者自信地使用。讨论了影响、局限性和未来的发展方向。
{"title":"The Big Five Inventory-2 in Korea: Validation and Cross-Cultural Comparisons with the U.S. and Chinese Versions.","authors":"Jinsoo Choi, Nanhee Kim, Bo Zhang, Sang Woo Park, Seonghee Cho, Young Woo Sohn, Christopher J Soto, Oliver P John","doi":"10.1177/10731911251357466","DOIUrl":"https://doi.org/10.1177/10731911251357466","url":null,"abstract":"<p><p>The Big Five Inventory-2 (BFI-2) has enjoyed global popularity due to its good balance of content coverage and brevity and has been officially translated into 12 languages in addition to the original English version. The current study aimed to further enhance the cultural accessibility of the BFI-2 by translating it into the Korean language and comprehensively validating the Korean version in two South Korean samples: working adults and college students. Across the two samples, the Korean BFI-2 demonstrated good reliability (e.g., test-retest reliability, Cronbach's alpha), construct validity (e.g., convergent/discriminant validity), and criterion-related validity with a wide range of outcome measures. Additionally, we compared the psychometric properties of the Korean BFI-2 to those of the original English and Chinese versions to further establish its comparability with other language versions. Overall, our results demonstrate that the Korean BFI-2 is a reliable and valid personality measure that can be confidently used by other researchers. Implications, limitations, and future directions were discussed.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251357466"},"PeriodicalIF":3.4,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145028861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}