Pub Date : 2024-09-01Epub Date: 2024-07-18DOI: 10.1037/pas0001327
Haley E Green, Lindsay N Gabel, Emma K Stewart, Yuliya Kotelnikova, Elizabeth P Hayden
Measurement tools from which valid interpretations can be made are critical for assessing early emerging depressive symptoms, as depressive symptoms in childhood are associated with increased risk for early-onset depressive disorder, recurrence, suicidality, and other psychopathology. The Depression Self-Rating Scale for Children (DSRS) is a widely used self-report scale assessing youth depressive symptoms. The relatively few studies investigating the DSRS' latent structure have yielded mixed results, and measurement invariance (MI) based on sex and age has not been examined. We examined the factor structure and MI of the DSRS across sex and age in a community sample of 6-9-year-olds (N = 352; Mage = 7.57 years, SD = .70). Consistent with the largest prior structural study of the DSRS, a two-factor structure, with factors reflecting elevated negative affect (NA) and low positive affect (PA), showed strong model fit. Although this structure was consistent across sex and age (i.e., configural invariance), loadings of DSRS items varied across sex and age (i.e., metric noninvariance). Allowing the loadings of items contributing to noninvariance to vary across groups improved model fit. Implications for the clinical and research utility of the DSRS and suggestions for future research are discussed. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Latent structure and measurement invariance of the Depression Self-Rating Scale for Children across sex and age.","authors":"Haley E Green, Lindsay N Gabel, Emma K Stewart, Yuliya Kotelnikova, Elizabeth P Hayden","doi":"10.1037/pas0001327","DOIUrl":"10.1037/pas0001327","url":null,"abstract":"<p><p>Measurement tools from which valid interpretations can be made are critical for assessing early emerging depressive symptoms, as depressive symptoms in childhood are associated with increased risk for early-onset depressive disorder, recurrence, suicidality, and other psychopathology. The Depression Self-Rating Scale for Children (DSRS) is a widely used self-report scale assessing youth depressive symptoms. The relatively few studies investigating the DSRS' latent structure have yielded mixed results, and measurement invariance (MI) based on sex and age has not been examined. We examined the factor structure and MI of the DSRS across sex and age in a community sample of 6-9-year-olds (<i>N</i> = 352; <i>M</i><sub>age</sub> = 7.57 years, <i>SD</i> = .70). Consistent with the largest prior structural study of the DSRS, a two-factor structure, with factors reflecting elevated negative affect (NA) and low positive affect (PA), showed strong model fit. Although this structure was consistent across sex and age (i.e., configural invariance), loadings of DSRS items varied across sex and age (i.e., metric noninvariance). Allowing the loadings of items contributing to noninvariance to vary across groups improved model fit. Implications for the clinical and research utility of the DSRS and suggestions for future research are discussed. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"552-561"},"PeriodicalIF":3.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141634303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-01Epub Date: 2024-06-20DOI: 10.1037/pas0001328
Sophie M Kudryk, Jolie T K Ho, Joshua R C Budge, David A Moscovitch
The use of analogue samples, as opposed to clinical groups, is common in mental health research, including research on social anxiety disorder (SAD). Recent observational and statistical evidence has raised doubts about the validity of current methods for establishing analogue samples of individuals with clinically significant social anxiety. Here, we used data from large community samples of clinical and nonclinical participants to determine new cutoff scores on self-report measures of social anxiety symptoms and symptom-related impairment. We then examined whether using these newly determined cutoff scores alone or in combination improves the identification of individuals who have SAD from those who do not, revealing the most ideal cutoff combination to be 34 or above on the Social Phobia Inventory and 11 or above on the Sheehan Disability Scale. Finally, we compared the effects of our new cutoff scores with old cutoff scores by extracting analogue samples of participants with high social anxiety from historical data on seven large groups of undergraduate Psychology research participants from the authors' institution spanning the past 5 years (2018-2023). We observed that the new combined cutoff scores identified markedly fewer students as having high social anxiety, lending credibility to their utility. We also observed a striking increase in levels of social anxiety symptoms in the undergraduate population from before to after the COVID-19 pandemic. Of note, most participants were under 30 and identified as Caucasian or Asian women, indicating that future research is needed to examine whether our findings generalize to diverse populations. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Identifying analogue samples of individuals with clinically significant social anxiety: Updating and combining cutoff scores on the Social Phobia Inventory and Sheehan Disability Scale.","authors":"Sophie M Kudryk, Jolie T K Ho, Joshua R C Budge, David A Moscovitch","doi":"10.1037/pas0001328","DOIUrl":"10.1037/pas0001328","url":null,"abstract":"<p><p>The use of analogue samples, as opposed to clinical groups, is common in mental health research, including research on social anxiety disorder (SAD). Recent observational and statistical evidence has raised doubts about the validity of current methods for establishing analogue samples of individuals with clinically significant social anxiety. Here, we used data from large community samples of clinical and nonclinical participants to determine new cutoff scores on self-report measures of social anxiety symptoms and symptom-related impairment. We then examined whether using these newly determined cutoff scores alone or in combination improves the identification of individuals who have SAD from those who do not, revealing the most ideal cutoff combination to be 34 or above on the Social Phobia Inventory and 11 or above on the Sheehan Disability Scale. Finally, we compared the effects of our new cutoff scores with old cutoff scores by extracting analogue samples of participants with high social anxiety from historical data on seven large groups of undergraduate Psychology research participants from the authors' institution spanning the past 5 years (2018-2023). We observed that the new combined cutoff scores identified markedly fewer students as having high social anxiety, lending credibility to their utility. We also observed a striking increase in levels of social anxiety symptoms in the undergraduate population from before to after the COVID-19 pandemic. Of note, most participants were under 30 and identified as Caucasian or Asian women, indicating that future research is needed to examine whether our findings generalize to diverse populations. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"513-525"},"PeriodicalIF":3.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141427456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Su-Jong Kim-Dorner,Heike Saßmann,Juliane R Framme,Bettina Heidtmann,Thomas M Kapellen,Olga Kordonouri,Karolin M E Nettelrodt,Nicole Pisarek,Roland Schweizer,Simone von Sengbusch,Karin Lange
Children with Type 1 diabetes (T1D) and their parent-caregivers often experience diabetes distress due to the daily demands of diabetes management. Regular screening for diabetes distress is needed to prevent the deterioration of metabolic control and the development of mental health disorders. The aim of this analysis was to examine the psychometric properties of the German versions of the Problem Areas in Diabetes Scale for Children (PAID-C) and for caregiver burden in Parents (P-PAID-C). Data were collected from 136 children aged 7-12 years (46.7% females) and 304 parents (Mage = 42.9 (SD 6.1) years; 78% mothers) by using linguistically translated questionnaires in a multicenter study. Confirmatory factor analysis and correlational analyses were conducted. Results confirmed the two-factor model for the PAID-C and the four-factor model for the P-PAID-C with a slight modification. Cronbach's αs for children and parents were 0.88 and 0.92, respectively. The PAID-C and P-PAID-C scores had small positive associations with HbA1c (rs = .220 and .139, respectively, all p < .05) and strong inverse association with the KIDSCREEN-10 index (r = -.643 and -.520, respectively, all p < .001). P-PAID-C scores increased with increasing depressive symptoms measured in nine-item Patient Health Questionnaire among parents (rs = .534, p < .001). The scores produced by the German PAID-C and P-PAID-C were reliable and valid in measuring diabetes burdens. These German versions of PAID can be utilized to assess diabetes-specific distress and to design interventions for children and their parents experiencing high levels of diabetes distress. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Psychometric properties of the German versions of the Problem Areas in Diabetes Scale for Children (PAID-C) with Type 1 Diabetes and Their Parents (P-PAID-C).","authors":"Su-Jong Kim-Dorner,Heike Saßmann,Juliane R Framme,Bettina Heidtmann,Thomas M Kapellen,Olga Kordonouri,Karolin M E Nettelrodt,Nicole Pisarek,Roland Schweizer,Simone von Sengbusch,Karin Lange","doi":"10.1037/pas0001338","DOIUrl":"https://doi.org/10.1037/pas0001338","url":null,"abstract":"Children with Type 1 diabetes (T1D) and their parent-caregivers often experience diabetes distress due to the daily demands of diabetes management. Regular screening for diabetes distress is needed to prevent the deterioration of metabolic control and the development of mental health disorders. The aim of this analysis was to examine the psychometric properties of the German versions of the Problem Areas in Diabetes Scale for Children (PAID-C) and for caregiver burden in Parents (P-PAID-C). Data were collected from 136 children aged 7-12 years (46.7% females) and 304 parents (Mage = 42.9 (SD 6.1) years; 78% mothers) by using linguistically translated questionnaires in a multicenter study. Confirmatory factor analysis and correlational analyses were conducted. Results confirmed the two-factor model for the PAID-C and the four-factor model for the P-PAID-C with a slight modification. Cronbach's αs for children and parents were 0.88 and 0.92, respectively. The PAID-C and P-PAID-C scores had small positive associations with HbA1c (rs = .220 and .139, respectively, all p < .05) and strong inverse association with the KIDSCREEN-10 index (r = -.643 and -.520, respectively, all p < .001). P-PAID-C scores increased with increasing depressive symptoms measured in nine-item Patient Health Questionnaire among parents (rs = .534, p < .001). The scores produced by the German PAID-C and P-PAID-C were reliable and valid in measuring diabetes burdens. These German versions of PAID can be utilized to assess diabetes-specific distress and to design interventions for children and their parents experiencing high levels of diabetes distress. (PsycInfo Database Record (c) 2024 APA, all rights reserved).","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"19 1","pages":"e38-e50"},"PeriodicalIF":3.6,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142165993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-01Epub Date: 2024-06-20DOI: 10.1037/pas0001322
Emily C Kemp, James V Ray, Paul J Frick, Laura C Thornton, Tina D Wall Myers, Emily L Robertson, Laurence Steinberg, Elizabeth Cauffman
The Inventory of Callous-Unemotional Traits (ICU) is a widely used measure of callous-unemotional (CU) traits that may aid in the assessment of the diagnostic specifier "with limited prosocial emotions," which has been added to diagnostic criteria for conduct disorder. Though there is substantial support for use of the ICU total score, the scale's factor structure has been highly debated. Inconsistencies in past factor analyses may be largely attributed to failure to control for method variance due to item wording (i.e., half of the items being worded in the callous direction and half worded in the prosocial direction). Thus, the present study used a multitrait-multimethod confirmatory factor analytic approach that models both trait and method variance to test the factor structure of the ICU self-report in a clinically relevant, high-risk sample of justice-involved male adolescents (N = 1,216). When comparing the fit of empirical and theoretical models, goodness of fit indices (χ² = 1105.877, df = 190, root-mean-square error of approximation = .063, comparative fit index = .916, Tucker-Lewis index = .878, standardized root-mean-square residual = .051) provided support for a hierarchical four-factor model (i.e., one overarching callous-unemotional factor, four latent trait factors) when accounting for method variance (i.e., covarying positively worded items). This factor structure is consistent with the way the ICU was constructed and with criteria for the limited prosocial emotions specifier. In addition, measurement invariance of this factor structure across age, race, and ethnicity was supported, and the predictive validity of the ICU was supported across these demographic groups in predicting self-reported antisocial behavior and rearrests over a 5-year period following an adolescent's first arrest. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"The Inventory of Callous-Unemotional Traits (ICU) self-report version: Factor structure, measurement invariance, and predictive validity in justice-involved male adolescents.","authors":"Emily C Kemp, James V Ray, Paul J Frick, Laura C Thornton, Tina D Wall Myers, Emily L Robertson, Laurence Steinberg, Elizabeth Cauffman","doi":"10.1037/pas0001322","DOIUrl":"10.1037/pas0001322","url":null,"abstract":"<p><p>The Inventory of Callous-Unemotional Traits (ICU) is a widely used measure of callous-unemotional (CU) traits that may aid in the assessment of the diagnostic specifier \"with limited prosocial emotions,\" which has been added to diagnostic criteria for conduct disorder. Though there is substantial support for use of the ICU total score, the scale's factor structure has been highly debated. Inconsistencies in past factor analyses may be largely attributed to failure to control for method variance due to item wording (i.e., half of the items being worded in the callous direction and half worded in the prosocial direction). Thus, the present study used a multitrait-multimethod confirmatory factor analytic approach that models both trait and method variance to test the factor structure of the ICU self-report in a clinically relevant, high-risk sample of justice-involved male adolescents (<i>N</i> = 1,216). When comparing the fit of empirical and theoretical models, goodness of fit indices (χ² = 1105.877, <i>df</i> = 190, root-mean-square error of approximation = .063, comparative fit index = .916, Tucker-Lewis index = .878, standardized root-mean-square residual = .051) provided support for a hierarchical four-factor model (i.e., one overarching callous-unemotional factor, four latent trait factors) when accounting for method variance (i.e., covarying positively worded items). This factor structure is consistent with the way the ICU was constructed and with criteria for the limited prosocial emotions specifier. In addition, measurement invariance of this factor structure across age, race, and ethnicity was supported, and the predictive validity of the ICU was supported across these demographic groups in predicting self-reported antisocial behavior and rearrests over a 5-year period following an adolescent's first arrest. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"562-571"},"PeriodicalIF":3.3,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141427460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-05-23DOI: 10.1037/pas0001319
Lindsey C Stewart, Shayan Asadi, Craig Rodriguez-Seijas, Sylia Wilson, Giorgia Michelini, Roman Kotov, David C Cicero, Thomas M Olino
There are numerous studies examining differences in the experience of disorders and symptoms of psychopathology in adolescents across racial or ethnic groups and sex. Though there is substantial research exploring potential factors that may influence these differences, few studies have considered the potential contribution of measurement properties to these differences. Therefore, this study examined whether there are differences across racial or ethnic groups and sex in the measurement of psychopathology, assessed in mother-reported behavior of 9-11 year old youth from the Adolescent Brain Cognitive Development study sample using updated Child Behavior Checklist scales (CBCL; Achenbach & Rescorla, 2001). Tests of measurement invariance of the CBCL utilized the higher order factor structure identified by Michelini et al. (2019) using this same Adolescent Brain Cognitive Development cohort. The dimensions include internalizing, somatoform, detachment, externalizing, and neurodevelopmental problems. The configural model had a good-to-excellent fit on all subscales of the CBCL across racial or ethnic groups and sex. The metric and scalar models fit just as well as the configural models, indicating that the scales are measuring the same constructs across racial or ethnic groups and sex and are not influenced by measurement properties of items on the CBCL, although some high-severity response options were not endorsed for youth in all racial or ethnic groups. These findings support the use of the CBCL in research examining psychopathology in racially or ethnically diverse samples of youth. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Measurement invariance of the Child Behavior Checklist (CBCL) across race/ethnicity and sex in the Adolescent Brain and Cognitive Development (ABCD) study.","authors":"Lindsey C Stewart, Shayan Asadi, Craig Rodriguez-Seijas, Sylia Wilson, Giorgia Michelini, Roman Kotov, David C Cicero, Thomas M Olino","doi":"10.1037/pas0001319","DOIUrl":"10.1037/pas0001319","url":null,"abstract":"<p><p>There are numerous studies examining differences in the experience of disorders and symptoms of psychopathology in adolescents across racial or ethnic groups and sex. Though there is substantial research exploring potential factors that may influence these differences, few studies have considered the potential contribution of measurement properties to these differences. Therefore, this study examined whether there are differences across racial or ethnic groups and sex in the measurement of psychopathology, assessed in mother-reported behavior of 9-11 year old youth from the Adolescent Brain Cognitive Development study sample using updated Child Behavior Checklist scales (CBCL; Achenbach & Rescorla, 2001). Tests of measurement invariance of the CBCL utilized the higher order factor structure identified by Michelini et al. (2019) using this same Adolescent Brain Cognitive Development cohort. The dimensions include internalizing, somatoform, detachment, externalizing, and neurodevelopmental problems. The configural model had a good-to-excellent fit on all subscales of the CBCL across racial or ethnic groups and sex. The metric and scalar models fit just as well as the configural models, indicating that the scales are measuring the same constructs across racial or ethnic groups and sex and are not influenced by measurement properties of items on the CBCL, although some high-severity response options were not endorsed for youth in all racial or ethnic groups. These findings support the use of the CBCL in research examining psychopathology in racially or ethnically diverse samples of youth. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"441-451"},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11801408/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141082122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-05-06DOI: 10.1037/pas0001320
Courtney M Goetz, Taylor A Miller, Paul J Frick
Recent changes to diagnostic criteria for serious conduct problems in children and adolescents have included the presence of elevated callous-unemotional traits to define etiologically and clinically important subgroups of youth with a conduct problem diagnosis. The Clinical Assessment of Prosocial Emotions (CAPE) is an intensive assessment of the symptoms of this limited prosocial emotions specifier that uses a structured professional judgment method of scoring, which may make it useful in clinical settings when diagnoses may require more information than that provided by behavior rating scales. The present study adds to the limited tests of the CAPE's reliability and validity, using a sample of clinic-referred children ages 6-17 years of age, who were all administered the CAPE by trained clinicians. The mean age of the sample was 10.13 years (SD = 2.64); 54% of the sample identified as male and 46% identified as female; and 67% of participants identified as White, 29% identified as Black, and 52% identified as another race/ethnicity (i.e., Asian, Hispanic/Latinx, or other). The findings indicated that CAPE scores demonstrated strong interrater reliability. The scores also were associated with measures of conduct problems and aggression, even when controlling for behavior ratings of callous-unemotional traits. Further, when children with conduct problem diagnoses were divided into groups based on the presence of the limited prosocial emotions specifier from the CAPE, the subgroup with the specifier showed more severe conduct problems and aggression. The results support cautious clinical use of the CAPE, its further development and testing, and research into ways to make its use feasible in many clinical settings. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"The Clinical Assessment of Prosocial Emotions (CAPE): Initial tests of reliability and validity in a clinic-referred sample of children and adolescents.","authors":"Courtney M Goetz, Taylor A Miller, Paul J Frick","doi":"10.1037/pas0001320","DOIUrl":"10.1037/pas0001320","url":null,"abstract":"<p><p>Recent changes to diagnostic criteria for serious conduct problems in children and adolescents have included the presence of elevated callous-unemotional traits to define etiologically and clinically important subgroups of youth with a conduct problem diagnosis. The Clinical Assessment of Prosocial Emotions (CAPE) is an intensive assessment of the symptoms of this limited prosocial emotions specifier that uses a structured professional judgment method of scoring, which may make it useful in clinical settings when diagnoses may require more information than that provided by behavior rating scales. The present study adds to the limited tests of the CAPE's reliability and validity, using a sample of clinic-referred children ages 6-17 years of age, who were all administered the CAPE by trained clinicians. The mean age of the sample was 10.13 years (<i>SD</i> = 2.64); 54% of the sample identified as male and 46% identified as female; and 67% of participants identified as White, 29% identified as Black, and 52% identified as another race/ethnicity (i.e., Asian, Hispanic/Latinx, or other). The findings indicated that CAPE scores demonstrated strong interrater reliability. The scores also were associated with measures of conduct problems and aggression, even when controlling for behavior ratings of callous-unemotional traits. Further, when children with conduct problem diagnoses were divided into groups based on the presence of the limited prosocial emotions specifier from the CAPE, the subgroup with the specifier showed more severe conduct problems and aggression. The results support cautious clinical use of the CAPE, its further development and testing, and research into ways to make its use feasible in many clinical settings. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"452-461"},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140851157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-05-16DOI: 10.1037/pas0001317
Daniel J Reis, Adam R Kinney, Jeri E Forster, Kelly A Stearns-Yoder, Julie A Kittel, Amanda E Wood, David W Oslin, Lisa A Brenner, Joseph A Simonetti
Comparing self-reported symptom scores across time requires longitudinal measurement invariance (LMI), a psychometric property that means the measure is functioning identically across all time points. Despite its prominence as a measure of depression symptom severity in both research and health care, LMI has yet to be firmly established for the Patient Health Questionnaire-9 depression module (PHQ-9), particularly over the course of antidepressant pharmacotherapy. Accordingly, the objective of this study was to assess for LMI of the PHQ-9 during pharmacotherapy for major depressive disorder. This was a secondary analysis of data collected during a randomized controlled trial. A total of 1,944 veterans began antidepressant monotherapy and completed the PHQ-9 six times over 24 weeks of treatment. LMI was assessed using a series of four confirmatory factor analysis models that included all six time points, with estimated parameters increasingly constrained across models to test for different aspects of invariance. Root-mean-square error of approximation of the chi-square difference test values below 0.06 indicated the presence of LMI. Exploratory LMI analyses were also performed for separate sex, age, and race subgroups. Root-mean-square error of approximation of the chi-square difference test showed minimal change in model fits during invariance testing (≤ 0.06 for all steps), supporting full LMI for the PHQ-9. LMI was also supported for all tested veteran subgroups. As such, PHQ-9 sum scores can be compared across extended pharmacotherapy treatment durations. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Longitudinal invariance of the Patient Health Questionnaire-9 among patients receiving pharmacotherapy for major depressive disorder: A secondary analysis of clinical trial data.","authors":"Daniel J Reis, Adam R Kinney, Jeri E Forster, Kelly A Stearns-Yoder, Julie A Kittel, Amanda E Wood, David W Oslin, Lisa A Brenner, Joseph A Simonetti","doi":"10.1037/pas0001317","DOIUrl":"10.1037/pas0001317","url":null,"abstract":"<p><p>Comparing self-reported symptom scores across time requires longitudinal measurement invariance (LMI), a psychometric property that means the measure is functioning identically across all time points. Despite its prominence as a measure of depression symptom severity in both research and health care, LMI has yet to be firmly established for the Patient Health Questionnaire-9 depression module (PHQ-9), particularly over the course of antidepressant pharmacotherapy. Accordingly, the objective of this study was to assess for LMI of the PHQ-9 during pharmacotherapy for major depressive disorder. This was a secondary analysis of data collected during a randomized controlled trial. A total of 1,944 veterans began antidepressant monotherapy and completed the PHQ-9 six times over 24 weeks of treatment. LMI was assessed using a series of four confirmatory factor analysis models that included all six time points, with estimated parameters increasingly constrained across models to test for different aspects of invariance. Root-mean-square error of approximation of the chi-square difference test values below 0.06 indicated the presence of LMI. Exploratory LMI analyses were also performed for separate sex, age, and race subgroups. Root-mean-square error of approximation of the chi-square difference test showed minimal change in model fits during invariance testing (≤ 0.06 for all steps), supporting full LMI for the PHQ-9. LMI was also supported for all tested veteran subgroups. As such, PHQ-9 sum scores can be compared across extended pharmacotherapy treatment durations. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"462-471"},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140945720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-06-20DOI: 10.1037/pas0001321
Keanan J Joyner, Keenan Roberts, Ashley L Watts, Kelsey L Lowman, Robert D Latzman, Scott O Lilienfeld, Christopher J Patrick
The triarchic model posits that distinct trait constructs of boldness, meanness, and disinhibition underlie psychopathy. The triarchic model traits are conceptualized as biobehavioral dimensions that can be assessed using different sets of indicators from alternative measurement modalities; as such, the triarchic model would hypothesize that these traits are not confined to any one item set. The present study tested whether the triarchic model dimensions would emerge from a hierarchical-structural analysis of the facet scales of the Elemental Psychopathy Assessment (EPA), an inventory designed to comprehensively index psychopathy according to the five-factor personality model. Study participants (Ns = 811, 170) completed the EPA and three different scale sets assessing the triarchic traits along with criterion measures of antisocial/externalizing behaviors. Bass-ackwards modeling of the EPA facet scales revealed a four-level structure, with factors at the third level appearing similar to the triarchic trait dimensions. An analysis in which scores for the Level-3 EPA factors were regressed onto corresponding latent-trait dimensions defined using the different triarchic scale sets revealed extremely high convergence (βs = .84-.91). The Level-3 EPA factors also evidenced validity in relation to relevant criteria, approximating and sometimes exceeding that evident for the Level-4 EPA factors. Together, these results indicate that the triarchic trait constructs are embedded in a psychopathy inventory designed to align with a general personality model and effectively predict pertinent external criteria. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Locating triarchic model constructs in the hierarchical structure of a comprehensive trait-based psychopathy measure: Implications for research and clinical assessment.","authors":"Keanan J Joyner, Keenan Roberts, Ashley L Watts, Kelsey L Lowman, Robert D Latzman, Scott O Lilienfeld, Christopher J Patrick","doi":"10.1037/pas0001321","DOIUrl":"10.1037/pas0001321","url":null,"abstract":"<p><p>The triarchic model posits that distinct trait constructs of boldness, meanness, and disinhibition underlie psychopathy. The triarchic model traits are conceptualized as biobehavioral dimensions that can be assessed using different sets of indicators from alternative measurement modalities; as such, the triarchic model would hypothesize that these traits are not confined to any one item set. The present study tested whether the triarchic model dimensions would emerge from a hierarchical-structural analysis of the facet scales of the Elemental Psychopathy Assessment (EPA), an inventory designed to comprehensively index psychopathy according to the five-factor personality model. Study participants (<i>N</i>s = 811, 170) completed the EPA and three different scale sets assessing the triarchic traits along with criterion measures of antisocial/externalizing behaviors. Bass-ackwards modeling of the EPA facet scales revealed a four-level structure, with factors at the third level appearing similar to the triarchic trait dimensions. An analysis in which scores for the Level-3 EPA factors were regressed onto corresponding latent-trait dimensions defined using the different triarchic scale sets revealed extremely high convergence (βs = .84-.91). The Level-3 EPA factors also evidenced validity in relation to relevant criteria, approximating and sometimes exceeding that evident for the Level-4 EPA factors. Together, these results indicate that the triarchic trait constructs are embedded in a psychopathy inventory designed to align with a general personality model and effectively predict pertinent external criteria. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"472-487"},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11879147/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141427457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-06-27DOI: 10.1037/pas0001323
Reza de Souza Brümmer, Karolin Rose Krause, Giovanni Abrahão Salum, Marcelo Pio de Almeida Fleck, Ighor Miron Porto, João Villanova do Amaral, João Pedro Gonçalves Pacheco, Bettina Moltrecht, Eoin McElroy, Mauricio Scopel Hoffmann
Assessment tools for depression and anxiety usually inquire about the frequency of symptoms. However, evidence suggests that different question framings might trigger different responses. Our aim is to test if asking about symptom's context, ability, duration, and botherment adds validity to Patient Health Questionnaire-9, General Anxiety Disorder-7, and Patient-Related Outcome Measurement Information Systems depression and anxiety. Participants came from two cross-sectional convenience-sampled surveys (N = 1,871) of adults (66% females, aged 33.4 ± 13.2), weighted to approximate with the state-level population. We examined measurement invariance across the different question frames, estimated whether framing affected mean scores, and tested their independent validity using covariate-adjusted and sample-weighted structural equation models. Validity was tested using tools assessing general disability, alcohol use, loneliness, well-being, grit, and frequency-based questions from depression and anxiety questionnaires. A bifactor model was applied to test the internal consistency of the question frames under the presence of a general factor (i.e., depression or anxiety). Measurement invariance was supported across the different frames. Framing questions as ability (i.e., "How easily …") produced a higher score, compared with framing by context (i.e., "In which daily situations …"). Construct and criterion validity analysis demonstrate that variance explained using multiple question frames was similar to using only one. We detected a strong overarching factor for each instrument, with little variances left to be explained by the question frame. Therefore, it is unlikely that using different adverbial phrasings can help clinicians and researchers to improve their ability to detect depression or anxiety. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
{"title":"Beyond frequency: Evaluating the validity of assessing the context, duration, ability, and botherment of depression and anxiety symptoms in South Brazil.","authors":"Reza de Souza Brümmer, Karolin Rose Krause, Giovanni Abrahão Salum, Marcelo Pio de Almeida Fleck, Ighor Miron Porto, João Villanova do Amaral, João Pedro Gonçalves Pacheco, Bettina Moltrecht, Eoin McElroy, Mauricio Scopel Hoffmann","doi":"10.1037/pas0001323","DOIUrl":"10.1037/pas0001323","url":null,"abstract":"<p><p>Assessment tools for depression and anxiety usually inquire about the frequency of symptoms. However, evidence suggests that different question framings might trigger different responses. Our aim is to test if asking about symptom's context, ability, duration, and botherment adds validity to Patient Health Questionnaire-9, General Anxiety Disorder-7, and Patient-Related Outcome Measurement Information Systems depression and anxiety. Participants came from two cross-sectional convenience-sampled surveys (<i>N</i> = 1,871) of adults (66% females, aged 33.4 ± 13.2), weighted to approximate with the state-level population. We examined measurement invariance across the different question frames, estimated whether framing affected mean scores, and tested their independent validity using covariate-adjusted and sample-weighted structural equation models. Validity was tested using tools assessing general disability, alcohol use, loneliness, well-being, grit, and frequency-based questions from depression and anxiety questionnaires. A bifactor model was applied to test the internal consistency of the question frames under the presence of a general factor (i.e., depression or anxiety). Measurement invariance was supported across the different frames. Framing questions as ability (i.e., \"How easily …\") produced a higher score, compared with framing by context (i.e., \"In which daily situations …\"). Construct and criterion validity analysis demonstrate that variance explained using multiple question frames was similar to using only one. We detected a strong overarching factor for each instrument, with little variances left to be explained by the question frame. Therefore, it is unlikely that using different adverbial phrasings can help clinicians and researchers to improve their ability to detect depression or anxiety. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"488-504"},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141458963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This article illustrates novel quantitative methods to estimate classification consistency in machine learning models used for screening measures. Screening measures are used in psychology and medicine to classify individuals into diagnostic classifications. In addition to achieving high accuracy, it is ideal for the screening process to have high classification consistency, which means that respondents would be classified into the same group every time if the assessment was repeated. Although machine learning models are increasingly being used to predict a screening classification based on individual item responses, methods to describe the classification consistency of machine learning models have not yet been developed. This article addresses this gap by describing methods to estimate classification inconsistency in machine learning models arising from two different sources: sampling error during model fitting and measurement error in the item responses. These methods use data resampling techniques such as the bootstrap and Monte Carlo sampling. These methods are illustrated using three empirical examples predicting a health condition/diagnosis from item responses. R code is provided to facilitate the implementation of the methods. This article highlights the importance of considering classification consistency alongside accuracy when studying screening measures and provides the tools and guidance necessary for applied researchers to obtain classification consistency indices in their machine learning research on diagnostic assessments. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
本文阐述了新颖的定量方法,用于估算筛查措施所用机器学习模型的分类一致性。筛查方法被用于心理学和医学领域,以将个体划分为诊断类别。除了要达到高准确度外,筛查过程还必须具有高分类一致性,这意味着如果重复进行评估,受访者每次都会被归入同一组别。尽管机器学习模型越来越多地被用于预测基于单个项目反应的筛选分类,但描述机器学习模型分类一致性的方法尚未开发出来。本文针对这一空白,介绍了估算机器学习模型分类不一致性的方法,这种不一致性由两个不同的来源引起:模型拟合过程中的抽样误差和项目回答中的测量误差。这些方法使用了数据重采样技术,如自举法和蒙特卡罗采样。这些方法通过三个从项目回答中预测健康状况/诊断的经验示例进行了说明。本文提供了 R 代码,以方便方法的实施。本文强调了在研究筛查措施时考虑分类一致性和准确性的重要性,并为应用研究人员在诊断评估的机器学习研究中获取分类一致性指数提供了必要的工具和指导。(PsycInfo 数据库记录 (c) 2024 APA,保留所有权利)。
{"title":"Estimating classification consistency of machine learning models for screening measures.","authors":"Oscar Gonzalez, A R Georgeson, William E Pelham","doi":"10.1037/pas0001313","DOIUrl":"https://doi.org/10.1037/pas0001313","url":null,"abstract":"<p><p>This article illustrates novel quantitative methods to estimate classification consistency in machine learning models used for screening measures. Screening measures are used in psychology and medicine to classify individuals into diagnostic classifications. In addition to achieving high accuracy, it is ideal for the screening process to have high classification consistency, which means that respondents would be classified into the same group every time if the assessment was repeated. Although machine learning models are increasingly being used to predict a screening classification based on individual item responses, methods to describe the classification consistency of machine learning models have not yet been developed. This article addresses this gap by describing methods to estimate classification inconsistency in machine learning models arising from two different sources: sampling error during model fitting and measurement error in the item responses. These methods use data resampling techniques such as the bootstrap and Monte Carlo sampling. These methods are illustrated using three empirical examples predicting a health condition/diagnosis from item responses. R code is provided to facilitate the implementation of the methods. This article highlights the importance of considering classification consistency alongside accuracy when studying screening measures and provides the tools and guidance necessary for applied researchers to obtain classification consistency indices in their machine learning research on diagnostic assessments. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"36 6-7","pages":"395-406"},"PeriodicalIF":3.6,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141200498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}