The estimates of intraclass correlations are known to be biased, but there are few analytical ways to assess the amount of bias. The analytical approach requires the normality assumption to estimate bias. Bootstrap requires no such assumption and can, therefore, be used to estimate bias, regardless of the model assumption. We utilize cluster bootstrapping to calculate the bias in estimating the intraclass correlation. A well-known dataset is provided to illustrate the bias estimation in a typical study design of intraclass correlation, and its implications for other study designs are also discussed.
{"title":"Bootstrap Estimate of Bias for Intraclass Correlation.","authors":"Xiaofeng Steven Liu, Kelvin Terrell Pompey","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>The estimates of intraclass correlations are known to be biased, but there are few analytical ways to assess the amount of bias. The analytical approach requires the normality assumption to estimate bias. Bootstrap requires no such assumption and can, therefore, be used to estimate bias, regardless of the model assumption. We utilize cluster bootstrapping to calculate the bias in estimating the intraclass correlation. A well-known dataset is provided to illustrate the bias estimation in a typical study design of intraclass correlation, and its implications for other study designs are also discussed.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 1","pages":"101-108"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"37704087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
To understand the role of fit statistics in Rasch measurement is simple: applied researchers can only benefit from the desirable properties of the Rasch model when the data fit the model. The purpose of the current study was to assess the Q-Index robustness (Ostini and Nering, 2006), and its performance was compared to the current popular fit statistics known as MSQ Infit, MSQ Outfit, and standardized Infit and Outfit (ZSTDs) under varying conditions of test length, sample size, item difficulty (normal and uniform), and dimensionality utilizing a Monte Carlo simulation. The Type I and Type II error rates are also examined across fit indices. This study provides applied researchers guidelines the robustness and appropriateness of the use of the Q-Index, which is an alternative to the currently available item fit statistics. The Q-Index was slightly more sensitive to the levels of multidimensionality set in the study while MSQ Infit, Outfit, and standardized Infit and Outfit (ZSTDs) failed to identify the multidimensional conditions. The Type I error rate of the Q-Index was lower than the rest of the fit indices; however, the Type II error rate was higher than the anticipated beta = .20 across all fit indices.
{"title":"Evaluating the Impact of Multidimensionality on Type I and Type II Error Rates using the Q-Index Item Fit Statistic for the Rasch Model.","authors":"Samantha Estrada","doi":"10.31219/osf.io/kh7vq","DOIUrl":"https://doi.org/10.31219/osf.io/kh7vq","url":null,"abstract":"To understand the role of fit statistics in Rasch measurement is simple: applied researchers can only benefit from the desirable properties of the Rasch model when the data fit the model. The purpose of the current study was to assess the Q-Index robustness (Ostini and Nering, 2006), and its performance was compared to the current popular fit statistics known as MSQ Infit, MSQ Outfit, and standardized Infit and Outfit (ZSTDs) under varying conditions of test length, sample size, item difficulty (normal and uniform), and dimensionality utilizing a Monte Carlo simulation. The Type I and Type II error rates are also examined across fit indices. This study provides applied researchers guidelines the robustness and appropriateness of the use of the Q-Index, which is an alternative to the currently available item fit statistics. The Q-Index was slightly more sensitive to the levels of multidimensionality set in the study while MSQ Infit, Outfit, and standardized Infit and Outfit (ZSTDs) failed to identify the multidimensional conditions. The Type I error rate of the Q-Index was lower than the rest of the fit indices; however, the Type II error rate was higher than the anticipated beta = .20 across all fit indices.","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 4 1","pages":"496-514"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"69636647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The purpose of this study is to demonstrate an application of the many-facet Rasch model (MFRM) in evaluating the impromptu speech skills of pre-service principals in Taiwan. The findings showed that the topics of speech did not exhibit different difficulty measures. With respect to scoring criteria, time control was the most difficult aspect among the scoring criteria. Regarding gender difference in raters, female raters gave lower scores than male raters, but there was no statistical evidence for gender-related bias. However, raters exhibited statistically significant differences in rater severity. The results of this study demonstrates that the MFRM provides a scientific approach to assessment, which can reveal some useful diagnostic information from the original ordinal rating scores on impromptu speech.
{"title":"Examining the Pre-service School Principals' Impromptu Speech Skills with a Many-Facet Rasch Model.","authors":"Mingchuan Hsieh, Akihito Kamata","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>The purpose of this study is to demonstrate an application of the many-facet Rasch model (MFRM) in evaluating the impromptu speech skills of pre-service principals in Taiwan. The findings showed that the topics of speech did not exhibit different difficulty measures. With respect to scoring criteria, time control was the most difficult aspect among the scoring criteria. Regarding gender difference in raters, female raters gave lower scores than male raters, but there was no statistical evidence for gender-related bias. However, raters exhibited statistically significant differences in rater severity. The results of this study demonstrates that the MFRM provides a scientific approach to assessment, which can reveal some useful diagnostic information from the original ordinal rating scores on impromptu speech.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 3","pages":"282-293"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38978106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Multidimensional pairwise comparison (MPC) items have been widely used to assess career interest, value and personality to avoid response bias in educational sectors. In reality, a statement in an MPC item may have different utilities for different groups, which is referred to as differential statement functioning (DSF). Few studies have been investigated DSF assessment. Based on a Rasch model for MPC items, this study adapts three methods to detect DSF for polytomous MPC items: the equal-mean-utility (EMU) method, the all-other-statement (AOS) method and the constant-statement (CS) method. Simulation study was conducted to evaluate the recovery of parameters as well as the performance of the proposed methods. Results showed that when the test contains DSF statement(s), the CS method where one or more DSF-free statements are chosen as an anchor will yield accurate estimates and perform well for DSF assessment. An empirical example of career interest assessment was provided. .
{"title":"Assessing Differential Statement Functioning in Polytomous Multidimensional Pairwise Comparison Items.","authors":"Xue-Lan Qiu","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Multidimensional pairwise comparison (MPC) items have been widely used to assess career interest, value and personality to avoid response bias in educational sectors. In reality, a statement in an MPC item may have different utilities for different groups, which is referred to as differential statement functioning (DSF). Few studies have been investigated DSF assessment. Based on a Rasch model for MPC items, this study adapts three methods to detect DSF for polytomous MPC items: the equal-mean-utility (EMU) method, the all-other-statement (AOS) method and the constant-statement (CS) method. Simulation study was conducted to evaluate the recovery of parameters as well as the performance of the proposed methods. Results showed that when the test contains DSF statement(s), the CS method where one or more DSF-free statements are chosen as an anchor will yield accurate estimates and perform well for DSF assessment. An empirical example of career interest assessment was provided. .</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 3","pages":"329-346"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38978109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Nicholas Marosszeky, E Arthur Shores, Michael P Jones, Rassoul Sadeghi
Streiner, Norman and Cairney (2015) "Health Measurement Scales: A practical guide to their development and use", now in its fifth edition, is one of the foundational texts of the health outcomes movement. It states that "the differences between scales constructed with IRT and CTT are trivial." (Streiner, Norman and Cairney, 2015, p. 299) This statement is representative of the view which emphasizes the equivalence of True-Score Theory (TST) (also known as Classical Test Theory [CTT]) and the Rasch Measurement Model [RMM]). This view is widely held and has been one factor in limiting the application of RMM in the development of health outcome measures. However, this equivalence view relies heavily on a paper by Fan (1998) which examined the item statistics derived from TST, IRT (Item Response Theory) and the RMM for a large educational dataset. While subject to a number of theoretical and practical criticisms from a RMM perspective this paper has not been replicated with a large sample. This paper by replicating and extending the paper by Fan (1998) challenges the finding that item difficulty indexes derived from high and low ability samples using TST techniques are invariant. They are not. On the other hand, item locations derived from the RMM have a high degree of invariance. This secondary data analysis, by working through the methods used by Fan (1998) also demonstrates that a reliance on the magnitude of correlational coefficients cannot be used to determine the invariance of item difficulty indexes. An investigation into the linearity of the correlations using scatter plots is also required. Finally, an item analysis derived from the item difficulty indexes which displays a picture of the test as a whole shows that, for this large sample, the differences between scales constructed with TST and the RMM are not trivial.
斯特雷纳、诺曼和凯恩(2015)《健康测量量表:编制和使用的实用指南》现已出版第五版,是健康成果运动的基础文本之一。它指出“用IRT和CTT构建的量表之间的差异是微不足道的。”(Streiner, Norman and Cairney, 2015, p. 299)这一说法代表了强调真分数理论(TST)(也称为经典测试理论[CTT])和Rasch测量模型[RMM]等价的观点。这一观点被广泛接受,并且是限制在制定保健结果措施中应用RMM的一个因素。然而,这种等效性观点在很大程度上依赖于范(1998)的一篇论文,该论文对大型教育数据集的项目统计数据进行了研究,这些统计数据来自TST、IRT(项目反应理论)和RMM。虽然从RMM的角度来看,这篇论文受到了许多理论和实践的批评,但没有得到大样本的复制。本文复制并扩展了Fan(1998)的论文,挑战了使用TST技术从高能力和低能力样本中得出的项目难度指数不变的发现。事实并非如此。另一方面,从RMM派生的项目位置具有高度的不变性。通过Fan(1998)使用的方法进行的二次数据分析也表明,不能使用对相关系数大小的依赖来确定项目难度指数的不变性。还需要使用散点图对相关性的线性进行调查。最后,从项目难度指数中得出的项目分析显示了测试的整体情况,对于这个大样本,用TST和RMM构建的量表之间的差异并非微不足道。
{"title":"A Psychometric Replication of Fan (1998) Item Response Theory and Classical Test Theory: An Empirical Comparison of their Item/Person Statistics.","authors":"Nicholas Marosszeky, E Arthur Shores, Michael P Jones, Rassoul Sadeghi","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Streiner, Norman and Cairney (2015) \"Health Measurement Scales: A practical guide to their development and use\", now in its fifth edition, is one of the foundational texts of the health outcomes movement. It states that \"the differences between scales constructed with IRT and CTT are trivial.\" (Streiner, Norman and Cairney, 2015, p. 299) This statement is representative of the view which emphasizes the equivalence of True-Score Theory (TST) (also known as Classical Test Theory [CTT]) and the Rasch Measurement Model [RMM]). This view is widely held and has been one factor in limiting the application of RMM in the development of health outcome measures. However, this equivalence view relies heavily on a paper by Fan (1998) which examined the item statistics derived from TST, IRT (Item Response Theory) and the RMM for a large educational dataset. While subject to a number of theoretical and practical criticisms from a RMM perspective this paper has not been replicated with a large sample. This paper by replicating and extending the paper by Fan (1998) challenges the finding that item difficulty indexes derived from high and low ability samples using TST techniques are invariant. They are not. On the other hand, item locations derived from the RMM have a high degree of invariance. This secondary data analysis, by working through the methods used by Fan (1998) also demonstrates that a reliance on the magnitude of correlational coefficients cannot be used to determine the invariance of item difficulty indexes. An investigation into the linearity of the correlations using scatter plots is also required. Finally, an item analysis derived from the item difficulty indexes which displays a picture of the test as a whole shows that, for this large sample, the differences between scales constructed with TST and the RMM are not trivial.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 4","pages":"456-480"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38912689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Many assessment scales in the social sciences are composed of multiple items that form a subscale structure. They have this structure because more than one aspect of the variable is assessed and more than one item assesses each aspect. Nevertheless, generally, a single measurement is required from the scale. A characteristic of this measurement is that the greater the number of items, and categories within an item, that assess an aspect, the greater its influence on the final measurement. One way to control this influence is to include the desired relative number of items and categories to assess each aspect in the scale. However, there are circumstances where designing the required number of items and categories for each aspect is challenging. This paper shows a method of controlling the influence of the number of items and categories assessing each aspect by a-priori weighting of items at the person measurement stage with the Rasch model.
{"title":"A-priori Weighting of Items with the Rasch Model.","authors":"David Andrich, Sonia Sappl","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Many assessment scales in the social sciences are composed of multiple items that form a subscale structure. They have this structure because more than one aspect of the variable is assessed and more than one item assesses each aspect. Nevertheless, generally, a single measurement is required from the scale. A characteristic of this measurement is that the greater the number of items, and categories within an item, that assess an aspect, the greater its influence on the final measurement. One way to control this influence is to include the desired relative number of items and categories to assess each aspect in the scale. However, there are circumstances where designing the required number of items and categories for each aspect is challenging. This paper shows a method of controlling the influence of the number of items and categories assessing each aspect by a-priori weighting of items at the person measurement stage with the Rasch model.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 3","pages":"243-255"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38978134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Georgios D Sideridis, Abdulrahman Al-Samrani, Bjorn Norrbom
The purpose of the present report was to assess congruence between a language-based national examination (termed English placement test - EPT) and the Common European Framework of Reference for Languages (CEFR) levels. To this end, a series of methodological steps were put forth to accumulate evidence suggesting that language performance based on the EPT instrument can be split onto meaningful subgroups based on theoretical (expert judgement on difficulty level and CEFR correspondence) and empirical considerations (i.e., how well these levels and subgroups emerged). Participants were 2642 high school graduates who took on the EPT instrument as part of their entry criteria to the university and for the purposes of the present study only the structure subscale is presented. Items were classified as reflecting specific CEFR levels and a person-based analysis attempted to classify individuals sharing the same behavioral patterns. Results using a latent class analysis (LCA) indicated that a Pre-A1, an A1 an A2 a B1 and a B2 levels were present with regard to the structure domain of language. Results showed a strong alignment between the EPT structure domain and CEFR guidelines using various methodological approaches.
{"title":"Alignment of a Language Instrument Scores to CEFR Levels: Methodological and Empirical Considerations.","authors":"Georgios D Sideridis, Abdulrahman Al-Samrani, Bjorn Norrbom","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>The purpose of the present report was to assess congruence between a language-based national examination (termed English placement test - EPT) and the Common European Framework of Reference for Languages (CEFR) levels. To this end, a series of methodological steps were put forth to accumulate evidence suggesting that language performance based on the EPT instrument can be split onto meaningful subgroups based on theoretical (expert judgement on difficulty level and CEFR correspondence) and empirical considerations (i.e., how well these levels and subgroups emerged). Participants were 2642 high school graduates who took on the EPT instrument as part of their entry criteria to the university and for the purposes of the present study only the structure subscale is presented. Items were classified as reflecting specific CEFR levels and a person-based analysis attempted to classify individuals sharing the same behavioral patterns. Results using a latent class analysis (LCA) indicated that a Pre-A1, an A1 an A2 a B1 and a B2 levels were present with regard to the structure domain of language. Results showed a strong alignment between the EPT structure domain and CEFR guidelines using various methodological approaches.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 1","pages":"68-90"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"37704117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Stephen M Ponisciak, Rob Meyer, Anna Brown, Tracy Schatzberg
A consensus has developed that high-quality teacher evaluation systems require multiple measures. We examine multiple measures from a large urban school district, which has included observational ratings and value-added ratings in its system since 2010. Evaluation systems that do not account for observer severity, classroom context, and other factors may yield different results from systems that do account for these factors. Choosing a simpler system involves a trade-off regarding a system's robustness or defensibility. Using a many-faceted Rasch model, we explore rating components like observer, time of year, and subdomain. We find high reliability of the resulting teacher ratings, some impact of adjusting for observer differences and differences between subdomains, and positive correlation with value-added measures. A comprehensive analysis like MFRM should be part of a district's evaluation system, even if only as a robustness check, and districts should examine how observational scores and classroom context are related.
{"title":"Trade-Offs in the Implementation of Observational Ratings Systems.","authors":"Stephen M Ponisciak, Rob Meyer, Anna Brown, Tracy Schatzberg","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>A consensus has developed that high-quality teacher evaluation systems require multiple measures. We examine multiple measures from a large urban school district, which has included observational ratings and value-added ratings in its system since 2010. Evaluation systems that do not account for observer severity, classroom context, and other factors may yield different results from systems that do account for these factors. Choosing a simpler system involves a trade-off regarding a system's robustness or defensibility. Using a many-faceted Rasch model, we explore rating components like observer, time of year, and subdomain. We find high reliability of the resulting teacher ratings, some impact of adjusting for observer differences and differences between subdomains, and positive correlation with value-added measures. A comprehensive analysis like MFRM should be part of a district's evaluation system, even if only as a robustness check, and districts should examine how observational scores and classroom context are related.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 1","pages":"50-67"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"37704116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Accurate parameter estimation in the Rasch model involves the assumption of conditional independence, also termed local independence. Conditional on ability, the responses to items A and B should be independent. Two types of conditional dependence are detailed in this pedagogical piece: trait dependency and response dependency. The bias in difficulty and reliability and the estimates of fit and correlated residuals resulting from these dependencies are compared and contrasted to results from using models that account for the dependency. Contrasts with results from a 2-parameter item response theory model are also briefly noted.
{"title":"Comparing Causes of Dependency: Shared Latent Trait or Dependence on Observed Response.","authors":"Christine E DeMars","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Accurate parameter estimation in the Rasch model involves the assumption of conditional independence, also termed local independence. Conditional on ability, the responses to items A and B should be independent. Two types of conditional dependence are detailed in this pedagogical piece: trait dependency and response dependency. The bias in difficulty and reliability and the estimates of fit and correlated residuals resulting from these dependencies are compared and contrasted to results from using models that account for the dependency. Contrasts with results from a 2-parameter item response theory model are also briefly noted.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 4","pages":"400-419"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38912686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Katherine Wentzel, Judith A Vessey, Lori Laffel, Larry Ludlow
The emotional burden of living with Type 1 diabetes (T1D) is experienced differently in each life stage. Thus the measurement of diabetes distress (DD) warrants tailoring to particular developmental stages, specifically emerging adulthood (ages 18-30). The new measure entitled the Problem Areas in Diabetes- Emerging Adult version (PAID-EA) is intended to be a developmentally-embedded measure of DD for use in clinical and research settings. The goal of the present study was to use Rasch psychometric analysis to reduce and refine the PAID-EA. Emerging adults with T1D (n = 194) completed the 30-item online survey. Evaluation of response category functioning, measurement precision, redundancy, unidimensionality and targeting guided item reduction through iterative revisions. The reduced and refined PAID-EA consists of 25 items and shows promising utility for clinicians and researchers.
{"title":"Diabetes Distress in Emerging Adults: Refining the Problem Areas in Diabetes-Emerging Adult Version using Rasch Analysis.","authors":"Katherine Wentzel, Judith A Vessey, Lori Laffel, Larry Ludlow","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>The emotional burden of living with Type 1 diabetes (T1D) is experienced differently in each life stage. Thus the measurement of diabetes distress (DD) warrants tailoring to particular developmental stages, specifically emerging adulthood (ages 18-30). The new measure entitled the Problem Areas in Diabetes- Emerging Adult version (PAID-EA) is intended to be a developmentally-embedded measure of DD for use in clinical and research settings. The goal of the present study was to use Rasch psychometric analysis to reduce and refine the PAID-EA. Emerging adults with T1D (n = 194) completed the 30-item online survey. Evaluation of response category functioning, measurement precision, redundancy, unidimensionality and targeting guided item reduction through iterative revisions. The reduced and refined PAID-EA consists of 25 items and shows promising utility for clinicians and researchers.</p>","PeriodicalId":73608,"journal":{"name":"Journal of applied measurement","volume":"21 4","pages":"481-495"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38912690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}