This paper introduces two new Item Response Theory (IRT) models, based on the Generalized Extreme Value (GEV) distribution. These new models have asymmetric item characteristic curves (ICC) which have drawn growing interest, as they may better model actual item response behaviours in specific scenarios. The analysis of the models is carried out using a Bayesian approach, and their properties are examined and discussed. The validity of the models is verified by means of extensive simulation studies to evaluate the sensitivity of the model to the choice of priors on the new item parameter introduced, the accuracy of the parameters' recovery, as well as an assessment of the capacity of model comparison criteria to choose the best model against other IRT models. The new models are exemplified using real data from two mathematics tests, one applied in Peruvian public schools and another one administered to incoming university students in Chile. In both cases, the proposed models showed to be a promising alternative to asymmetric IRT models, offering new insights into item response modelling.
{"title":"Generalized extreme value IRT models.","authors":"Jessica Alves, Jorge Bazán, Jorge González","doi":"10.1111/bmsp.70015","DOIUrl":"https://doi.org/10.1111/bmsp.70015","url":null,"abstract":"<p><p>This paper introduces two new Item Response Theory (IRT) models, based on the Generalized Extreme Value (GEV) distribution. These new models have asymmetric item characteristic curves (ICC) which have drawn growing interest, as they may better model actual item response behaviours in specific scenarios. The analysis of the models is carried out using a Bayesian approach, and their properties are examined and discussed. The validity of the models is verified by means of extensive simulation studies to evaluate the sensitivity of the model to the choice of priors on the new item parameter introduced, the accuracy of the parameters' recovery, as well as an assessment of the capacity of model comparison criteria to choose the best model against other IRT models. The new models are exemplified using real data from two mathematics tests, one applied in Peruvian public schools and another one administered to incoming university students in Chile. In both cases, the proposed models showed to be a promising alternative to asymmetric IRT models, offering new insights into item response modelling.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145497315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Debora de Chiusole, Andrea Spoto, Umberto Granziol, Luca Stefanutti
In knowledge structure theory (KST) framework, this study evaluates the reliability of knowledge state estimation by introducing two key measures: the expected accuracy rate and the expected discrepancy. The accuracy rate quantifies the likelihood that the estimated knowledge state aligns with the true state, while the expected discrepancy measures the average deviation when misclassification occurs. To support the theoretical framework, we provide an in-depth discussion of these indices, supplemented by two simulation studies and an empirical example. The simulation results reveal a trade-off between the number of items and the size of the knowledge structure. Specifically, smaller structures exhibit consistent accuracy across different error levels, while larger structures show increasing discrepancies as error rates rise. Nevertheless, accuracy improves with a greater number of items in larger structures, mitigating the impact of errors. Additionally, the expected discrepancy analysis shows that when misclassification occurs, the estimated state is generally close to the true one, minimizing the effect of errors in the assessment. Finally, an empirical application using real assessment data demonstrates the practical relevance of the proposed measures. This suggests that KST-based assessments provide reliable and meaningful diagnostic information, highlighting their potential for use in educational and psychological testing.
{"title":"Reliability measures in knowledge structure theory.","authors":"Debora de Chiusole, Andrea Spoto, Umberto Granziol, Luca Stefanutti","doi":"10.1111/bmsp.70013","DOIUrl":"https://doi.org/10.1111/bmsp.70013","url":null,"abstract":"<p><p>In knowledge structure theory (KST) framework, this study evaluates the reliability of knowledge state estimation by introducing two key measures: the expected accuracy rate and the expected discrepancy. The accuracy rate quantifies the likelihood that the estimated knowledge state aligns with the true state, while the expected discrepancy measures the average deviation when misclassification occurs. To support the theoretical framework, we provide an in-depth discussion of these indices, supplemented by two simulation studies and an empirical example. The simulation results reveal a trade-off between the number of items and the size of the knowledge structure. Specifically, smaller structures exhibit consistent accuracy across different error levels, while larger structures show increasing discrepancies as error rates rise. Nevertheless, accuracy improves with a greater number of items in larger structures, mitigating the impact of errors. Additionally, the expected discrepancy analysis shows that when misclassification occurs, the estimated state is generally close to the true one, minimizing the effect of errors in the assessment. Finally, an empirical application using real assessment data demonstrates the practical relevance of the proposed measures. This suggests that KST-based assessments provide reliable and meaningful diagnostic information, highlighting their potential for use in educational and psychological testing.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145423528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yong Zhang, Anja F Ernst, Ginette Lafit, Ward B Eiling, Laura F Bringmann
The stationary autoregressive model forms an important base of time-series analysis in today's psychology research. Diverse nonstationary extensions of this model are developed to capture different types of changing temporal dynamics. However, researchers do not always have a solid theoretical base to rely on for deciding a-priori which of these nonstationary models is the most appropriate for a given time-series. In this case, correct model selection becomes a crucial step to ensure an accurate understanding of the temporal dynamics. This study consists of two main parts. First, with a simulation study, we investigated the performance of in-sample (information criteria) and out-of-sample (cross-validation, out-of-sample prediction) model selection techniques in identifying six different univariate nonstationary processes. We found that the Bayesian information criteria (BIC) has an overall optimal performance whereas other techniques' performance depends largely on the time-series' length. Then, we re-analysed a 239-day-long time-series of positive and negative affect to illustrate the model selection process. Combining the simulation results and practical considerations from the empirical analysis, we argue that model selection for nonstationary time-series should not completely rely on data-driven approaches. Instead, more theory-driven approaches where researchers actively integrate their qualitative understanding will inform the data-driven approaches.
{"title":"An investigation into in-sample and out-of-sample model selection for nonstationary autoregressive models.","authors":"Yong Zhang, Anja F Ernst, Ginette Lafit, Ward B Eiling, Laura F Bringmann","doi":"10.1111/bmsp.70012","DOIUrl":"https://doi.org/10.1111/bmsp.70012","url":null,"abstract":"<p><p>The stationary autoregressive model forms an important base of time-series analysis in today's psychology research. Diverse nonstationary extensions of this model are developed to capture different types of changing temporal dynamics. However, researchers do not always have a solid theoretical base to rely on for deciding a-priori which of these nonstationary models is the most appropriate for a given time-series. In this case, correct model selection becomes a crucial step to ensure an accurate understanding of the temporal dynamics. This study consists of two main parts. First, with a simulation study, we investigated the performance of in-sample (information criteria) and out-of-sample (cross-validation, out-of-sample prediction) model selection techniques in identifying six different univariate nonstationary processes. We found that the Bayesian information criteria (BIC) has an overall optimal performance whereas other techniques' performance depends largely on the time-series' length. Then, we re-analysed a 239-day-long time-series of positive and negative affect to illustrate the model selection process. Combining the simulation results and practical considerations from the empirical analysis, we argue that model selection for nonstationary time-series should not completely rely on data-driven approaches. Instead, more theory-driven approaches where researchers actively integrate their qualitative understanding will inform the data-driven approaches.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145395303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Reinforcement learning (RL) powers the engine of adaptive learning systems which recommend customized learning materials to individual learners in their varying learning states to optimize learning effectiveness. However, some argue that only improving learning effectiveness may be insufficient, particularly if it overly extends learning efforts and requires additional time to work on the recommended materials. Learners with different amounts of prior knowledge consume different amounts of time on the same material. Therefore, designers should consider both the usefulness of the material and the time dedicated to making sense of the materials by individual learners with a specific amount of prior knowledge. To fill this gap, this study proposes a RL-based adaptive learning system wherein reward is improved by considering both factors. We then conducted Monte Carlo simulation studies to verify the effects of the improved reward and uncover the mechanisms for RL recommendation strategies. Results show that the improved reward reduces learners' learning duration substantially due to interpretable recommendation strategies, which results in growing learning efficiency for learners with varying prior knowledge.
{"title":"Reinforcement learning-based adaptive learning: Rewards improvement considering learning duration.","authors":"Tongxin Zhang, Canxi Cao, Tao Xin, Xiaoming Zhai","doi":"10.1111/bmsp.70014","DOIUrl":"https://doi.org/10.1111/bmsp.70014","url":null,"abstract":"<p><p>Reinforcement learning (RL) powers the engine of adaptive learning systems which recommend customized learning materials to individual learners in their varying learning states to optimize learning effectiveness. However, some argue that only improving learning effectiveness may be insufficient, particularly if it overly extends learning efforts and requires additional time to work on the recommended materials. Learners with different amounts of prior knowledge consume different amounts of time on the same material. Therefore, designers should consider both the usefulness of the material and the time dedicated to making sense of the materials by individual learners with a specific amount of prior knowledge. To fill this gap, this study proposes a RL-based adaptive learning system wherein reward is improved by considering both factors. We then conducted Monte Carlo simulation studies to verify the effects of the improved reward and uncover the mechanisms for RL recommendation strategies. Results show that the improved reward reduces learners' learning duration substantially due to interpretable recommendation strategies, which results in growing learning efficiency for learners with varying prior knowledge.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145356833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pasquale Anselmi, Debora de Chiusole, Egidio Robusto, Alice Bacherini, Giulia Balboni, Andrea Brancaccio, Ottavia M Epifania, Noemi Mazzoni, Luca Stefanutti
The basic local independence model (BLIM) is appropriate in situations where populations do not differ in the probabilities of the knowledge states and the probabilities of careless errors and lucky guesses of the items. In some situations, this is not the case. This work introduces the multiple observed classification local independence model (MOCLIM), which extends the BLIM by allowing the above probabilities to vary across populations. In the MOCLIM, each individual is characterized by proficiency, careless and guessing classes, which are observed and determine the probabilities of knowledge states, careless errors and lucky guesses of a population. Given a particular class type (proficiency, careless, or guessing), the probabilities are the same for populations with the same class but may vary between populations with different classes. Algorithms for maximum likelihood estimation of the MOCLIM parameters are provided. The results of a simulation study suggest that the true parameter values are well recovered by the estimation algorithm and that the true model can be uncovered by comparing the goodness-of-fit of alternative models. The results of an empirical application to data from Raven-like matrices suggest that the MOCLIM effectively discriminates between situations where group differences are expected and those where they are not.
{"title":"An extension of the basic local independence model to multiple observed classifications.","authors":"Pasquale Anselmi, Debora de Chiusole, Egidio Robusto, Alice Bacherini, Giulia Balboni, Andrea Brancaccio, Ottavia M Epifania, Noemi Mazzoni, Luca Stefanutti","doi":"10.1111/bmsp.70008","DOIUrl":"https://doi.org/10.1111/bmsp.70008","url":null,"abstract":"<p><p>The basic local independence model (BLIM) is appropriate in situations where populations do not differ in the probabilities of the knowledge states and the probabilities of careless errors and lucky guesses of the items. In some situations, this is not the case. This work introduces the multiple observed classification local independence model (MOCLIM), which extends the BLIM by allowing the above probabilities to vary across populations. In the MOCLIM, each individual is characterized by proficiency, careless and guessing classes, which are observed and determine the probabilities of knowledge states, careless errors and lucky guesses of a population. Given a particular class type (proficiency, careless, or guessing), the probabilities are the same for populations with the same class but may vary between populations with different classes. Algorithms for maximum likelihood estimation of the MOCLIM parameters are provided. The results of a simulation study suggest that the true parameter values are well recovered by the estimation algorithm and that the true model can be uncovered by comparing the goodness-of-fit of alternative models. The results of an empirical application to data from Raven-like matrices suggest that the MOCLIM effectively discriminates between situations where group differences are expected and those where they are not.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145115131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The Bayes factor, the data-based updating factor of the prior to posterior odds of two hypotheses, is a natural measure of statistical evidence for one hypothesis over the other. We show how Bayes factors can also be used for parameter estimation. The key idea is to consider the Bayes factor as a function of the parameter value under the null hypothesis. This 'support curve' is inverted to obtain point estimates ('maximum evidence estimates') and interval estimates ('support intervals'), similar to how p-value functions are inverted to obtain point estimates and confidence intervals. This provides data analysts with a unified inference framework as Bayes factors (for any tested parameter value), support intervals (at any level), and point estimates can be easily read off from a plot of the support curve. This approach shares similarities but is also distinct from conventional Bayesian and frequentist approaches: It uses the Bayesian evidence calculus, but without synthesizing data and prior, and it defines statistical evidence in terms of (integrated) likelihood ratios, but also includes a natural way for dealing with nuisance parameters. Applications to meta-analysis, replication studies and logistic regression illustrate how our framework is of practical value for making quantitative inferences.
{"title":"A Bayes factor framework for unified parameter estimation and hypothesis testing.","authors":"Samuel Pawel","doi":"10.1111/bmsp.70011","DOIUrl":"https://doi.org/10.1111/bmsp.70011","url":null,"abstract":"<p><p>The Bayes factor, the data-based updating factor of the prior to posterior odds of two hypotheses, is a natural measure of statistical evidence for one hypothesis over the other. We show how Bayes factors can also be used for parameter estimation. The key idea is to consider the Bayes factor as a function of the parameter value under the null hypothesis. This 'support curve' is inverted to obtain point estimates ('maximum evidence estimates') and interval estimates ('support intervals'), similar to how p-value functions are inverted to obtain point estimates and confidence intervals. This provides data analysts with a unified inference framework as Bayes factors (for any tested parameter value), support intervals (at any level), and point estimates can be easily read off from a plot of the support curve. This approach shares similarities but is also distinct from conventional Bayesian and frequentist approaches: It uses the Bayesian evidence calculus, but without synthesizing data and prior, and it defines statistical evidence in terms of (integrated) likelihood ratios, but also includes a natural way for dealing with nuisance parameters. Applications to meta-analysis, replication studies and logistic regression illustrate how our framework is of practical value for making quantitative inferences.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145082536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Intensive longitudinal data are often found to be non-stationary, namely, showing changes in statistical properties, such as means and variance-covariance structures, over time. One way to accommodate non-stationarity is to specify key parameters that show over-time changes as time-varying parameters (TVPs). However, the nature and dynamics of TVPs may themselves be heterogeneous across time, contexts, developmental stages, individuals and as related to other biopsychosocial-cultural influences. We propose an outlier detection method designed to facilitate the detection of critical shifts in any differentiable linear and non-linear dynamic functions, including dynamic functions for TVPs. This approach can be readily applied to various data scenarios, including single-subject and multisubject, univariate and multivariate processes, as well as with and without latent variables. We demonstrate the utility and performance of this approach with three sets of simulation studies and an empirical illustration using facial electromyography data from a laboratory emotion induction study.
{"title":"Detecting Critical Change in Dynamics Through Outlier Detection with Time-Varying Parameters.","authors":"Meng Chen, Michael D Hunter, Sy-Miin Chow","doi":"10.1111/bmsp.70010","DOIUrl":"https://doi.org/10.1111/bmsp.70010","url":null,"abstract":"<p><p>Intensive longitudinal data are often found to be non-stationary, namely, showing changes in statistical properties, such as means and variance-covariance structures, over time. One way to accommodate non-stationarity is to specify key parameters that show over-time changes as time-varying parameters (TVPs). However, the nature and dynamics of TVPs may themselves be heterogeneous across time, contexts, developmental stages, individuals and as related to other biopsychosocial-cultural influences. We propose an outlier detection method designed to facilitate the detection of critical shifts in any differentiable linear and non-linear dynamic functions, including dynamic functions for TVPs. This approach can be readily applied to various data scenarios, including single-subject and multisubject, univariate and multivariate processes, as well as with and without latent variables. We demonstrate the utility and performance of this approach with three sets of simulation studies and an empirical illustration using facial electromyography data from a laboratory emotion induction study.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145066348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Psychological research has traditionally relied on linear models to test scientific hypotheses. However, the emergence of machine learning (ML) algorithms has opened new opportunities for exploring variable relationships beyond linear constraints. To interpret the outcomes of these 'black-box' algorithms, various tools for assessing feature importance have been developed. However, most of these tools are descriptive and do not facilitate statistical inference. To address this gap, our study introduces two versions of residual permutation tests (RPTs), designed to assess the significance of a target feature in predicting the label. The first variant, RPT on Y (RPT-Y), permutes the residuals of the label conditioned on features other than the target. The second variant, RPT on X (RPT-X), permutes the residuals of the target feature conditioned on the other features. Through a comprehensive simulation study, we show that RPT-X maintains empirical Type I error rates under the nominal level across a wide range of ML algorithms and demonstrates appropriate statistical power in both regression and classification contexts. These findings suggest the utility of RPT-X for hypothesis testing in ML applications.
心理学研究传统上依靠线性模型来检验科学假设。然而,机器学习(ML)算法的出现为探索超越线性约束的变量关系开辟了新的机会。为了解释这些“黑盒”算法的结果,已经开发了各种评估特征重要性的工具。然而,这些工具大多是描述性的,不便于统计推断。为了解决这一差距,我们的研究引入了两个版本的残差排列测试(RPTs),旨在评估目标特征在预测标签中的重要性。第一种变体,RPT on Y (RPT-Y),根据目标以外的特征来排列标签的残差。第二个变体,RPT on X (RPT-X),将目标特征的残差以其他特征为条件进行排列。通过全面的模拟研究,我们表明RPT-X在广泛的ML算法中保持经验I型错误率低于标称水平,并在回归和分类上下文中显示出适当的统计能力。这些发现表明RPT-X在机器学习应用中的假设检验的效用。
{"title":"Residual permutation tests for feature importance in machine learning.","authors":"Po-Hsien Huang","doi":"10.1111/bmsp.70009","DOIUrl":"https://doi.org/10.1111/bmsp.70009","url":null,"abstract":"<p><p>Psychological research has traditionally relied on linear models to test scientific hypotheses. However, the emergence of machine learning (ML) algorithms has opened new opportunities for exploring variable relationships beyond linear constraints. To interpret the outcomes of these 'black-box' algorithms, various tools for assessing feature importance have been developed. However, most of these tools are descriptive and do not facilitate statistical inference. To address this gap, our study introduces two versions of residual permutation tests (RPTs), designed to assess the significance of a target feature in predicting the label. The first variant, RPT on Y (RPT-Y), permutes the residuals of the label conditioned on features other than the target. The second variant, RPT on X (RPT-X), permutes the residuals of the target feature conditioned on the other features. Through a comprehensive simulation study, we show that RPT-X maintains empirical Type I error rates under the nominal level across a wide range of ML algorithms and demonstrates appropriate statistical power in both regression and classification contexts. These findings suggest the utility of RPT-X for hypothesis testing in ML applications.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":" ","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144979556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cornelis J. Potgieter, Akihito Kamata, Yusuf Kara, Xin Qiao
In this study, we explore parameter estimation for a joint count-time data model with a two-factor latent trait structure, representing accuracy and speed. Each count-time variable pair corresponds to a specific item on a measurement instrument, where each item consists of a fixed number of tasks. The count variable represents the number of successfully completed tasks and is modeled using a Beta-binomial distribution to account for potential over-dispersion. The time variable, representing the duration needed to complete the tasks, is modeled using a normal distribution on a logarithmic scale. To characterize the model structure, we derive marginal moments that inform a set of method-of-moments (MOM) estimators, which serve as initial values for maximum likelihood estimation (MLE) via the Monte Carlo Expectation-Maximization (MCEM) algorithm. Standard errors are estimated using both the observed information matrix and bootstrap resampling, with simulation results indicating superior performance of the bootstrap, especially near boundary values of the dispersion parameter. A comprehensive simulation study investigates estimator accuracy and computational efficiency. To demonstrate the methodology, we analyze oral reading fluency (ORF) data, showing substantial variation in item-level dispersion and providing evidence for the improved model fit of the Beta-binomial specification, assessed using standardized root mean square residuals (SRMSR).
{"title":"Joint analysis of dispersed count-time data using a bivariate latent factor model","authors":"Cornelis J. Potgieter, Akihito Kamata, Yusuf Kara, Xin Qiao","doi":"10.1111/bmsp.70005","DOIUrl":"10.1111/bmsp.70005","url":null,"abstract":"<p>In this study, we explore parameter estimation for a joint count-time data model with a two-factor latent trait structure, representing accuracy and speed. Each count-time variable pair corresponds to a specific item on a measurement instrument, where each item consists of a fixed number of tasks. The count variable represents the number of successfully completed tasks and is modeled using a Beta-binomial distribution to account for potential over-dispersion. The time variable, representing the duration needed to complete the tasks, is modeled using a normal distribution on a logarithmic scale. To characterize the model structure, we derive marginal moments that inform a set of method-of-moments (MOM) estimators, which serve as initial values for maximum likelihood estimation (MLE) via the Monte Carlo Expectation-Maximization (MCEM) algorithm. Standard errors are estimated using both the observed information matrix and bootstrap resampling, with simulation results indicating superior performance of the bootstrap, especially near boundary values of the dispersion parameter. A comprehensive simulation study investigates estimator accuracy and computational efficiency. To demonstrate the methodology, we analyze oral reading fluency (ORF) data, showing substantial variation in item-level dispersion and providing evidence for the improved model fit of the Beta-binomial specification, assessed using standardized root mean square residuals (SRMSR).</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"79 1","pages":"207-228"},"PeriodicalIF":1.8,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://bpspsychub.onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.70005","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144979632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper, we propose the generalized mixed reduced rank regression method, GMR3 for short. GMR3 is a regression method for a mix of numeric, binary and ordinal response variables. The predictor variables can be a mix of binary, nominal, ordinal and numeric variables. For dealing with the categorical predictors we use optimal scaling. A majorization-minimization algorithm is derived for maximum likelihood estimation. A series of simulation studies is shown (Section 4) to evaluate the performance of the algorithm with different types of predictor and response variables. In Section 5, we briefly discuss the choices to make when applying the model the empirical data and give suggestions for supporting such choices. In a second simulation study (Section 6), we further study the behaviour of the model and algorithm in different scenarios for the true rank in relation to sample size. In Section 7, we show an application of GMR3 using the Eurobarometer Surveys data set of 2023.
{"title":"Reduced rank regression for mixed predictor and response variables","authors":"Mark de Rooij, Lorenza Cotugno, Roberta Siciliano","doi":"10.1111/bmsp.70004","DOIUrl":"10.1111/bmsp.70004","url":null,"abstract":"<p>In this paper, we propose the generalized mixed reduced rank regression method, GMR<sup>3</sup> for short. GMR<sup>3</sup> is a regression method for a mix of numeric, binary and ordinal response variables. The predictor variables can be a mix of binary, nominal, ordinal and numeric variables. For dealing with the categorical predictors we use optimal scaling. A majorization-minimization algorithm is derived for maximum likelihood estimation. A series of simulation studies is shown (Section 4) to evaluate the performance of the algorithm with different types of predictor and response variables. In Section 5, we briefly discuss the choices to make when applying the model the empirical data and give suggestions for supporting such choices. In a second simulation study (Section 6), we further study the behaviour of the model and algorithm in different scenarios for the true rank in relation to sample size. In Section 7, we show an application of GMR<sup>3</sup> using the Eurobarometer Surveys data set of 2023.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"79 1","pages":"173-206"},"PeriodicalIF":1.8,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://bpspsychub.onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.70004","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144979605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}