Pub Date : 2020-01-01DOI: 10.1080/24709360.2019.1618653
Steven S. Henley, R. Golden, T. Kashner
ABSTRACT Statistical modeling methods are widely used in clinical science, epidemiology, and health services research to analyze data that has been collected in clinical trials as well as observational studies of existing data sources, such as claims files and electronic health records. Diagnostic and prognostic inferences from statistical models are critical to researchers advancing science, clinical practitioners making patient care decisions, and administrators and policy makers impacting the health care system to improve quality and reduce costs. The veracity of such inferences relies not only on the quality and completeness of the collected data, but also statistical model validity. A key component of establishing model validity is determining when a model is not correctly specified and therefore incapable of adequately representing the Data Generating Process (DGP). In this article, model validity is first described and methods designed for assessing model fit, specification, and selection are reviewed. Second, data transformations that improve the model’s ability to represent the DGP are addressed. Third, model search and validation methods are discussed. Finally, methods for evaluating predictive and classification performance are presented. Together, these methods provide a practical framework with recommendations to guide the development and evaluation of statistical models that provide valid statistical inferences.
{"title":"Statistical modeling methods: challenges and strategies","authors":"Steven S. Henley, R. Golden, T. Kashner","doi":"10.1080/24709360.2019.1618653","DOIUrl":"https://doi.org/10.1080/24709360.2019.1618653","url":null,"abstract":"ABSTRACT Statistical modeling methods are widely used in clinical science, epidemiology, and health services research to analyze data that has been collected in clinical trials as well as observational studies of existing data sources, such as claims files and electronic health records. Diagnostic and prognostic inferences from statistical models are critical to researchers advancing science, clinical practitioners making patient care decisions, and administrators and policy makers impacting the health care system to improve quality and reduce costs. The veracity of such inferences relies not only on the quality and completeness of the collected data, but also statistical model validity. A key component of establishing model validity is determining when a model is not correctly specified and therefore incapable of adequately representing the Data Generating Process (DGP). In this article, model validity is first described and methods designed for assessing model fit, specification, and selection are reviewed. Second, data transformations that improve the model’s ability to represent the DGP are addressed. Third, model search and validation methods are discussed. Finally, methods for evaluating predictive and classification performance are presented. Together, these methods provide a practical framework with recommendations to guide the development and evaluation of statistical models that provide valid statistical inferences.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"4 1","pages":"105 - 139"},"PeriodicalIF":0.0,"publicationDate":"2020-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1618653","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47377251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-10-15DOI: 10.1080/24709360.2019.1673623
Zheyu Wang
Latent variable modeling is often used in diagnostic studies where a gold standard reference test is not available. Its applications have become increasing popular with the fast discovery of novel biomarkers and the effort to improve healthcare for each individual. This paper attempt to provide a review on current developments and debates of these models with a focus in diagnostic studies and to discuss the value as well as cautionary considerations in the applications of these models.
{"title":"Developments and debates on latent variable modeling in diagnostic studies when there is no gold standard","authors":"Zheyu Wang","doi":"10.1080/24709360.2019.1673623","DOIUrl":"https://doi.org/10.1080/24709360.2019.1673623","url":null,"abstract":"Latent variable modeling is often used in diagnostic studies where a gold standard reference test is not available. Its applications have become increasing popular with the fast discovery of novel biomarkers and the effort to improve healthcare for each individual. This paper attempt to provide a review on current developments and debates of these models with a focus in diagnostic studies and to discuss the value as well as cautionary considerations in the applications of these models.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"5 1","pages":"100 - 117"},"PeriodicalIF":0.0,"publicationDate":"2019-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1673623","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45189867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1615770
A. Albatineh, M. Wilcox, B. Zogheib, M. Niewiadomska-Bugaj
Finding the number of clusters in a data set is considered as one of the fundamental problems in cluster analysis. This paper integrates maximum clustering similarity (MCS), for finding the optimal number of clusters, into R statistical software through the package MCSim. The similarity between the two clustering methods is calculated at the same number of clusters, using Rand [Objective criteria for the evaluation of clustering methods. J Am Stat Assoc. 1971;66:846–850.] and Jaccard [The distribution of the flora of the alpine zone. New Phytologist. 1912;11:37–50.] indices, corrected for chance agreement. The number of clusters at which the index attains its maximum with most frequency is a candidate for the optimal number of clusters. Unlike other criteria, MCS can be used with circular data. Seven clustering algorithms, existing in R, are implemented in MCSim. A graph of the number of clusters vs. clusters similarity using corrected similarity indices is produced. Values of the similarity indices and a clustering tree (dendrogram) are produced. Several examples including simulated, real, and circular data sets are presented to show how MCSim successfully works in practice.
找出数据集中的聚类数量被认为是聚类分析的基本问题之一。本文通过MCSim软件包将最大聚类相似性(MCS)集成到R统计软件中,以寻找最优聚类数。两种聚类方法之间的相似性是在相同数量的聚类下计算的,使用Rand[聚类方法评估的客观标准。J Am Stat Assoc.1971;66:846–850.]和Jaccard[高山区植物群的分布。新植物学家。1912;11:37–50.]指数,对偶然一致性进行校正。指数以最高频率达到最大值的聚类数量是最优聚类数量的候选者。与其他标准不同,MCS可用于循环数据。在MCSim中实现了R中存在的七种聚类算法。使用校正的相似性指数生成聚类数量与聚类相似性的关系图。生成相似性指数的值和聚类树(树状图)。给出了几个例子,包括模拟、真实和循环数据集,以展示MCSim是如何在实践中成功工作的。
{"title":"How many clusters exist? Answer via maximum clustering similarity implemented in R","authors":"A. Albatineh, M. Wilcox, B. Zogheib, M. Niewiadomska-Bugaj","doi":"10.1080/24709360.2019.1615770","DOIUrl":"https://doi.org/10.1080/24709360.2019.1615770","url":null,"abstract":"Finding the number of clusters in a data set is considered as one of the fundamental problems in cluster analysis. This paper integrates maximum clustering similarity (MCS), for finding the optimal number of clusters, into R statistical software through the package MCSim. The similarity between the two clustering methods is calculated at the same number of clusters, using Rand [Objective criteria for the evaluation of clustering methods. J Am Stat Assoc. 1971;66:846–850.] and Jaccard [The distribution of the flora of the alpine zone. New Phytologist. 1912;11:37–50.] indices, corrected for chance agreement. The number of clusters at which the index attains its maximum with most frequency is a candidate for the optimal number of clusters. Unlike other criteria, MCS can be used with circular data. Seven clustering algorithms, existing in R, are implemented in MCSim. A graph of the number of clusters vs. clusters similarity using corrected similarity indices is produced. Values of the similarity indices and a clustering tree (dendrogram) are produced. Several examples including simulated, real, and circular data sets are presented to show how MCSim successfully works in practice.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"62 - 79"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1615770","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42954294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1699341
Nathalie C. Moon, Leilei Zeng, R. Cook
Cohort studies are routinely conducted to learn about the incidence or progression rates of chronic diseases. The illness-death model offers a natural framework for joint consideration of non-fatal events in the semi-competing risks setting. We consider the design of prospective cohort studies where the goal is to estimate the effect of a marker on the risk of a non-fatal event which is subject to interval-censoring due to an intermittent observation scheme. The sample size is shown to depend on the effect of interest, the number of assessments, and the duration of follow-up. Minimum-cost designs are also developed to account for the different costs of recruitment and follow-up examination. We also consider the setting where the event status of individuals is observed subject to misclassification; the consequent need to increase the sample size to account for this error is illustrated through asymptotic calculations.
{"title":"Cohort study design for illness-death processes with disease status under intermittent observation","authors":"Nathalie C. Moon, Leilei Zeng, R. Cook","doi":"10.1080/24709360.2019.1699341","DOIUrl":"https://doi.org/10.1080/24709360.2019.1699341","url":null,"abstract":"Cohort studies are routinely conducted to learn about the incidence or progression rates of chronic diseases. The illness-death model offers a natural framework for joint consideration of non-fatal events in the semi-competing risks setting. We consider the design of prospective cohort studies where the goal is to estimate the effect of a marker on the risk of a non-fatal event which is subject to interval-censoring due to an intermittent observation scheme. The sample size is shown to depend on the effect of interest, the number of assessments, and the duration of follow-up. Minimum-cost designs are also developed to account for the different costs of recruitment and follow-up examination. We also consider the setting where the event status of individuals is observed subject to misclassification; the consequent need to increase the sample size to account for this error is illustrated through asymptotic calculations.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"178 - 200"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1699341","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47561742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1591072
Zhengyang Fang, J. Y. Han, N. Simon, Xiaoping Zhou
Sparse and functional principal component analysis is a technique to extract sparse and smooth principal components from a matrix. In this paper, we propose a modified sparse and functional principal component analysis model for feature extraction. We measure the tuning parameters by their robustness against random perturbation, and select the tuning parameters by derivative-free optimization. We test our algorithm on the ADNI dataset to distinguish between the patients with Alzheimer's disease and the control group. By applying proper classification methods for sparse features, we get better result than classic singular value decomposition, support vector machine and logistic regression.
{"title":"Modified sparse functional principal component analysis for fMRI data process","authors":"Zhengyang Fang, J. Y. Han, N. Simon, Xiaoping Zhou","doi":"10.1080/24709360.2019.1591072","DOIUrl":"https://doi.org/10.1080/24709360.2019.1591072","url":null,"abstract":"Sparse and functional principal component analysis is a technique to extract sparse and smooth principal components from a matrix. In this paper, we propose a modified sparse and functional principal component analysis model for feature extraction. We measure the tuning parameters by their robustness against random perturbation, and select the tuning parameters by derivative-free optimization. We test our algorithm on the ADNI dataset to distinguish between the patients with Alzheimer's disease and the control group. By applying proper classification methods for sparse features, we get better result than classic singular value decomposition, support vector machine and logistic regression.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"80 - 89"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1591072","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46473035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1660111
A. Biswas, Rahul Bhattacharya, Soumyadeep Das
ABSTRACT Weighing the cumulative odds ratios suitably, a two treatment response adaptive design for phase III clinical trial is proposed for ordinal categorical responses. Properties of the proposed design are investigated theoretically as well as empirically. Applicability of the design is further verified using a data pertaining to a real clinical trial with trauma patients, where the responses are observed in an ordinal categorical scale.
{"title":"A response adaptive design for ordinal categorical responses weighing the cumulative odds ratios","authors":"A. Biswas, Rahul Bhattacharya, Soumyadeep Das","doi":"10.1080/24709360.2019.1660111","DOIUrl":"https://doi.org/10.1080/24709360.2019.1660111","url":null,"abstract":"ABSTRACT Weighing the cumulative odds ratios suitably, a two treatment response adaptive design for phase III clinical trial is proposed for ordinal categorical responses. Properties of the proposed design are investigated theoretically as well as empirically. Applicability of the design is further verified using a data pertaining to a real clinical trial with trauma patients, where the responses are observed in an ordinal categorical scale.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"109 - 125"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1660111","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47455932","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1670513
D. Rubin
ABSTRACT Causal inference refers to the process of inferring what would happen in the future if we change what we are doing, or inferring what would have happened in the past, if we had done something different in the distant past. Humans adjust our behaviors by anticipating what will happen if we act in different ways, using past experiences to inform these choices. ‘Essential’ here means in the mathematical sense of excluding the unnecessary and including only the necessary, e.g. stating that the Pythagorean theorem works for an isosceles right triangle is bad mathematics because it includes the unnecessary adjective isosceles; of course this is not as bad as omitting the adjective ‘right.’ I find much of what is written about causal inference to be mathematically inapposite in one of these senses because the descriptions either include irrelevant clutter or omit conditions required for the correctness of the assertions. The history of formal causal inference is remarkable because its correct formulation is so recent, a twentieth century phenomenon, and its future is intriguing because it is currently undeveloped when applied to investigate interventions applied to conscious humans, and moreover will utilize tools impossible without modern computing.
{"title":"Essential concepts of causal inference: a remarkable history and an intriguing future","authors":"D. Rubin","doi":"10.1080/24709360.2019.1670513","DOIUrl":"https://doi.org/10.1080/24709360.2019.1670513","url":null,"abstract":"ABSTRACT Causal inference refers to the process of inferring what would happen in the future if we change what we are doing, or inferring what would have happened in the past, if we had done something different in the distant past. Humans adjust our behaviors by anticipating what will happen if we act in different ways, using past experiences to inform these choices. ‘Essential’ here means in the mathematical sense of excluding the unnecessary and including only the necessary, e.g. stating that the Pythagorean theorem works for an isosceles right triangle is bad mathematics because it includes the unnecessary adjective isosceles; of course this is not as bad as omitting the adjective ‘right.’ I find much of what is written about causal inference to be mathematically inapposite in one of these senses because the descriptions either include irrelevant clutter or omit conditions required for the correctness of the assertions. The history of formal causal inference is remarkable because its correct formulation is so recent, a twentieth century phenomenon, and its future is intriguing because it is currently undeveloped when applied to investigate interventions applied to conscious humans, and moreover will utilize tools impossible without modern computing.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"140 - 155"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1670513","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43617355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1663665
A. Masud, Zhangsheng Yu, W. Tu
Survival data with long-term survivors are common in clinical investigations. Such data are often analyzed with mixture cure rate models. Existing model selection procedures do not readily discriminate nonlinear effects from linear ones. Here, we propose a procedure for accommodating nonlinear effects and for determining the cure rate model composition. The procedure is based on the Least Absolute Shrinkage and Selection Operators (LASSO). Specifically, by partitioning each variable into linear and nonlinear components, we use LASSO to select linear and nonlinear components. Operationally, we model the nonlinear components by cubic B-splines. The procedure adds to the existing variable selection methods an ability to discover hidden nonlinear effects in a cure rate model setting. To implement, we ascertain the maximum likelihood estimates by using an Expectation Maximization (EM) algorithm. We conduct an extensive simulation study to assess the operating characteristics of the selection procedure. We illustrate the use of the method by analyzing data from a real clinical study.
{"title":"Variable selection and nonlinear effect discovery in partially linear mixture cure rate models","authors":"A. Masud, Zhangsheng Yu, W. Tu","doi":"10.1080/24709360.2019.1663665","DOIUrl":"https://doi.org/10.1080/24709360.2019.1663665","url":null,"abstract":"Survival data with long-term survivors are common in clinical investigations. Such data are often analyzed with mixture cure rate models. Existing model selection procedures do not readily discriminate nonlinear effects from linear ones. Here, we propose a procedure for accommodating nonlinear effects and for determining the cure rate model composition. The procedure is based on the Least Absolute Shrinkage and Selection Operators (LASSO). Specifically, by partitioning each variable into linear and nonlinear components, we use LASSO to select linear and nonlinear components. Operationally, we model the nonlinear components by cubic B-splines. The procedure adds to the existing variable selection methods an ability to discover hidden nonlinear effects in a cure rate model setting. To implement, we ascertain the maximum likelihood estimates by using an Expectation Maximization (EM) algorithm. We conduct an extensive simulation study to assess the operating characteristics of the selection procedure. We illustrate the use of the method by analyzing data from a real clinical study.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"156 - 177"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1663665","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47487854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1580463
E. Lorenz, C. Jenkner, W. Sauerbrei, H. Becher
Risk and prognostic factors in epidemiological and clinical research are often semicontinuous such that a proportion of individuals have exposure zero, and a continuous distribution among those exposed. We call this a spike at zero (SAZ). Typical examples are consumption of alcohol and tobacco, or hormone receptor levels. To additionally model non-linear functional relationships for SAZ variables, an extension of the fractional polynomial (FP) approach was proposed. To indicate whether or not a value is zero, a binary variable is added to the model. In a two-stage procedure, called FP-spike, it is assessed whether the binary variable and/or the continuous FP function for the positive part is required for a suitable fit. In this paper, we compared the performance of two approaches – standard FP and FP-spike – in the Cox model in a motivating example on breast cancer prognosis and a simulation study. The comparisons lead to the suggestion to generally using FP-spike rather than standard FP when the SAZ effect is considerably large because the method performed better in real data applications and simulation in terms of deviance and functional form. Abbreviations: CI: confidence interval; FP: fractional polynomial; FP1: first degree fractional polynomial; FP2: second degree fractional polynomial; FSP: function selection procedure; HT: hormone therapy; OR: odds ratio; SAZ: spike at zero
{"title":"Modeling exposures with a spike at zero: simulation study and practical application to survival data","authors":"E. Lorenz, C. Jenkner, W. Sauerbrei, H. Becher","doi":"10.1080/24709360.2019.1580463","DOIUrl":"https://doi.org/10.1080/24709360.2019.1580463","url":null,"abstract":"Risk and prognostic factors in epidemiological and clinical research are often semicontinuous such that a proportion of individuals have exposure zero, and a continuous distribution among those exposed. We call this a spike at zero (SAZ). Typical examples are consumption of alcohol and tobacco, or hormone receptor levels. To additionally model non-linear functional relationships for SAZ variables, an extension of the fractional polynomial (FP) approach was proposed. To indicate whether or not a value is zero, a binary variable is added to the model. In a two-stage procedure, called FP-spike, it is assessed whether the binary variable and/or the continuous FP function for the positive part is required for a suitable fit. In this paper, we compared the performance of two approaches – standard FP and FP-spike – in the Cox model in a motivating example on breast cancer prognosis and a simulation study. The comparisons lead to the suggestion to generally using FP-spike rather than standard FP when the SAZ effect is considerably large because the method performed better in real data applications and simulation in terms of deviance and functional form. Abbreviations: CI: confidence interval; FP: fractional polynomial; FP1: first degree fractional polynomial; FP2: second degree fractional polynomial; FSP: function selection procedure; HT: hormone therapy; OR: odds ratio; SAZ: spike at zero","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"23 - 37"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1580463","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48298967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2019-01-01DOI: 10.1080/24709360.2019.1660110
M. Soltanifar, A. Dupuis, R. Schachar, M. Escobar
The stop signal reaction time (SSRT), a measure of the latency of the stop signal process, has been theoretically formulated using a horse race model of go and stop signal processes by the American scientist Gordon Logan (1994). The SSRT assumes equal impact of the preceding trial type (go/stop) on its measurement. In the case of a violation of this assumption, we consider estimation of SSRT based on the idea of earlier analysis of cluster type go reaction times (GORT) and linear mixed model (LMM) data analysis results. Two clusters of trials were considered including those trials preceded by a go trial and other trials preceded by a stop trial. Given disparities between cluster type SSRTs, we need to consider some new indexes considering the unused cluster type information in the calculations. We introduce mixture SSRT and weighted SSRT as two new distinct indexes of SSRT that address the violated assumption. Mixture SSRT and weighted SSRT are theoretically asymptotically equivalent under special conditions. An example of stop single task (SST) real data is presented to show equivalency of these two new SSRT indexes and their larger magnitude compared to Logan's single 1994 SSRT. Abbreviations: ADHD: attention deficit hyperactivity disorder; ExG: Ex-Gaussiandistribution; GORT: reaction time in a go trial; GORTA: reaction time in a type A gotrial; GORTB: reaction time in a type B go trial; LMM: linear mixed model; SWAN:strengths and weakness of ADHD symptoms and normal behavior rating scale; SSD: stop signal delay; SR: signal respond; SRRT: reaction time in a failedstop trial; SSRT: stop signal reaction times in a stop trial; SST: stop signaltask.
{"title":"A frequentist mixture modeling of stop signal reaction times","authors":"M. Soltanifar, A. Dupuis, R. Schachar, M. Escobar","doi":"10.1080/24709360.2019.1660110","DOIUrl":"https://doi.org/10.1080/24709360.2019.1660110","url":null,"abstract":"The stop signal reaction time (SSRT), a measure of the latency of the stop signal process, has been theoretically formulated using a horse race model of go and stop signal processes by the American scientist Gordon Logan (1994). The SSRT assumes equal impact of the preceding trial type (go/stop) on its measurement. In the case of a violation of this assumption, we consider estimation of SSRT based on the idea of earlier analysis of cluster type go reaction times (GORT) and linear mixed model (LMM) data analysis results. Two clusters of trials were considered including those trials preceded by a go trial and other trials preceded by a stop trial. Given disparities between cluster type SSRTs, we need to consider some new indexes considering the unused cluster type information in the calculations. We introduce mixture SSRT and weighted SSRT as two new distinct indexes of SSRT that address the violated assumption. Mixture SSRT and weighted SSRT are theoretically asymptotically equivalent under special conditions. An example of stop single task (SST) real data is presented to show equivalency of these two new SSRT indexes and their larger magnitude compared to Logan's single 1994 SSRT. Abbreviations: ADHD: attention deficit hyperactivity disorder; ExG: Ex-Gaussiandistribution; GORT: reaction time in a go trial; GORTA: reaction time in a type A gotrial; GORTB: reaction time in a type B go trial; LMM: linear mixed model; SWAN:strengths and weakness of ADHD symptoms and normal behavior rating scale; SSD: stop signal delay; SR: signal respond; SRRT: reaction time in a failedstop trial; SSRT: stop signal reaction times in a stop trial; SST: stop signaltask.","PeriodicalId":37240,"journal":{"name":"Biostatistics and Epidemiology","volume":"3 1","pages":"108 - 90"},"PeriodicalIF":0.0,"publicationDate":"2019-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1080/24709360.2019.1660110","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41851070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}