We introduce a new class of zero-or-one inflated power logit (IPL) regression models, which serve as a versatile tool for analyzing bounded continuous data with observations at a boundary. These models are applied to explore the effects of climate changes on the distribution of tropical tuna within the North Atlantic Ocean. Our findings suggest that our modeling approach is adequate and capable of handling the outliers in the data. It exhibited superior performance compared to rival models in both diagnostic analysis and regarding the inference robustness. We offer a user-friendly method for fitting IPL regression models in practical applications.
{"title":"Modeling tropical tuna shifts: An inflated power logit regression approach","authors":"Francisco F. Queiroz, Silvia L. P. Ferrari","doi":"10.1002/bimj.202300288","DOIUrl":"https://doi.org/10.1002/bimj.202300288","url":null,"abstract":"<p>We introduce a new class of zero-or-one inflated power logit (IPL) regression models, which serve as a versatile tool for analyzing bounded continuous data with observations at a boundary. These models are applied to explore the effects of climate changes on the distribution of tropical tuna within the North Atlantic Ocean. Our findings suggest that our modeling approach is adequate and capable of handling the outliers in the data. It exhibited superior performance compared to rival models in both diagnostic analysis and regarding the inference robustness. We offer a user-friendly method for fitting IPL regression models in practical applications.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 4","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140820550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jiaxin Zhang, S. Ghazaleh Dashti, John B. Carlin, Katherine J. Lee, Margarita Moreno-Betancur
In the context of missing data, the identifiability or “recoverability” of the average causal effect (ACE) depends not only on the usual causal assumptions but also on missingness assumptions that can be depicted by adding variable-specific missingness indicators to causal diagrams, creating missingness directed acyclic graphs (m-DAGs). Previous research described canonical m-DAGs, representing typical multivariable missingness mechanisms in epidemiological studies, and examined mathematically the recoverability of the ACE in each case. However, this work assumed no effect modification and did not investigate methods for estimation across such scenarios. Here, we extend this research by determining the recoverability of the ACE in settings with effect modification and conducting a simulation study to evaluate the performance of widely used missing data methods when estimating the ACE using correctly specified g-computation. Methods assessed were complete case analysis (CCA) and various implementations of multiple imputation (MI) with varying degrees of compatibility with the outcome model used in g-computation. Simulations were based on an example from the Victorian Adolescent Health Cohort Study (VAHCS), where interest was in estimating the ACE of adolescent cannabis use on mental health in young adulthood. We found that the ACE is recoverable when no incomplete variable (exposure, outcome, or confounder) causes its own missingness, and nonrecoverable otherwise, in simplified versions of 10 canonical m-DAGs that excluded unmeasured common causes of missingness indicators. Despite this nonrecoverability, simulations showed that MI approaches that are compatible with the outcome model in g-computation may enable approximately unbiased estimation across all canonical m-DAGs considered, except when the outcome causes its own missingness or causes the missingness of a variable that causes its own missingness. In the latter settings, researchers may need to consider sensitivity analysis methods incorporating external information (e.g., delta-adjustment methods). The VAHCS case study illustrates the practical implications of these findings.
在数据缺失的情况下,平均因果效应(ACE)的可识别性或 "可恢复性 "不仅取决于通常的因果假设,还取决于缺失假设,缺失假设可以通过在因果图中添加特定变量的缺失指标来描述,从而创建缺失有向无环图(m-DAGs)。之前的研究描述了代表流行病学研究中典型多变量缺失机制的典型 m-DAG,并对每种情况下 ACE 的可恢复性进行了数学分析。但是,这项研究假设没有效应修正,也没有研究在这种情况下的估算方法。在此,我们扩展了这项研究,确定了在有效应修饰的情况下 ACE 的可恢复性,并进行了一项模拟研究,以评估广泛使用的缺失数据方法在使用正确指定的 g 计算估计 ACE 时的性能。评估的方法包括完整病例分析(CCA)和多重估算(MI)的各种实现方法,它们与 g 计算中使用的结果模型的兼容程度各不相同。模拟以维多利亚青少年健康队列研究(VAHCS)中的一个实例为基础,该研究的目的是估算青少年使用大麻对其成年后心理健康的影响。我们发现,在简化版的 10 个典型 m-DAG 中,如果没有不完整变量(暴露、结果或混杂因素)导致其自身的缺失,ACE 是可以恢复的,反之则不可恢复。尽管存在这种不可恢复性,但模拟结果表明,在 g 计算中与结果模型兼容的多元智能方法可以在所考虑的所有典型 m-DAG 中实现近似无偏估计,除非结果导致自身缺失或导致导致自身缺失的变量的缺失。在后一种情况下,研究人员可能需要考虑结合外部信息的敏感性分析方法(如三角调整方法)。VAHCS 案例研究说明了这些发现的实际意义。
{"title":"Recoverability and estimation of causal effects under typical multivariable missingness mechanisms","authors":"Jiaxin Zhang, S. Ghazaleh Dashti, John B. Carlin, Katherine J. Lee, Margarita Moreno-Betancur","doi":"10.1002/bimj.202200326","DOIUrl":"https://doi.org/10.1002/bimj.202200326","url":null,"abstract":"<p>In the context of missing data, the identifiability or “recoverability” of the average causal effect (ACE) depends not only on the usual causal assumptions but also on missingness assumptions that can be depicted by adding variable-specific missingness indicators to causal diagrams, creating missingness directed acyclic graphs (m-DAGs). Previous research described canonical m-DAGs, representing typical multivariable missingness mechanisms in epidemiological studies, and examined mathematically the recoverability of the ACE in each case. However, this work assumed no effect modification and did not investigate methods for estimation across such scenarios. Here, we extend this research by determining the recoverability of the ACE in settings with effect modification and conducting a simulation study to evaluate the performance of widely used missing data methods when estimating the ACE using correctly specified g-computation. Methods assessed were complete case analysis (CCA) and various implementations of multiple imputation (MI) with varying degrees of compatibility with the outcome model used in g-computation. Simulations were based on an example from the Victorian Adolescent Health Cohort Study (VAHCS), where interest was in estimating the ACE of adolescent cannabis use on mental health in young adulthood. We found that the ACE is recoverable when no incomplete variable (exposure, outcome, or confounder) causes its own missingness, and nonrecoverable otherwise, in simplified versions of 10 canonical m-DAGs that excluded unmeasured common causes of missingness indicators. Despite this nonrecoverability, simulations showed that MI approaches that are compatible with the outcome model in g-computation may enable approximately unbiased estimation across all canonical m-DAGs considered, except when the outcome causes its own missingness or causes the missingness of a variable that causes its own missingness. In the latter settings, researchers may need to consider sensitivity analysis methods incorporating external information (e.g., delta-adjustment methods). The VAHCS case study illustrates the practical implications of these findings.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.202200326","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In screening large populations a diagnostic test is frequently used repeatedly. An example is screening for bowel cancer using the fecal occult blood test (FOBT) on several occasions such as at 3 or 6 days. The question that is addressed here is how often should we repeat a diagnostic test when screening for a specific medical condition. Sensitivity is often used as a performance measure of a diagnostic test and is considered here for the individual application of the diagnostic test as well as for the overall screening procedure. The latter can involve an increasingly large number of repeated applications, but how many are sufficient? We demonstrate the issues involved in answering this question using real data on bowel cancer at St Vincents Hospital in Sydney. As data are only available for those testing positive at least once, an appropriate modeling technique is developed on the basis of the zero-truncated binomial distribution which allows for population heterogeneity. The latter is modeled using discrete nonparametric maximum likelihood. If we wish to achieve an overall sensitivity of 90%, the FOBT should be repeated for 2 weeks instead of the 1 week that was used at the time of the survey. A simulation study also shows consistency in the sense that bias and standard deviation for the estimated sensitivity decrease with an increasing number of repeated occasions as well as with increasing sample size.
{"title":"On repeated diagnostic testing in screening for a medical condition: How often should the diagnostic test be repeated?","authors":"Patarawan Sangnawakij, Dankmar Böhning","doi":"10.1002/bimj.202300175","DOIUrl":"https://doi.org/10.1002/bimj.202300175","url":null,"abstract":"<p>In screening large populations a diagnostic test is frequently used repeatedly. An example is screening for bowel cancer using the fecal occult blood test (FOBT) on several occasions such as at 3 or 6 days. The question that is addressed here is how often should we repeat a diagnostic test when screening for a specific medical condition. Sensitivity is often used as a performance measure of a diagnostic test and is considered here for the individual application of the diagnostic test as well as for the overall screening procedure. The latter can involve an increasingly large number of repeated applications, but how many are sufficient? We demonstrate the issues involved in answering this question using real data on bowel cancer at St Vincents Hospital in Sydney. As data are only available for those testing positive at least once, an appropriate modeling technique is developed on the basis of the zero-truncated binomial distribution which allows for population heterogeneity. The latter is modeled using discrete nonparametric maximum likelihood. If we wish to achieve an overall sensitivity of 90%, the FOBT should be repeated for 2 weeks instead of the 1 week that was used at the time of the survey. A simulation study also shows consistency in the sense that bias and standard deviation for the estimated sensitivity decrease with an increasing number of repeated occasions as well as with increasing sample size.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper, we consider online multiple testing with familywise error rate (FWER) control, where the probability of committing at least one type I error will remain under control while testing a possibly infinite sequence of hypotheses over time. Currently, adaptive-discard (ADDIS) procedures seem to be the most promising online procedures with FWER control in terms of power. Now, our main contribution is a uniform improvement of the ADDIS principle and thus of all ADDIS procedures. This means, the methods we propose reject as least as much hypotheses as ADDIS procedures and in some cases even more, while maintaining FWER control. In addition, we show that there is no other FWER controlling procedure that enlarges the event of rejecting any hypothesis. Finally, we apply the new principle to derive uniform improvements of the ADDIS-Spending and ADDIS-Graph.
{"title":"An exhaustive ADDIS principle for online FWER control","authors":"Lasse Fischer, Marta Bofill Roig, Werner Brannath","doi":"10.1002/bimj.202300237","DOIUrl":"https://doi.org/10.1002/bimj.202300237","url":null,"abstract":"<p>In this paper, we consider online multiple testing with familywise error rate (FWER) control, where the probability of committing at least one type I error will remain under control while testing a possibly infinite sequence of hypotheses over time. Currently, adaptive-discard (ADDIS) procedures seem to be the most promising online procedures with FWER control in terms of power. Now, our main contribution is a uniform improvement of the ADDIS principle and thus of all ADDIS procedures. This means, the methods we propose reject as least as much hypotheses as ADDIS procedures and in some cases even more, while maintaining FWER control. In addition, we show that there is no other FWER controlling procedure that enlarges the event of rejecting any hypothesis. Finally, we apply the new principle to derive uniform improvements of the ADDIS-Spending and ADDIS-Graph.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.202300237","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In order to assess prognostic risk for individuals in precision health research, risk prediction models are increasingly used, in which statistical models are used to estimate the risk of future outcomes based on clinical and nonclinical characteristics. The predictive accuracy of a risk score must be assessed before it can be used in routine clinical decision making, where the receiver operator characteristic curves, precision–recall curves, and their corresponding area under the curves are commonly used metrics to evaluate the discriminatory ability of a continuous risk score. Among these the precision–recall curves have been shown to be more informative when dealing with unbalanced biomarker distribution between classes, which is common in rare event, even though except one, all existing methods are proposed for classic uncensored data. This paper is therefore to propose a novel nonparametric estimation approach for the time-dependent precision–recall curve and its associated area under the curve for right-censored data. A simulation is conducted to show the better finite sample property of the proposed estimator over the existing method and a real-world data from primary biliary cirrhosis trial is used to demonstrate the practical applicability of the proposed estimator.
{"title":"A novel nonparametric time-dependent precision–recall curve estimator for right-censored survival data","authors":"Kassu Mehari Beyene, Ding-Geng Chen, Yehenew Getachew Kifle","doi":"10.1002/bimj.202300135","DOIUrl":"https://doi.org/10.1002/bimj.202300135","url":null,"abstract":"<p>In order to assess prognostic risk for individuals in precision health research, risk prediction models are increasingly used, in which statistical models are used to estimate the risk of future outcomes based on clinical and nonclinical characteristics. The predictive accuracy of a risk score must be assessed before it can be used in routine clinical decision making, where the receiver operator characteristic curves, precision–recall curves, and their corresponding area under the curves are commonly used metrics to evaluate the discriminatory ability of a continuous risk score. Among these the precision–recall curves have been shown to be more informative when dealing with unbalanced biomarker distribution between classes, which is common in rare event, even though except one, all existing methods are proposed for classic uncensored data. This paper is therefore to propose a novel nonparametric estimation approach for the time-dependent precision–recall curve and its associated area under the curve for right-censored data. A simulation is conducted to show the better finite sample property of the proposed estimator over the existing method and a real-world data from primary biliary cirrhosis trial is used to demonstrate the practical applicability of the proposed estimator.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.202300135","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Theodoros Evrenoglou, Silvia Metelli, Johannes-Schneider Thomas, Spyridon Siafis, Rebecca M. Turner, Stefan Leucht, Anna Chaimani
Network meta-analysis (NMA) usually provides estimates of the relative effects with the highest possible precision. However, sparse networks with few available studies and limited direct evidence can arise, threatening the robustness and reliability of NMA estimates. In these cases, the limited amount of available information can hamper the formal evaluation of the underlying NMA assumptions of transitivity and consistency. In addition, NMA estimates from sparse networks are expected to be imprecise and possibly biased as they rely on large-sample approximations that are invalid in the absence of sufficient data. We propose a Bayesian framework that allows sharing of information between two networks that pertain to different population subgroups. Specifically, we use the results from a subgroup with a lot of direct evidence (a dense network) to construct informative priors for the relative effects in the target subgroup (a sparse network). This is a two-stage approach where at the first stage, we extrapolate the results of the dense network to those expected from the sparse network. This takes place by using a modified hierarchical NMA model where we add a location parameter that shifts the distribution of the relative effects to make them applicable to the target population. At the second stage, these extrapolated results are used as prior information for the sparse network. We illustrate our approach through a motivating example of psychiatric patients. Our approach results in more precise and robust estimates of the relative effects and can adequately inform clinical practice in presence of sparse networks.
{"title":"Sharing information across patient subgroups to draw conclusions from sparse treatment networks","authors":"Theodoros Evrenoglou, Silvia Metelli, Johannes-Schneider Thomas, Spyridon Siafis, Rebecca M. Turner, Stefan Leucht, Anna Chaimani","doi":"10.1002/bimj.202200316","DOIUrl":"https://doi.org/10.1002/bimj.202200316","url":null,"abstract":"<p>Network meta-analysis (NMA) usually provides estimates of the relative effects with the highest possible precision. However, sparse networks with few available studies and limited direct evidence can arise, threatening the robustness and reliability of NMA estimates. In these cases, the limited amount of available information can hamper the formal evaluation of the underlying NMA assumptions of transitivity and consistency. In addition, NMA estimates from sparse networks are expected to be imprecise and possibly biased as they rely on large-sample approximations that are invalid in the absence of sufficient data. We propose a Bayesian framework that allows sharing of information between two networks that pertain to different population subgroups. Specifically, we use the results from a subgroup with a lot of direct evidence (a dense network) to construct informative priors for the relative effects in the target subgroup (a sparse network). This is a two-stage approach where at the first stage, we extrapolate the results of the dense network to those expected from the sparse network. This takes place by using a modified hierarchical NMA model where we add a location parameter that shifts the distribution of the relative effects to make them applicable to the target population. At the second stage, these extrapolated results are used as prior information for the sparse network. We illustrate our approach through a motivating example of psychiatric patients. Our approach results in more precise and robust estimates of the relative effects and can adequately inform clinical practice in presence of sparse networks.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.202200316","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Rank methods are well-established tools for comparing two or multiple (independent) groups. Statistical planning methods for the computing the required sample size(s) to detect a specific alternative with predefined power are lacking. In the present paper, we develop numerical algorithms for sample size planning of pseudo-rank-based multiple contrast tests. We discuss the treatment effects and different ways to approximate variance parameters within the estimation scheme. We further compare pairwise with global rank methods in detail. Extensive simulation studies show that the sample size estimators are accurate. A real data example illustrates the application of the methods.
{"title":"Sample size planning for rank-based multiple contrast tests","authors":"Anna Pöhlmann, Edgar Brunner, Frank Konietschke","doi":"10.1002/bimj.202300240","DOIUrl":"https://doi.org/10.1002/bimj.202300240","url":null,"abstract":"<p>Rank methods are well-established tools for comparing two or multiple (independent) groups. Statistical planning methods for the computing the required sample size(s) to detect a specific alternative with predefined power are lacking. In the present paper, we develop numerical algorithms for sample size planning of pseudo-rank-based multiple contrast tests. We discuss the treatment effects and different ways to approximate variance parameters within the estimation scheme. We further compare pairwise with global rank methods in detail. Extensive simulation studies show that the sample size estimators are accurate. A real data example illustrates the application of the methods.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.202300240","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140619793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The research on the quantitative trait locus (QTL) mapping of count data has aroused the wide attention of researchers. There are frequent problems in applied research that limit the application of the conventional Poisson model in the analysis of count phenotypes, which include the overdispersion and excess zeros and ones. In this article, a novel model, that is, the zero-and-one-inflated generalized Poisson (ZOIGP) model, is proposed to deal with these problems. Based on the proposed model, a score test is performed for the inflation parameter, in which the ZOIGP model with a constant proportion of excess zeros and ones is compared with a standard generalized Poisson model. To illustrate the practicability of the ZOIGP model, we extend it to the QTL interval mapping application that underpins count phenotype with excess zeros and excess ones. The genetic effects are estimated utilizing the expectation–maximization algorithm embedded with the Newton–Raphson algorithm, and the genome-wide scan and likelihood ratio test is performed to map and test the potential QTLs. The statistical properties exhibited by the proposed method are investigated through simulation. Finally, a real data analysis example is used to illustrate the utility of the proposed method for QTL mapping.
{"title":"Mapping QTL controlling count traits with excess zeros and ones using a zero-and-one-inflated generalized Poisson regression model","authors":"Jinling Chi, Jimin Ye, Ying Zhou","doi":"10.1002/bimj.202200342","DOIUrl":"https://doi.org/10.1002/bimj.202200342","url":null,"abstract":"<p>The research on the quantitative trait locus (QTL) mapping of count data has aroused the wide attention of researchers. There are frequent problems in applied research that limit the application of the conventional Poisson model in the analysis of count phenotypes, which include the overdispersion and excess zeros and ones. In this article, a novel model, that is, the zero-and-one-inflated generalized Poisson (ZOIGP) model, is proposed to deal with these problems. Based on the proposed model, a score test is performed for the inflation parameter, in which the ZOIGP model with a constant proportion of excess zeros and ones is compared with a standard generalized Poisson model. To illustrate the practicability of the ZOIGP model, we extend it to the QTL interval mapping application that underpins count phenotype with excess zeros and excess ones. The genetic effects are estimated utilizing the expectation–maximization algorithm embedded with the Newton–Raphson algorithm, and the genome-wide scan and likelihood ratio test is performed to map and test the potential QTLs. The statistical properties exhibited by the proposed method are investigated through simulation. Finally, a real data analysis example is used to illustrate the utility of the proposed method for QTL mapping.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140552879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In a two-way additive analysis of variance (ANOVA) model, we consider the problem of testing for homogeneity of both row and column effects against their simultaneous ordering. The error variances are assumed to be heterogeneous with unbalanced samples in each cell. Two simultaneous test procedures are developed—the first one using the likelihood ratio test (LRT) statistics of two independent hypotheses and another based on the consecutive pairwise differences of estimators of effects. The parametric bootstrap (PB) approach is used to find critical points of both the tests and the asymptotic accuracy of the bootstrap is established. An extensive simulation study shows that the proposed tests achieve the nominal size and have very good power performance. The robustness of the tests is also analyzed under deviation from normality. An “R” package is developed and shared on “GitHub” for ease of implementation of users. The proposed tests are illustrated using a real data set on the mortality due to alcoholic liver disease and it is shown that age and gender have a significant impact on the increasing incidence of mortality.
{"title":"Bootstrap tests for simultaneous monotone ordering of effects in a two-way ANOVA","authors":"Raju Dey, Anjana Mondal, Somesh Kumar","doi":"10.1002/bimj.202300238","DOIUrl":"https://doi.org/10.1002/bimj.202300238","url":null,"abstract":"<p>In a two-way additive analysis of variance (ANOVA) model, we consider the problem of testing for homogeneity of both row and column effects against their simultaneous ordering. The error variances are assumed to be heterogeneous with unbalanced samples in each cell. Two simultaneous test procedures are developed—the first one using the likelihood ratio test (LRT) statistics of two independent hypotheses and another based on the consecutive pairwise differences of estimators of effects. The parametric bootstrap (PB) approach is used to find critical points of both the tests and the asymptotic accuracy of the bootstrap is established. An extensive simulation study shows that the proposed tests achieve the nominal size and have very good power performance. The robustness of the tests is also analyzed under deviation from normality. An “R” package is developed and shared on “GitHub” for ease of implementation of users. The proposed tests are illustrated using a real data set on the mortality due to alcoholic liver disease and it is shown that age and gender have a significant impact on the increasing incidence of mortality.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"66 3","pages":""},"PeriodicalIF":1.7,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140351577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}