Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.10.002
Mohamed Chaouch , Naâmane Laïb , Elias Ould Saïd
The present paper deals with a nonparametric -estimation for right censored regression model with stationary ergodic data. Defined as an implicit function, a kernel-type estimator of a family of robust regression is considered when the covariate takes its values in () and the data are sampled from a stationary ergodic process. The strong consistency (with rate) and the asymptotic distribution of the estimator are established under mild assumptions. Moreover, a usable confidence interval is provided which does not depend on any unknown quantity. Our results hold without any mixing condition and do not require the existence of marginal densities. A comparison study based on simulated data is also provided.
{"title":"Nonparametric M-estimation for right censored regression model with stationary ergodic data","authors":"Mohamed Chaouch , Naâmane Laïb , Elias Ould Saïd","doi":"10.1016/j.stamet.2016.10.002","DOIUrl":"10.1016/j.stamet.2016.10.002","url":null,"abstract":"<div><p>The present paper deals with a nonparametric <span><math><mi>M</mi></math></span><span><span>-estimation for right censored regression model with stationary ergodic data. Defined as an implicit function, a kernel-type estimator of a family of robust regression is considered when the </span>covariate takes its values in </span><span><math><msup><mrow><mi>R</mi></mrow><mrow><mi>d</mi></mrow></msup></math></span> (<span><math><mi>d</mi><mo>≥</mo><mn>1</mn></math></span>) and the data are sampled from a <em>stationary ergodic process</em><span>. The strong consistency (with rate) and the asymptotic distribution of the estimator are established under mild assumptions. Moreover, a usable confidence interval is provided which does not depend on any unknown quantity. Our results hold without any mixing condition and do not require the existence of marginal densities. A comparison study based on simulated data is also provided.</span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 234-255"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.10.002","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125822228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.08.006
Sudipta Das, Anup Dewanji, Debasis Sengupta
In many situations, multiple copies of a software are tested in parallel with different test cases as input, and the detected errors from a particular round of testing are debugged together. In this article, we discuss a discrete time model of software reliability for such a scenario of periodic debugging. We propose likelihood based inference of the model parameters, including the initial number of errors, under the assumption that all errors are equally likely to be detected. The proposed method is used to estimate the reliability of the software. We establish asymptotic normality of the estimated model parameters. The performance of the proposed method is evaluated through a simulation study and its use is illustrated through the analysis of a dataset obtained from testing of a real-time flight control software. We also consider a more general model, in which different errors have different probabilities of detection.
{"title":"Discrete time software reliability modeling with periodic debugging schedule","authors":"Sudipta Das, Anup Dewanji, Debasis Sengupta","doi":"10.1016/j.stamet.2016.08.006","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.08.006","url":null,"abstract":"<div><p>In many situations, multiple copies of a software are tested in parallel with different test cases as input, and the detected errors from a particular round of testing are debugged together. In this article, we discuss a discrete time model of software reliability for such a scenario of periodic debugging. We propose likelihood based inference of the model parameters, including the initial number of errors, under the assumption that all errors are equally likely to be detected. The proposed method is used to estimate the reliability of the software. We establish asymptotic normality<span> of the estimated model parameters<span>. The performance of the proposed method is evaluated through a simulation study and its use is illustrated through the analysis of a dataset obtained from testing of a real-time flight control software. We also consider a more general model, in which different errors have different probabilities of detection.</span></span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 147-159"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.08.006","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.07.001
Sun-Hee Kim , Lei Li , Christos Faloutsos , Hyung-Jeong Yang , Seong-Whan Lee
Acute hypotensive episodes (AHEs) are serious clinical events in intensive care units (ICUs), and require immediate treatment to prevent patient injury. Reducing the risks associated with an AHE requires effective and efficient mining of data generated from multiple physiological time series. We propose HeartCast, a model that extracts essential features from such data to effectively predict AHE. HeartCast combines a non-linear support vector machine with best-feature extraction via analysis of the baseline threshold, quartile parameters, and window size of the physiological signals. Our approach has the following benefits: (a) it extracts the most relevant features; (b) it provides the best results for identification of an AHE event; (c) it is fast and scales with linear complexity over the length of the window; and (d) it can manage missing values and noise/outliers by using a best-feature extraction method. We performed experiments on data continuously captured from physiological time series of ICU patients (roughly 3 GB of processed data). HeartCast was found to outperform other state-of-the-art methods found in the literature with a 13.7% improvement in classification accuracy.
{"title":"HeartCast: Predicting acute hypotensive episodes in intensive care units","authors":"Sun-Hee Kim , Lei Li , Christos Faloutsos , Hyung-Jeong Yang , Seong-Whan Lee","doi":"10.1016/j.stamet.2016.07.001","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.07.001","url":null,"abstract":"<div><p>Acute hypotensive episodes (AHEs) are serious clinical events in intensive care units (ICUs), and require immediate treatment to prevent patient injury. Reducing the risks associated with an AHE requires effective and efficient mining of data generated from multiple physiological time series. We propose HeartCast, a model that extracts essential features from such data to effectively predict AHE. HeartCast combines a non-linear support vector machine with best-feature extraction via analysis of the baseline threshold, quartile parameters, and window size of the physiological signals. Our approach has the following benefits: (a) it extracts the most relevant features; (b) it provides the best results for identification of an AHE event; (c) it is fast and scales with linear complexity over the length of the window; and (d) it can manage missing values and noise/outliers by using a best-feature extraction method. We performed experiments on data continuously captured from physiological time series of ICU patients (roughly 3 GB of processed data). HeartCast was found to outperform other state-of-the-art methods found in the literature with a 13.7% improvement in classification accuracy.</p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 1-13"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.07.001","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.09.001
Gregory E. Wilding, Mark C. Baker
The testing of equality of several Pearson correlations can be found in a number of scientific fields. We surmise in many such cases that the alternatives of interest in practice are, in deed, order restricted, and therefore the researcher is best served by use of testing procedures developed for those specific alternatives. In this note we introduce a collection of tests for use in testing equality of correlation coefficients against order alternatives, with an emphasis on simple order. Specifically, we propose likelihood ratio tests and contrast tests based on the well known Fisher Z transformation as well as tests which make use of generalized variable methodologies. The proposed procedures are empirically compared with regard to type I and II error rates via Monte Carlo simulations studies, and the use of the approaches is illustrated using an example. These tests are found to be vastly superior to tests for the general alternative, and the contrast tests based on the Fisher Z transformation are recommended for practice based on the observed test properties and simplicity.
{"title":"Inference procedures about population correlations under order restrictions","authors":"Gregory E. Wilding, Mark C. Baker","doi":"10.1016/j.stamet.2016.09.001","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.09.001","url":null,"abstract":"<div><p><span>The testing of equality of several Pearson correlations can be found in a number of scientific fields. We surmise in many such cases that the alternatives of interest in practice are, in deed, order restricted, and therefore the researcher is best served by use of testing procedures developed for those specific alternatives. In this note we introduce a collection of tests for use in testing equality of </span><span><math><mi>k</mi></math></span><span><span> correlation coefficients against order alternatives, with an emphasis on simple order. Specifically, we propose </span>likelihood ratio tests<span> and contrast tests based on the well known Fisher Z transformation as well as tests which make use of generalized variable methodologies. The proposed procedures are empirically compared with regard to type I and II error rates via Monte Carlo simulations studies, and the use of the approaches is illustrated using an example. These tests are found to be vastly superior to tests for the general alternative, and the contrast tests based on the Fisher Z transformation are recommended for practice based on the observed test properties and simplicity.</span></span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 203-216"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.09.001","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.07.003
Elvan Ceyhan
We consider two types of graphs based on a family of proximity catch digraphs (PCDs) and study their edge density. In particular, the PCDs we use are a parameterized digraph family called proportional-edge (PE) PCDs and the two associated graph types are the “underlying graphs” and the newly introduced “reflexivity graphs” based on the PE-PCDs. These graphs are extensions of random geometric graphs where distance is replaced with a dissimilarity measure and the threshold is not fixed but depends on the location of the points. PCDs and the associated graphs are constructed based on data points from two classes, say and , where one class (say class ) forms the vertices of the PCD and the Delaunay tessellation of the other class (i.e., class ) yields the (Delaunay) cells which serve as the support of class points. We demonstrate that edge density of these graphs is a -statistic, hence obtain the asymptotic normality of it for data from any distribution that satisfies mild regulatory conditions. The rate of convergence to asymptotic normality is sharper for the edge density of the reflexivity and underlying graphs compared to the arc density of the PE-PCDs. For uniform data in Euclidean plane where Delaunay cells are triangles, we demonstrate that the distribution of the edge density is geometry invariant (i.e., independent of the shape of the triangular support). We compute the explicit forms of the asymptotic normal distribution for uniform data in one Delaunay triangle in the Euclidean plane utilizing this geometry invariance property. We also provide various versions of edge density in the multiple triangle case. The approach presented here can also be extended for application to data in higher dimensions.
{"title":"Edge density of new graph types based on a random digraph family","authors":"Elvan Ceyhan","doi":"10.1016/j.stamet.2016.07.003","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.07.003","url":null,"abstract":"<div><p>We consider two types of graphs based on a family of proximity catch digraphs (PCDs) and study their edge density. In particular, the PCDs we use are a parameterized digraph family called proportional-edge (PE) PCDs and the two associated graph types are the “underlying graphs” and the newly introduced “reflexivity graphs” based on the PE-PCDs. These graphs are extensions of random geometric graphs where distance is replaced with a dissimilarity measure and the threshold is not fixed but depends on the location of the points. PCDs and the associated graphs are constructed based on data points from two classes, say <span><math><mi>X</mi></math></span> and <span><math><mi>Y</mi></math></span>, where one class (say class <span><math><mi>X</mi></math></span>) forms the vertices of the PCD and the Delaunay tessellation of the other class (i.e., class <span><math><mi>Y</mi></math></span>) yields the (Delaunay) cells which serve as the support of class <span><math><mi>X</mi></math></span> points. We demonstrate that edge density of these graphs is a <span><math><mi>U</mi></math></span><span>-statistic, hence obtain the asymptotic normality<span><span> of it for data from any distribution that satisfies mild regulatory conditions. The rate of convergence to asymptotic normality is sharper for the edge density of the reflexivity and underlying graphs compared to the arc density of the PE-PCDs. For uniform data in </span>Euclidean plane<span> where Delaunay cells are triangles, we demonstrate that the distribution of the edge density is geometry invariant (i.e., independent of the shape of the triangular support). We compute the explicit forms of the asymptotic normal distribution<span><span> for uniform data in one Delaunay triangle in the Euclidean plane utilizing this geometry invariance property. We also provide various versions of edge density in the multiple triangle case. The approach presented here can also be extended for application to data in </span>higher dimensions.</span></span></span></span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 31-54"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.07.003","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.08.001
Michael Brimacombe
A general diagnostic approach to the evaluation of asymptotic approximation in likelihood based models is developed and applied to logistic regression. The expected asymptotic and observed log-likelihood functions are compared using a chi distribution in a directional Bayesian setting. This provides a general approach to assessing and visualizing non-convergence in higher dimensional models. Several well-known examples from the logistic regression literature are discussed.
{"title":"Large sample convergence diagnostics for likelihood based inference: Logistic regression","authors":"Michael Brimacombe","doi":"10.1016/j.stamet.2016.08.001","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.08.001","url":null,"abstract":"<div><p>A general diagnostic approach to the evaluation of asymptotic approximation in likelihood based models is developed and applied to logistic regression. The expected asymptotic and observed log-likelihood functions are compared using a chi distribution in a directional Bayesian setting. This provides a general approach to assessing and visualizing non-convergence in higher dimensional models. Several well-known examples from the logistic regression literature are discussed.</p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 114-130"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.08.001","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.08.005
Félix Almendra-Arao , José Juan Castro-Alva , Hortensia Reyes-Cervantes
In both statistical non-inferiority (NI) and superiority (S) tests, the critical region must be a Barnard convex set for two main reasons. One, being computational in nature, based on the fact that calculating test sizes is a computationally intensive problem due to the presence of a nuisance parameter. However, this calculation is considerably reduced when the critical region is a Barnard convex set. The other reason is that in order for the NI/S statistical tests to make sense, its critical regions must be Barnard convex sets. While it is indeed possible for NI/S tests’ critical regions to not be Barnard convex sets, for the reasons stated above, it is desirable that they are. Therefore, it is important to generate, from a given NI/S test, a test which guarantees that the critical regions are Barnard convex sets. We propose a method by which, from a given NI/S test, we construct another NI/S test, ensuring that the critical regions corresponding to the modified test are Barnard convex sets, we illustrate this through examples. This work is theoretical because the type of developments refers to the general framework of NI/S testing for two independent binomial proportions and it is applied because statistical tests that do not ensure that their critical regions are Barnard convex sets may appear in practice, particularly in the clinical trials area.
{"title":"Constructing tests to compare two proportions whose critical regions guarantee to be Barnard convex sets","authors":"Félix Almendra-Arao , José Juan Castro-Alva , Hortensia Reyes-Cervantes","doi":"10.1016/j.stamet.2016.08.005","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.08.005","url":null,"abstract":"<div><p>In both statistical non-inferiority (NI) and superiority (S) tests, the critical region must be a Barnard convex set<span> for two main reasons. One, being computational in nature, based on the fact that calculating test sizes is a computationally intensive problem due to the presence of a nuisance parameter<span>. However, this calculation is considerably reduced when the critical region is a Barnard convex set. The other reason is that in order for the NI/S statistical tests to make sense, its critical regions must be Barnard convex sets. While it is indeed possible for NI/S tests’ critical regions to not be Barnard convex sets, for the reasons stated above, it is desirable that they are. Therefore, it is important to generate, from a given NI/S test, a test which guarantees that the critical regions are Barnard convex sets. We propose a method by which, from a given NI/S test, we construct another NI/S test, ensuring that the critical regions corresponding to the modified test are Barnard convex sets, we illustrate this through examples. This work is theoretical because the type of developments refers to the general framework of NI/S testing for two independent binomial proportions and it is applied because statistical tests that do not ensure that their critical regions are Barnard convex sets may appear in practice, particularly in the clinical trials area.</span></span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 160-171"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.08.005","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.09.004
Rasul A. Khan
A problem for estimating the number of trials in the binomial distribution , is revisited by considering the large sample model and the associated maximum likelihood estimator (MLE) and some sequential procedures. Asymptotic properties of the MLE of via the normal model are briefly described. Beyond the asymptotic properties, our main focus is on the sequential estimation of . Let be iid random variables with an unknown mean and variance , where is known. The sequential estimation of is explored by a method initiated by Robbins (1970) and further pursued by Khan (1973). Various properties of the procedure including the error probability and the expected sample size are determined. An asymptotic optimality of the procedure is given. Sequential interval estimation and point estimation are also briefly discussed.
{"title":"Estimating the integer mean of a normal model related to binomial distribution","authors":"Rasul A. Khan","doi":"10.1016/j.stamet.2016.09.004","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.09.004","url":null,"abstract":"<div><p>A problem for estimating the number of trials <span><math><mi>n</mi></math></span><span> in the binomial distribution </span><span><math><mi>B</mi><mrow><mo>(</mo><mi>n</mi><mo>,</mo><mi>p</mi><mo>)</mo></mrow></math></span>, is revisited by considering the large sample model <span><math><mi>N</mi><mrow><mo>(</mo><mi>μ</mi><mo>,</mo><mi>c</mi><mi>μ</mi><mo>)</mo></mrow></math></span><span><span> and the associated maximum likelihood estimator (MLE) and some sequential procedures. </span>Asymptotic properties of the MLE of </span><span><math><mi>n</mi></math></span> via the normal model <span><math><mi>N</mi><mrow><mo>(</mo><mi>μ</mi><mo>,</mo><mi>c</mi><mi>μ</mi><mo>)</mo></mrow></math></span> are briefly described. Beyond the asymptotic properties, our main focus is on the sequential estimation of <span><math><mi>n</mi></math></span>. Let <span><math><msub><mrow><mi>X</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>,</mo><msub><mrow><mi>X</mi></mrow><mrow><mn>2</mn></mrow></msub><mo>,</mo><mo>…</mo><mo>,</mo><msub><mrow><mi>X</mi></mrow><mrow><mi>m</mi></mrow></msub><mo>,</mo><mo>…</mo></math></span> be iid <span><math><mi>N</mi><mrow><mo>(</mo><mi>μ</mi><mo>,</mo><mi>c</mi><mi>μ</mi><mo>)</mo></mrow></math></span><span><math><mrow><mo>(</mo><mi>c</mi><mo>></mo><mn>0</mn><mo>)</mo></mrow></math></span> random variables with an unknown mean <span><math><mi>μ</mi><mo>=</mo><mn>1</mn><mo>,</mo><mn>2</mn><mo>,</mo><mo>…</mo></math></span> and variance <span><math><mi>c</mi><mspace></mspace><mi>μ</mi></math></span>, where <span><math><mi>c</mi></math></span> is known. The sequential estimation of <span><math><mi>μ</mi></math></span><span> is explored by a method initiated by Robbins (1970) and further pursued by Khan (1973). Various properties of the procedure including the error probability<span> and the expected sample size are determined. An asymptotic optimality<span> of the procedure is given. Sequential interval estimation and point estimation are also briefly discussed.</span></span></span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 192-202"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.09.004","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.07.002
David Stibůrek
In statistical inference on the drift parameter in the process , where and are known, deterministic functions, there is known a large number of options how to do it. We may, for example, base this inference on the differences between the observed values of the process at discrete times and their normality. Although such methods are very simple, it turns out that it is more appropriate to use sequential methods. For the hypotheses testing about the drift parameter , it is more proper to standardize the observed process and to use sequential methods based on the first exit time of the observed process of a pre-specified interval until some given time. These methods can be generalized to the case of random part being a symmetric Itô integral or continuous symmetric martingale.
{"title":"Sequential testing of hypotheses about drift for Gaussian diffusions","authors":"David Stibůrek","doi":"10.1016/j.stamet.2016.07.002","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.07.002","url":null,"abstract":"<div><p><span>In statistical inference<span> on the drift parameter </span></span><span><math><mi>θ</mi></math></span> in the process <span><math><msub><mrow><mi>X</mi></mrow><mrow><mi>t</mi></mrow></msub><mo>=</mo><mi>θ</mi><mi>a</mi><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow><mo>+</mo><msubsup><mrow><mo>∫</mo></mrow><mrow><mn>0</mn></mrow><mrow><mi>t</mi></mrow></msubsup><mi>b</mi><mrow><mo>(</mo><mi>s</mi><mo>)</mo></mrow><mstyle><mi>d</mi></mstyle><msub><mrow><mi>W</mi></mrow><mrow><mi>s</mi></mrow></msub></math></span>, where <span><math><mi>a</mi><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow></math></span> and <span><math><mi>b</mi><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow></math></span><span><span> are known, deterministic functions, there is known a large number of options how to do it. We may, for example, base this inference on the differences between the observed values of the process at discrete times and their normality. Although such methods are very simple, it turns out that it is more appropriate to use sequential methods. For the </span>hypotheses testing about the drift parameter </span><span><math><mi>θ</mi></math></span><span>, it is more proper to standardize the observed process and to use sequential methods based on the first exit time of the observed process of a pre-specified interval until some given time. These methods can be generalized to the case of random part being a symmetric Itô integral or continuous symmetric martingale.</span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 14-30"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.07.002","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2016-12-01DOI: 10.1016/j.stamet.2016.08.004
Clécio S. Ferreira , Víctor H. Lachos
Normal nonlinear regression models are applied in some areas of the sciences and engineering to explain or describe the phenomena under study. However, it is well known that several phenomena are not always represented by the normal model due to lack of symmetry or the presence of heavy- and light-tailed distributions related to the normal law in the data. This paper proposes an extension of nonlinear regression models using the skew-scale mixtures of normal (SSMN) distributions proposed by Ferreira et al. (2011). This class of models provides a useful generalization of the symmetrical nonlinear regression models since the random term distributions cover both asymmetric and heavy-tailed distributions, such as the skew--normal, skew-slash and skew-contaminated normal, among others. An expectation–maximization (EM) algorithm for maximum likelihood (ML) estimates is presented and the observed information matrix is derived analytically. Some simulation studies are presented to examine the performance of the proposed methods, with relation to robustness and asymptotic properties of the ML estimates. Finally, an illustration of the method is presented considering a dataset previously analyzed under normal and skew-normal (SN) nonlinear regression models. The main conclusion is that the ML estimates from the heavy tails SSMN nonlinear models are more robust against outlying observations compared to the corresponding SN estimates.
正态非线性回归模型被应用于科学和工程的某些领域来解释或描述所研究的现象。然而,众所周知,由于缺乏对称性或数据中存在与正态律相关的重尾和轻尾分布,一些现象并不总是用正态模型来表示。本文利用Ferreira et al.(2011)提出的斜尺度混合正态分布(SSMN)对非线性回归模型进行了扩展。这类模型提供了对称非线性回归模型的有用推广,因为随机项分布涵盖了不对称和重尾分布,如斜t正态、斜斜线和斜污染正态等。提出了一种最大似然估计的期望最大化算法,并解析导出了观测到的信息矩阵。提出了一些仿真研究来检验所提出的方法的性能,以及与ML估计的鲁棒性和渐近特性的关系。最后,以正态和偏态正态(SN)非线性回归模型下分析的数据集为例说明了该方法。主要结论是,与相应的SN估计相比,来自重尾SSMN非线性模型的ML估计对外围观测值更具鲁棒性。
{"title":"Nonlinear regression models under skew scale mixtures of normal distributions","authors":"Clécio S. Ferreira , Víctor H. Lachos","doi":"10.1016/j.stamet.2016.08.004","DOIUrl":"https://doi.org/10.1016/j.stamet.2016.08.004","url":null,"abstract":"<div><p><span>Normal nonlinear regression models are applied in some areas of the sciences and engineering to explain or describe the phenomena under study. However, it is well known that several phenomena are not always represented by the normal model due to lack of symmetry or the presence of heavy- and light-tailed distributions related to the normal law in the data. This paper proposes an extension of nonlinear regression models using the skew-scale mixtures of normal (SSMN) distributions proposed by Ferreira et al. (2011). This class of models provides a useful generalization of the symmetrical nonlinear regression models since the random term distributions cover both asymmetric and heavy-tailed distributions, such as the skew-</span><span><math><mi>t</mi></math></span><span>-normal, skew-slash and skew-contaminated normal, among others. An expectation–maximization (EM) algorithm for maximum likelihood (ML) estimates is presented and the observed information matrix is derived analytically. Some simulation studies are presented to examine the performance of the proposed methods, with relation to robustness and asymptotic properties<span> of the ML estimates. Finally, an illustration of the method is presented considering a dataset previously analyzed under normal and skew-normal (SN) nonlinear regression models. The main conclusion is that the ML estimates from the heavy tails SSMN nonlinear models are more robust against outlying observations compared to the corresponding SN estimates.</span></span></p></div>","PeriodicalId":48877,"journal":{"name":"Statistical Methodology","volume":"33 ","pages":"Pages 131-146"},"PeriodicalIF":0.0,"publicationDate":"2016-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.stamet.2016.08.004","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136837486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}