Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2020.08.009
Victor Chernozhukov , Iván Fernández-Val , Martin Weidner
This paper provides a method to construct simultaneous confidence bands for quantile functions and quantile effects in nonlinear network and panel models with unobserved two-way effects, strictly exogenous covariates, and possibly discrete outcome variables. The method is based upon projection of simultaneous confidence bands for distribution functions constructed from fixed effects distribution regression estimators. These fixed effects estimators are debiased to deal with the incidental parameter problem. Under asymptotic sequences where both dimensions of the data set grow at the same rate, the confidence bands for the quantile functions and effects have correct joint coverage in large samples. An empirical application to gravity models of trade illustrates the applicability of the methods to network data.
{"title":"Network and panel quantile effects via distribution regression","authors":"Victor Chernozhukov , Iván Fernández-Val , Martin Weidner","doi":"10.1016/j.jeconom.2020.08.009","DOIUrl":"https://doi.org/10.1016/j.jeconom.2020.08.009","url":null,"abstract":"<div><p>This paper provides a method to construct simultaneous confidence bands for quantile functions and quantile effects in nonlinear network and panel models with unobserved two-way effects, strictly exogenous covariates, and possibly discrete outcome variables. The method is based upon projection of simultaneous confidence bands for distribution functions constructed from fixed effects distribution regression estimators. These fixed effects estimators are debiased to deal with the incidental parameter problem. Under asymptotic sequences where both dimensions of the data set grow at the same rate, the confidence bands for the quantile functions and effects have correct joint coverage in large samples. An empirical application to gravity models of trade illustrates the applicability of the methods to network data.</p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105009"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0304407620303390/pdfft?md5=95e8d9cc3d7ec76bead0f99a33c2f2f7&pid=1-s2.0-S0304407620303390-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140145344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2022.06.011
Bryan S. Graham , Fengshi Niu , James L. Powell
We study nonparametric estimation of density functions for undirected dyadic random variables (i.e., random variables defined for all unordered pairs of agents/nodes in a weighted network of order ). These random variables satisfy a local dependence property: any random variables in the network that share one or two indices may be dependent, while those sharing no indices in common are independent. In this setting, we show that density functions may be estimated by an application of the kernel estimation method of Rosenblatt (1956) and Parzen (1962). We suggest an estimate of their asymptotic variances inspired by a combination of (i) Newey’s (1994) method of variance estimation for kernel estimators in the “monadic” setting and (ii) a variance estimator for the (estimated) density of a simple network first suggested by Holland and Leinhardt (1976). More unusual are the rates of convergence and asymptotic (normal) distributions of our dyadic density estimates. Specifically, we show that they converge at the same rate as the (unconditional) dyadic sample mean: the square root of the number, , of nodes. This differs from the results for nonparametric estimation of densities and regression functions for monadic data, which generally have a slower rate of convergence than their corresponding sample mean.
我们研究的是对无向二元随机变量(即为阶数为 N 的加权网络中所有 n≡defN2 无序代理/节点对定义的随机变量)密度函数的非参数估计。这些随机变量满足局部依赖特性:网络中任何共享一个或两个索引的随机变量都可能是依赖的,而那些不共享索引的随机变量则是独立的。在这种情况下,我们可以应用 Rosenblatt(1956 年)和 Parzen(1962 年)的核估计方法来估计密度函数。我们提出了一种对其渐近方差的估计方法,其灵感来自于 (i) Newey(1994 年)在 "一元 "设置中对核估计器进行方差估计的方法和 (ii) Holland 和 Leinhardt(1976 年)首次提出的简单网络(估计)密度的方差估计器。更特别的是我们的二元密度估计的收敛率和渐近(正态)分布。具体来说,我们证明它们的收敛速度与(无条件的)二元样本平均值相同:即节点数 N 的平方根。这不同于对一元数据的密度和回归函数进行非参数估计的结果,后者的收敛速度通常慢于相应的样本平均值。
{"title":"Kernel density estimation for undirected dyadic data","authors":"Bryan S. Graham , Fengshi Niu , James L. Powell","doi":"10.1016/j.jeconom.2022.06.011","DOIUrl":"https://doi.org/10.1016/j.jeconom.2022.06.011","url":null,"abstract":"<div><p><span>We study nonparametric estimation of density functions for undirected dyadic random variables (i.e., random variables defined for all </span><span><math><mrow><mi>n</mi><mover><mrow><mo>≡</mo></mrow><mrow><mi>d</mi><mi>e</mi><mi>f</mi></mrow></mover><mfenced><mfrac><mrow><mi>N</mi></mrow><mrow><mn>2</mn></mrow></mfrac></mfenced></mrow></math></span><span> unordered pairs of agents/nodes in a weighted network of order </span><span><math><mi>N</mi></math></span><span><span><span>). These random variables satisfy a local dependence property: any random variables in the network that share one or two indices may be dependent, while those sharing no indices in common are independent. In this setting, we show that density functions may be estimated by an application of the kernel estimation method of </span>Rosenblatt<span> (1956) and Parzen (1962). We suggest an estimate of their asymptotic variances<span> inspired by a combination of (i) Newey’s (1994) method of variance estimation for kernel estimators in the “monadic” setting and (ii) a </span></span></span>variance estimator<span> for the (estimated) density of a simple network first suggested by Holland and Leinhardt (1976). More unusual are the rates of convergence and asymptotic (normal) distributions of our dyadic density estimates. Specifically, we show that they converge at the same rate as the (unconditional) dyadic sample mean: the square root of the number, </span></span><span><math><mi>N</mi></math></span><span>, of nodes. This differs from the results for nonparametric estimation of densities and regression functions for monadic data, which generally have a slower rate of convergence than their corresponding sample mean.</span></p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105336"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140145459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2021.01.003
Joshua D. Higbee , James B. McDonald
Several families of statistical distributions have been used to model financial data. The four-parameter generalized beta of the second kind (GB2) and five-parameter skewed generalized t (SGT) have been fit to return and log-return data, respectively. We introduce the skewed generalized log-t (SGLT) distribution and note that the GB2 and SGLT share such distributions as the asymmetric log-Laplace (ALL), log-Laplace (LL), and log-normal (LN). We then compare the relative performance of the GB2 and SGLT in modeling the distribution of daily, weekly, and monthly stock return data. We find that the GB2 and SGLT perform similarly and that the three-parameter log-t (LT) distribution is quite robust.
{"title":"A comparison of the GB2 and skewed generalized log-t distributions with an application in finance","authors":"Joshua D. Higbee , James B. McDonald","doi":"10.1016/j.jeconom.2021.01.003","DOIUrl":"10.1016/j.jeconom.2021.01.003","url":null,"abstract":"<div><p>Several families of statistical distributions have been used to model financial data. The four-parameter generalized beta of the second kind (GB2) and five-parameter skewed generalized t (SGT) have been fit to return and log-return data, respectively. We introduce the skewed generalized log-t (SGLT) distribution and note that the GB2 and SGLT share such distributions as the asymmetric log-Laplace (ALL), log-Laplace (LL), and log-normal (LN). We then compare the relative performance of the GB2 and SGLT in modeling the distribution of daily, weekly, and monthly stock return data. We find that the GB2 and SGLT perform similarly and that the three-parameter log-t (LT) distribution is quite robust.</p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105064"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44576544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2023.02.005
Alberto Abadie , Jiaying Gu , Shu Shen
We propose a simple data-driven procedure that exploits heterogeneity in the first-stage correlation between an instrument and an endogenous variable to improve the asymptotic mean squared error (MSE) of instrumental variable estimators. We show that the resulting gains in asymptotic MSE can be quite large in settings where there is substantial heterogeneity in the first-stage parameters. We also show that a naive procedure used in some applied work, which consists of selecting the composition of the sample based on the value of the first-stage -statistic, may cause substantial over-rejection of a null hypothesis on a second-stage parameter. We apply the methods to study (1) the return to schooling using the minimum school leaving age as the exogenous instrument and (2) the effect of local economic conditions on voter turnout using energy supply shocks as the source of identification.
我们提出了一个简单的数据驱动程序,利用工具和内生变量之间第一阶段相关性的异质性来改善工具变量估计值的渐近均方误差(MSE)。我们的研究表明,在第一阶段参数存在大量异质性的情况下,渐近平均误差的增益可能相当大。我们还表明,一些应用研究中使用的天真程序,即根据第一阶段 t 统计量的值选择样本的组成,可能会导致对第二阶段参数的无效假设产生严重的过度拒绝。我们运用这些方法研究了(1)以最低离校年龄为外生工具的就学回报率,以及(2)以能源供应冲击为识别源的地方经济条件对投票率的影响。
{"title":"Instrumental variable estimation with first-stage heterogeneity","authors":"Alberto Abadie , Jiaying Gu , Shu Shen","doi":"10.1016/j.jeconom.2023.02.005","DOIUrl":"10.1016/j.jeconom.2023.02.005","url":null,"abstract":"<div><p><span><span>We propose a simple data-driven procedure that exploits heterogeneity in the first-stage correlation between an instrument and an endogenous variable<span> to improve the asymptotic mean squared error (MSE) of </span></span>instrumental variable estimators. We show that the resulting gains in asymptotic MSE can be quite large in settings where there is substantial heterogeneity in the first-stage parameters. We also show that a naive procedure used in some applied work, which consists of selecting the composition of the sample based on the value of the first-stage </span><span><math><mi>t</mi></math></span><span><span>-statistic, may cause substantial over-rejection of a null hypothesis on a second-stage parameter. We apply the methods to study (1) the return to schooling using the minimum school leaving age as the exogenous instrument and (2) the effect of local economic conditions on </span>voter turnout using energy supply shocks as the source of identification.</span></p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105425"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42651450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2021.01.006
Matias D. Cattaneo , Michael Jansson , Xinwei Ma
This paper investigates the large sample properties of local regression distribution estimators, which include a class of boundary adaptive density estimators as a prime example. First, we establish a pointwise Gaussian large sample distributional approximation in a unified way, allowing for both boundary and interior evaluation points simultaneously. Using this result, we study the asymptotic efficiency of the estimators, and show that a carefully crafted minimum distance implementation based on “redundant” regressors can lead to efficiency gains. Second, we establish uniform linearizations and strong approximations for the estimators, and employ these results to construct valid confidence bands. Third, we develop extensions to weighted distributions with estimated weights and to local estimation. Finally, we illustrate our methods with two applications in program evaluation: counterfactual density testing, and IV specification and heterogeneity density analysis. Companion software packages in Stata and R are available.
本文研究了局部回归分布估计器的大样本特性,其中一类边界自适应密度估计器就是最好的例子。首先,我们以统一的方式建立了点式高斯大样本分布近似,同时允许边界和内部评估点。利用这一结果,我们研究了估计器的渐进效率,并表明基于 "冗余 "回归因子的精心设计的最小距离实现可以提高效率。其次,我们建立了估计器的统一线性化和强近似,并利用这些结果构建了有效的置信区间。第三,我们对带有估计权重的加权分布和局部 L2 估计进行了扩展。最后,我们用项目评估中的两个应用来说明我们的方法:反事实密度检验以及 IV 规范和异质性密度分析。可提供 Stata 和 R 的配套软件包。
{"title":"Local regression distribution estimators","authors":"Matias D. Cattaneo , Michael Jansson , Xinwei Ma","doi":"10.1016/j.jeconom.2021.01.006","DOIUrl":"10.1016/j.jeconom.2021.01.006","url":null,"abstract":"<div><p><span><span><span>This paper investigates the large sample properties of local regression distribution estimators, which include a class of boundary adaptive density estimators as a prime example. First, we establish a pointwise<span> Gaussian large sample distributional approximation in a unified way, allowing for both boundary and interior evaluation points simultaneously. Using this result, we study the </span></span>asymptotic efficiency of the estimators, and show that a carefully crafted minimum distance implementation based on “redundant” </span>regressors can lead to efficiency gains. Second, we establish uniform linearizations and strong approximations for the estimators, and employ these results to construct valid confidence bands. Third, we develop extensions to weighted distributions with estimated weights and to local </span><span><math><msup><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> estimation. Finally, we illustrate our methods with two applications in program evaluation: counterfactual density testing, and IV specification and heterogeneity density analysis. Companion software packages in <span>Stata</span> and <span>R</span> are available.</p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105074"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48826557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2020.09.013
Susan Athey , Guido W. Imbens , Jonas Metzger , Evan Munro
When researchers develop new econometric methods it is common practice to compare the performance of the new methods to those of existing methods in Monte Carlo studies. The credibility of such Monte Carlo studies is often limited because of the discretion the researcher has in choosing the Monte Carlo designs reported. To improve the credibility we propose using a class of generative models that has recently been developed in the machine learning literature, termed Generative Adversarial Networks (GANs) which can be used to systematically generate artificial data that closely mimics existing datasets. Thus, in combination with existing real data sets, GANs can be used to limit the degrees of freedom in Monte Carlo study designs for the researcher, making any comparisons more convincing. In addition, if an applied researcher is concerned with the performance of a particular statistical method on a specific data set (beyond its theoretical properties in large samples), she can use such GANs to assess the performance of the proposed method, e.g. the coverage rate of confidence intervals or the bias of the estimator, using simulated data which closely resembles the exact setting of interest. To illustrate these methods we apply Wasserstein GANs (WGANs) to the estimation of average treatment effects. In this example, we find that there is not a single estimator that outperforms the others in all three settings, so researchers should tailor their analytic approach to a given setting, systematic simulation studies can be helpful for selecting among competing methods in this situation, and the generated data closely resemble the actual data.
{"title":"Using Wasserstein Generative Adversarial Networks for the design of Monte Carlo simulations","authors":"Susan Athey , Guido W. Imbens , Jonas Metzger , Evan Munro","doi":"10.1016/j.jeconom.2020.09.013","DOIUrl":"https://doi.org/10.1016/j.jeconom.2020.09.013","url":null,"abstract":"<div><p><span><span>When researchers develop new econometric methods it is common practice to compare the performance of the new methods to those of existing methods in </span>Monte Carlo studies<span>. The credibility of such Monte Carlo studies is often limited because of the discretion the researcher has in choosing the Monte Carlo designs reported. To improve the credibility we propose using a class of generative models that has recently been developed in the machine learning literature, termed Generative Adversarial Networks (GANs) which can be used to systematically generate artificial data that closely mimics existing datasets. Thus, in combination with existing real data sets, GANs can be used to limit the degrees of freedom in Monte Carlo study designs for the researcher, making any comparisons more convincing. In addition, if an applied researcher is concerned with the performance of a particular statistical method on a specific data set (beyond its theoretical properties in large samples), she can use such GANs to assess the performance of the proposed method, </span></span><em>e.g.</em><span> the coverage rate of confidence intervals or the bias of the estimator, using simulated data<span> which closely resembles the exact setting of interest. To illustrate these methods we apply Wasserstein GANs (WGANs) to the estimation of average treatment effects. In this example, we find that </span></span><span><math><mrow><mo>(</mo><mi>i</mi><mo>)</mo></mrow></math></span> there is not a single estimator that outperforms the others in all three settings, so researchers should tailor their analytic approach to a given setting, <span><math><mrow><mo>(</mo><mi>i</mi><mi>i</mi><mo>)</mo></mrow></math></span> systematic simulation studies can be helpful for selecting among competing methods in this situation, and <span><math><mrow><mo>(</mo><mi>i</mi><mi>i</mi><mi>i</mi><mo>)</mo></mrow></math></span> the generated data closely resemble the actual data.</p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105076"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140145357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2021.07.016
Shakeeb Khan , Denis Nekipelov
This paper explores the uniformity of inference for parameters of interest in nonlinear econometric models with endogeneity. Here the notion of uniformity arises because the behavior of estimators of parameters of interest is shown to vary with where either they or nuisance parameters lie in the parameter space. As a result, inference becomes nonstandard in a fashion that is loosely analogous to inference complications found in the unit root and weak instruments literature, as well as the models recently studied in Andrews and Cheng (2012), Chen et al. (2014), Han and McCloskey (2019). Our main illustrative example is the standard sample selection model, where the parameter of interest is the intercept term as in Heckman (1990), Andrews and Schafgans (1998) and Lewbel (2007). We show here there is a discontinuity in the limiting distribution for an estimator of this parameter despite it being uniformly consistent. This discontinuity prevents standard inference procedures from being valid, and motivates the development of new methods, for which we establish asymptotic properties. Finite sample properties of the procedure are explored through a simulation study and an empirical illustration using the Mroz (1987) data set as in Newey, Powell, and Walker (1990).
{"title":"On uniform inference in nonlinear models with endogeneity","authors":"Shakeeb Khan , Denis Nekipelov","doi":"10.1016/j.jeconom.2021.07.016","DOIUrl":"10.1016/j.jeconom.2021.07.016","url":null,"abstract":"<div><p><span><span>This paper explores the uniformity of inference for parameters of interest in nonlinear econometric models with endogeneity. Here the notion of uniformity arises because the behavior of estimators of parameters of interest is shown to vary with where either they or </span>nuisance parameters lie in the parameter space. As a result, inference becomes nonstandard in a fashion that is loosely analogous to inference complications found in the unit root and weak instruments literature, as well as the models recently studied in Andrews and Cheng (2012), Chen et al. (2014), Han and McCloskey (2019). Our main illustrative example is the standard sample selection model, where the parameter of interest is the intercept term as in Heckman (1990), Andrews and Schafgans (1998) and Lewbel (2007). We show here there is a </span><em>discontinuity</em><span> in the limiting distribution for an estimator of this parameter despite it being uniformly consistent. This discontinuity prevents standard inference procedures from being valid, and motivates the development of new methods, for which we establish asymptotic properties. Finite sample properties of the procedure are explored through a simulation study and an empirical illustration using the Mroz (1987) data set as in Newey, Powell, and Walker (1990).</span></p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105261"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132163820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2021.03.007
Frank Windmeijer
This paper develops the links between overidentification tests, underidentification tests, score tests and the Cragg and Donald (1993, 1997) and Kleibergen and Paap (2006) rank tests in linear instrumental variable (IV) models. For the structural linear model , with the endogenous explanatory variables partitioned as , this general framework shows that standard underidentification tests are tests for overidentification in an auxiliary linear model, , estimated by IV estimation methods using the same instruments as for the original model. This simple structure makes it possible to establish valid robust underidentification tests for linear IV models where these have not been proposed or used before, like clustered dynamic panel data models estimated by GMM. The framework also applies to tests for the rank of general parameter matrices. Invariant rank tests are based on the LIML or continuously updated GMM estimators of both structural and first-stage parameters. This insight leads to the proposal of new two-step invariant asymptotically efficient GMM estimators, and a new iterated GMM estimator that, if it converges, converges to the continuously updated GMM estimator.
{"title":"Testing underidentification in linear models, with applications to dynamic panel and asset pricing models","authors":"Frank Windmeijer","doi":"10.1016/j.jeconom.2021.03.007","DOIUrl":"10.1016/j.jeconom.2021.03.007","url":null,"abstract":"<div><p><span>This paper develops the links between overidentification tests, underidentification tests, score tests and the Cragg and Donald (1993, 1997) and Kleibergen and Paap (2006) rank tests<span> in linear instrumental variable (IV) models. For the structural linear model </span></span><span><math><mrow><mi>y</mi><mo>=</mo><mi>X</mi><mi>β</mi><mo>+</mo><mi>u</mi></mrow></math></span><span>, with the endogenous explanatory variables partitioned as </span><span><math><mrow><mi>X</mi><mo>=</mo><mfenced><mrow><msub><mrow><mi>x</mi></mrow><mrow><mn>1</mn></mrow></msub><mspace></mspace><msub><mrow><mi>X</mi></mrow><mrow><mn>2</mn></mrow></msub></mrow></mfenced></mrow></math></span>, this general framework shows that standard underidentification tests are tests for overidentification in an auxiliary linear model, <span><math><mrow><msub><mrow><mi>x</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>=</mo><msub><mrow><mi>X</mi></mrow><mrow><mn>2</mn></mrow></msub><mi>δ</mi><mo>+</mo><mi>ɛ</mi></mrow></math></span><span>, estimated by IV estimation methods using the same instruments as for the original model. This simple structure makes it possible to establish valid robust underidentification tests for linear IV models where these have not been proposed or used before, like clustered dynamic panel data models estimated by GMM. The framework also applies to tests for the rank of general parameter matrices. Invariant rank tests are based on the LIML or continuously updated GMM estimators of both structural and first-stage parameters. This insight leads to the proposal of new two-step invariant asymptotically efficient GMM estimators, and a new iterated GMM estimator that, if it converges, converges to the continuously updated GMM estimator.</span></p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105104"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43824339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2020.09.012
Xavier D’Haultfœuille , Stefan Hoderlein , Yuya Sasaki
The control function approach which employs an instrumental variable excluded from the outcome equation is a very common solution to deal with the problem of endogeneity in nonseparable models. Exclusion restrictions, however, are frequently controversial. We first argue that, in a nonparametric triangular structure typical of the control function literature, one can actually test this exclusion restriction provided the instrument satisfies a local irrelevance condition. Second, we investigate identification without such exclusion restrictions, i.e., if the “instrument” that is independent of the unobservables in the outcome equation also directly affects the outcome variable. In particular, we show that identification of average causal effects can be achieved in the two most common special cases of the general nonseparable model: linear random coefficients models and single index models.
{"title":"Testing and relaxing the exclusion restriction in the control function approach","authors":"Xavier D’Haultfœuille , Stefan Hoderlein , Yuya Sasaki","doi":"10.1016/j.jeconom.2020.09.012","DOIUrl":"10.1016/j.jeconom.2020.09.012","url":null,"abstract":"<div><p><span>The control function approach which employs an </span>instrumental variable<span> excluded from the outcome equation is a very common solution to deal with the problem of endogeneity in nonseparable models. Exclusion restrictions, however, are frequently controversial. We first argue that, in a nonparametric triangular structure typical of the control function literature, one can actually test this exclusion restriction provided the instrument satisfies a local irrelevance condition. Second, we investigate identification without such exclusion restrictions, i.e., if the “instrument” that is independent of the unobservables in the outcome equation also directly affects the outcome variable. In particular, we show that identification of average causal effects can be achieved in the two most common special cases of the general nonseparable model: linear random coefficients models and single index models.</span></p></div>","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105075"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46289137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-01DOI: 10.1016/j.jeconom.2022.12.013
Thomas Kolokotrones , James H. Stock , Christopher D. Walker
<div><p><span><span>Newey–West (1987) standard errors are the dominant standard errors used for heteroskedasticity and autocorrelation robust (HAR) inference in </span>time series<span> regression. The Newey–West estimator uses the Bartlett kernel, which is a first-order kernel, meaning that its characteristic exponent, </span></span><span><math><mi>q</mi></math></span>, is equal to 1, where <span><math><mi>q</mi></math></span> is defined as the largest value of <span><math><mi>r</mi></math></span> for which the quantity <span><math><mrow><msup><mrow><mi>k</mi></mrow><mrow><mrow><mo>[</mo><mi>r</mi><mo>]</mo></mrow></mrow></msup><mrow><mo>(</mo><mn>0</mn><mo>)</mo></mrow><mo>=</mo><msub><mrow><mo>lim</mo></mrow><mrow><mi>t</mi><mo>→</mo><mn>0</mn></mrow></msub><msup><mrow><mrow><mo>|</mo><mi>t</mi><mo>|</mo></mrow></mrow><mrow><mo>−</mo><mi>r</mi></mrow></msup><mrow><mo>(</mo><mi>k</mi><mrow><mo>(</mo><mn>0</mn><mo>)</mo></mrow><mo>−</mo><mi>k</mi><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow><mo>)</mo></mrow></mrow></math></span> is defined and finite. This raises the apparently uninvestigated question of whether the Bartlett kernel is optimal among first-order kernels. We demonstrate that, for <span><math><mrow><mi>q</mi><mo><</mo><mn>2</mn></mrow></math></span>, there is no optimal <span><math><mi>q</mi></math></span><span>th-order kernel for HAR testing in the Gaussian<span><span> location model or for minimizing the MSE in </span>spectral density estimation. In fact, for any </span></span><span><math><mrow><mi>q</mi><mo><</mo><mn>2</mn></mrow></math></span>, the space of <span><math><mi>q</mi></math></span>th-order positive-semidefinite kernels is not closed and, moreover, all continuous <span><math><mi>q</mi></math></span>th-order kernels can be decomposed into a weighted sum of <span><math><mi>q</mi></math></span>th and second-order kernels, which suggests that there is no meaningful notion of ‘pure’ <span><math><mi>q</mi></math></span>th-order kernels for <span><math><mrow><mi>q</mi><mo><</mo><mn>2</mn></mrow></math></span>. Nevertheless, it is possible to rank any given collection of <span><math><mi>q</mi></math></span>th-order kernels using the functional <span><math><mrow><msub><mrow><mi>I</mi></mrow><mrow><mi>q</mi></mrow></msub><mrow><mo>[</mo><mi>k</mi><mo>]</mo></mrow><mo>=</mo><msup><mrow><mfenced><mrow><msup><mrow><mi>k</mi></mrow><mrow><mrow><mo>[</mo><mi>q</mi><mo>]</mo></mrow></mrow></msup><mrow><mo>(</mo><mn>0</mn><mo>)</mo></mrow></mrow></mfenced></mrow><mrow><mn>1</mn><mo>/</mo><mi>q</mi></mrow></msup><mo>∫</mo><msup><mrow><mi>k</mi></mrow><mrow><mn>2</mn></mrow></msup><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow><mi>d</mi><mi>t</mi></mrow></math></span> with smaller values corresponding to better asymptotic performance. We examine the value of <span><math><mrow><msub><mrow><mi>I</mi></mrow><mrow><mi>q</mi></mrow></msub><mrow><mo>[</mo><mi>k</mi><mo>]</mo></mrow></mrow></math></span> for a wide variety of first-order es
{"title":"Is Newey–West optimal among first-order kernels?","authors":"Thomas Kolokotrones , James H. Stock , Christopher D. Walker","doi":"10.1016/j.jeconom.2022.12.013","DOIUrl":"10.1016/j.jeconom.2022.12.013","url":null,"abstract":"<div><p><span><span>Newey–West (1987) standard errors are the dominant standard errors used for heteroskedasticity and autocorrelation robust (HAR) inference in </span>time series<span> regression. The Newey–West estimator uses the Bartlett kernel, which is a first-order kernel, meaning that its characteristic exponent, </span></span><span><math><mi>q</mi></math></span>, is equal to 1, where <span><math><mi>q</mi></math></span> is defined as the largest value of <span><math><mi>r</mi></math></span> for which the quantity <span><math><mrow><msup><mrow><mi>k</mi></mrow><mrow><mrow><mo>[</mo><mi>r</mi><mo>]</mo></mrow></mrow></msup><mrow><mo>(</mo><mn>0</mn><mo>)</mo></mrow><mo>=</mo><msub><mrow><mo>lim</mo></mrow><mrow><mi>t</mi><mo>→</mo><mn>0</mn></mrow></msub><msup><mrow><mrow><mo>|</mo><mi>t</mi><mo>|</mo></mrow></mrow><mrow><mo>−</mo><mi>r</mi></mrow></msup><mrow><mo>(</mo><mi>k</mi><mrow><mo>(</mo><mn>0</mn><mo>)</mo></mrow><mo>−</mo><mi>k</mi><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow><mo>)</mo></mrow></mrow></math></span> is defined and finite. This raises the apparently uninvestigated question of whether the Bartlett kernel is optimal among first-order kernels. We demonstrate that, for <span><math><mrow><mi>q</mi><mo><</mo><mn>2</mn></mrow></math></span>, there is no optimal <span><math><mi>q</mi></math></span><span>th-order kernel for HAR testing in the Gaussian<span><span> location model or for minimizing the MSE in </span>spectral density estimation. In fact, for any </span></span><span><math><mrow><mi>q</mi><mo><</mo><mn>2</mn></mrow></math></span>, the space of <span><math><mi>q</mi></math></span>th-order positive-semidefinite kernels is not closed and, moreover, all continuous <span><math><mi>q</mi></math></span>th-order kernels can be decomposed into a weighted sum of <span><math><mi>q</mi></math></span>th and second-order kernels, which suggests that there is no meaningful notion of ‘pure’ <span><math><mi>q</mi></math></span>th-order kernels for <span><math><mrow><mi>q</mi><mo><</mo><mn>2</mn></mrow></math></span>. Nevertheless, it is possible to rank any given collection of <span><math><mi>q</mi></math></span>th-order kernels using the functional <span><math><mrow><msub><mrow><mi>I</mi></mrow><mrow><mi>q</mi></mrow></msub><mrow><mo>[</mo><mi>k</mi><mo>]</mo></mrow><mo>=</mo><msup><mrow><mfenced><mrow><msup><mrow><mi>k</mi></mrow><mrow><mrow><mo>[</mo><mi>q</mi><mo>]</mo></mrow></mrow></msup><mrow><mo>(</mo><mn>0</mn><mo>)</mo></mrow></mrow></mfenced></mrow><mrow><mn>1</mn><mo>/</mo><mi>q</mi></mrow></msup><mo>∫</mo><msup><mrow><mi>k</mi></mrow><mrow><mn>2</mn></mrow></msup><mrow><mo>(</mo><mi>t</mi><mo>)</mo></mrow><mi>d</mi><mi>t</mi></mrow></math></span> with smaller values corresponding to better asymptotic performance. We examine the value of <span><math><mrow><msub><mrow><mi>I</mi></mrow><mrow><mi>q</mi></mrow></msub><mrow><mo>[</mo><mi>k</mi><mo>]</mo></mrow></mrow></math></span> for a wide variety of first-order es","PeriodicalId":15629,"journal":{"name":"Journal of Econometrics","volume":"240 2","pages":"Article 105399"},"PeriodicalIF":6.3,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45646742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}