首页 > 最新文献

arXiv - ECON - Econometrics最新文献

英文 中文
Conditional nonparametric variable screening by neural factor regression 通过神经因子回归进行条件非参数变量筛选
Pub Date : 2024-08-20 DOI: arxiv-2408.10825
Jianqing FanPrinceton University, Weining WangUniversity of Groningen, Yue ZhaoUniversity of York
High-dimensional covariates often admit linear factor structure. Toeffectively screen correlated covariates in high-dimension, we propose aconditional variable screening test based on non-parametric regression usingneural networks due to their representation power. We ask the question whetherindividual covariates have additional contributions given the latent factors ormore generally a set of variables. Our test statistics are based on theestimated partial derivative of the regression function of the candidatevariable for screening and a observable proxy for the latent factors. Hence,our test reveals how much predictors contribute additionally to thenon-parametric regression after accounting for the latent factors. Ourderivative estimator is the convolution of a deep neural network regressionestimator and a smoothing kernel. We demonstrate that when the neural networksize diverges with the sample size, unlike estimating the regression functionitself, it is necessary to smooth the partial derivative of the neural networkestimator to recover the desired convergence rate for the derivative. Moreover,our screening test achieves asymptotic normality under the null after finelycentering our test statistics that makes the biases negligible, as well asconsistency for local alternatives under mild conditions. We demonstrate theperformance of our test in a simulation study and two real world applications.
高维协变量通常具有线性因子结构。为了有效筛选高维相关协变量,我们提出了一种基于非参数回归的条件变量筛选测试,利用神经网络的表征能力进行筛选。我们提出的问题是,在潜在因素或更广泛的变量集合中,单个协变量是否有额外的贡献。我们的检验统计基于筛选候选变量的回归函数的估计偏导数和潜在因素的可观测替代变量。因此,我们的检验揭示了在考虑潜在因素后,预测因子对非参数回归的额外贡献程度。我们的衍生估计器是深度神经网络回归估计器和平滑核的卷积。我们证明,当神经网络大小随样本大小发散时,与估计回归函数本身不同,有必要平滑神经网络估计器的偏导数,以恢复所需的导数收敛速率。此外,我们的筛选检验在对检验统计量进行精细中心化处理后,实现了空值下的渐近正态性,使偏差可以忽略不计,并在温和条件下实现了局部替代的一致性。我们在一项模拟研究和两个实际应用中证明了我们的测试性能。
{"title":"Conditional nonparametric variable screening by neural factor regression","authors":"Jianqing FanPrinceton University, Weining WangUniversity of Groningen, Yue ZhaoUniversity of York","doi":"arxiv-2408.10825","DOIUrl":"https://doi.org/arxiv-2408.10825","url":null,"abstract":"High-dimensional covariates often admit linear factor structure. To\u0000effectively screen correlated covariates in high-dimension, we propose a\u0000conditional variable screening test based on non-parametric regression using\u0000neural networks due to their representation power. We ask the question whether\u0000individual covariates have additional contributions given the latent factors or\u0000more generally a set of variables. Our test statistics are based on the\u0000estimated partial derivative of the regression function of the candidate\u0000variable for screening and a observable proxy for the latent factors. Hence,\u0000our test reveals how much predictors contribute additionally to the\u0000non-parametric regression after accounting for the latent factors. Our\u0000derivative estimator is the convolution of a deep neural network regression\u0000estimator and a smoothing kernel. We demonstrate that when the neural network\u0000size diverges with the sample size, unlike estimating the regression function\u0000itself, it is necessary to smooth the partial derivative of the neural network\u0000estimator to recover the desired convergence rate for the derivative. Moreover,\u0000our screening test achieves asymptotic normality under the null after finely\u0000centering our test statistics that makes the biases negligible, as well as\u0000consistency for local alternatives under mild conditions. We demonstrate the\u0000performance of our test in a simulation study and two real world applications.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"1587 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gradient Wild Bootstrap for Instrumental Variable Quantile Regressions with Weak and Few Clusters 用于弱聚类和少聚类工具变量量值回归的梯度野生引导法
Pub Date : 2024-08-20 DOI: arxiv-2408.10686
Wenjie Wang, Yichong Zhang
We study the gradient wild bootstrap-based inference for instrumentalvariable quantile regressions in the framework of a small number of largeclusters in which the number of clusters is viewed as fixed, and the number ofobservations for each cluster diverges to infinity. For the Wald inference, weshow that our wild bootstrap Wald test, with or without studentization usingthe cluster-robust covariance estimator (CRVE), controls size asymptotically upto a small error as long as the parameter of endogenous variable is stronglyidentified in at least one of the clusters. We further show that the wildbootstrap Wald test with CRVE studentization is more powerful for distant localalternatives than that without. Last, we develop a wild bootstrapAnderson-Rubin (AR) test for the weak-identification-robust inference. We showit controls size asymptotically up to a small error, even under weak or partialidentification for all clusters. We illustrate the good finite-sampleperformance of the new inference methods using simulations and provide anempirical application to a well-known dataset about US local labor markets.
我们研究了基于梯度野生引导的工具变量量化回归推断,该推断是在少数大型聚类的框架下进行的,其中聚类的数量被视为固定的,而每个聚类的观察数会发散到无穷大。对于 Wald 推理,我们表明,只要内生变量的参数至少在其中一个聚类中被强识别,我们的野生自举 Wald 检验,无论是否使用聚类稳健协方差估计器(CRVE)进行学生化,都能在很小的误差范围内渐进地控制规模。我们进一步证明,与不使用 CRVE 的情况相比,使用 CRVE 的野生自回归 Wald 检验对遥远的本地替代变量更有效。最后,我们开发了一种用于弱识别稳健推断的野生自举安德森-鲁宾(AR)检验。我们证明,即使在所有聚类的弱识别或部分识别情况下,它也能控制大小,直至误差很小。我们通过模拟说明了新推断方法良好的有限样本性能,并提供了一个关于美国地方劳动力市场的著名数据集的经验应用。
{"title":"Gradient Wild Bootstrap for Instrumental Variable Quantile Regressions with Weak and Few Clusters","authors":"Wenjie Wang, Yichong Zhang","doi":"arxiv-2408.10686","DOIUrl":"https://doi.org/arxiv-2408.10686","url":null,"abstract":"We study the gradient wild bootstrap-based inference for instrumental\u0000variable quantile regressions in the framework of a small number of large\u0000clusters in which the number of clusters is viewed as fixed, and the number of\u0000observations for each cluster diverges to infinity. For the Wald inference, we\u0000show that our wild bootstrap Wald test, with or without studentization using\u0000the cluster-robust covariance estimator (CRVE), controls size asymptotically up\u0000to a small error as long as the parameter of endogenous variable is strongly\u0000identified in at least one of the clusters. We further show that the wild\u0000bootstrap Wald test with CRVE studentization is more powerful for distant local\u0000alternatives than that without. Last, we develop a wild bootstrap\u0000Anderson-Rubin (AR) test for the weak-identification-robust inference. We show\u0000it controls size asymptotically up to a small error, even under weak or partial\u0000identification for all clusters. We illustrate the good finite-sample\u0000performance of the new inference methods using simulations and provide an\u0000empirical application to a well-known dataset about US local labor markets.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"32 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
kendallknight: Efficient Implementation of Kendall's Correlation Coefficient Computation kendallknight:肯德尔相关系数计算的高效实现
Pub Date : 2024-08-19 DOI: arxiv-2408.09618
Mauricio Vargas Sepúlveda
The kendallknight package introduces an efficient implementation of Kendall'scorrelation coefficient computation, significantly improving the processingtime for large datasets without sacrificing accuracy. The kendallknightpackage, following Knight (1966) and posterior literature, reduces thecomputational complexity resulting in drastic reductions in computation time,transforming operations that would take minutes or hours into milliseconds orminutes, while maintaining precision and correctly handling edge cases anderrors. The package is particularly advantageous in econometric and statisticalcontexts where rapid and accurate calculation of Kendall's correlationcoefficient is desirable. Benchmarks demonstrate substantial performance gainsover the base R implementation, especially for large datasets.
kendallknight 软件包引入了肯德尔相关系数计算的高效实现方法,在不牺牲准确性的前提下,显著改善了大型数据集的处理时间。kendallknight 软件包遵循 Knight (1966) 和后继文献,降低了计算复杂度,从而大幅减少了计算时间,将需要几分钟或几小时的操作转化为几毫秒或几分钟,同时保持了精度,并正确处理了边缘情况和错误。该软件包在计量经济学和统计领域尤其具有优势,因为这些领域需要快速、准确地计算肯德尔相关系数。基准测试表明,与基本 R 实现相比,该软件包的性能大幅提升,尤其是在大型数据集上。
{"title":"kendallknight: Efficient Implementation of Kendall's Correlation Coefficient Computation","authors":"Mauricio Vargas Sepúlveda","doi":"arxiv-2408.09618","DOIUrl":"https://doi.org/arxiv-2408.09618","url":null,"abstract":"The kendallknight package introduces an efficient implementation of Kendall's\u0000correlation coefficient computation, significantly improving the processing\u0000time for large datasets without sacrificing accuracy. The kendallknight\u0000package, following Knight (1966) and posterior literature, reduces the\u0000computational complexity resulting in drastic reductions in computation time,\u0000transforming operations that would take minutes or hours into milliseconds or\u0000minutes, while maintaining precision and correctly handling edge cases and\u0000errors. The package is particularly advantageous in econometric and statistical\u0000contexts where rapid and accurate calculation of Kendall's correlation\u0000coefficient is desirable. Benchmarks demonstrate substantial performance gains\u0000over the base R implementation, especially for large datasets.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"157 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184215","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anytime-Valid Inference for Double/Debiased Machine Learning of Causal Parameters 因果参数双重/偏差机器学习的随时有效推理
Pub Date : 2024-08-18 DOI: arxiv-2408.09598
Abhinandan Dalal, Patrick Blöbaum, Shiva Kasiviswanathan, Aaditya Ramdas
Double (debiased) machine learning (DML) has seen widespread use in recentyears for learning causal/structural parameters, in part due to its flexibilityand adaptability to high-dimensional nuisance functions as well as its abilityto avoid bias from regularization or overfitting. However, the classicdouble-debiased framework is only valid asymptotically for a predeterminedsample size, thus lacking the flexibility of collecting more data if sharperinference is needed, or stopping data collection early if useful inferences canbe made earlier than expected. This can be of particular concern in large scaleexperimental studies with huge financial costs or human lives at stake, as wellas in observational studies where the length of confidence of intervals do notshrink to zero even with increasing sample size due to partial identifiabilityof a structural parameter. In this paper, we present time-uniform counterpartsto the asymptotic DML results, enabling valid inference and confidenceintervals for structural parameters to be constructed at any arbitrary(possibly data-dependent) stopping time. We provide conditions which are onlyslightly stronger than the standard DML conditions, but offer the strongerguarantee for anytime-valid inference. This facilitates the transformation ofany existing DML method to provide anytime-valid guarantees with minimalmodifications, making it highly adaptable and easy to use. We illustrate ourprocedure using two instances: a) local average treatment effect in onlineexperiments with non-compliance, and b) partial identification of averagetreatment effect in observational studies with potential unmeasuredconfounding.
近年来,双(去偏)机器学习(DML)在因果/结构参数学习中得到了广泛应用,部分原因在于它的灵活性和对高维骚扰函数的适应性,以及避免正则化或过度拟合产生偏差的能力。然而,经典的双偏差框架仅在预定样本量下渐进有效,因此缺乏灵活性,无法在需要更清晰推断时收集更多数据,或在比预期更早做出有用推断时提前停止数据收集。这一点在涉及巨大经济成本或人命的大型实验研究中,以及在观察性研究中尤为重要,因为在观察性研究中,由于结构参数的部分可识别性,即使样本量增加,置信区间的长度也不会缩减为零。在本文中,我们提出了与渐近 DML 结果相对应的时间均匀性,从而使结构参数的有效推断和置信区间可以在任意(可能取决于数据)停止时间构建。我们提供的条件只比标准 DML 条件稍强,但却为任何时间的有效推断提供了更强的保证。这有助于将任何现有的 DML 方法转化为提供随时有效保证的方法,而只需做极少的修改,从而使其具有很强的适应性和易用性。我们用两个例子来说明我们的方法:a) 在线实验中的局部平均治疗效果,有不遵守的情况;b) 观察性研究中的平均治疗效果的部分识别,有潜在的非测量混淆。
{"title":"Anytime-Valid Inference for Double/Debiased Machine Learning of Causal Parameters","authors":"Abhinandan Dalal, Patrick Blöbaum, Shiva Kasiviswanathan, Aaditya Ramdas","doi":"arxiv-2408.09598","DOIUrl":"https://doi.org/arxiv-2408.09598","url":null,"abstract":"Double (debiased) machine learning (DML) has seen widespread use in recent\u0000years for learning causal/structural parameters, in part due to its flexibility\u0000and adaptability to high-dimensional nuisance functions as well as its ability\u0000to avoid bias from regularization or overfitting. However, the classic\u0000double-debiased framework is only valid asymptotically for a predetermined\u0000sample size, thus lacking the flexibility of collecting more data if sharper\u0000inference is needed, or stopping data collection early if useful inferences can\u0000be made earlier than expected. This can be of particular concern in large scale\u0000experimental studies with huge financial costs or human lives at stake, as well\u0000as in observational studies where the length of confidence of intervals do not\u0000shrink to zero even with increasing sample size due to partial identifiability\u0000of a structural parameter. In this paper, we present time-uniform counterparts\u0000to the asymptotic DML results, enabling valid inference and confidence\u0000intervals for structural parameters to be constructed at any arbitrary\u0000(possibly data-dependent) stopping time. We provide conditions which are only\u0000slightly stronger than the standard DML conditions, but offer the stronger\u0000guarantee for anytime-valid inference. This facilitates the transformation of\u0000any existing DML method to provide anytime-valid guarantees with minimal\u0000modifications, making it highly adaptable and easy to use. We illustrate our\u0000procedure using two instances: a) local average treatment effect in online\u0000experiments with non-compliance, and b) partial identification of average\u0000treatment effect in observational studies with potential unmeasured\u0000confounding.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"24 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Experimental Design For Causal Inference Through An Optimization Lens 通过优化视角进行因果推理的实验设计
Pub Date : 2024-08-18 DOI: arxiv-2408.09607
Jinglong Zhao
The study of experimental design offers tremendous benefits for answeringcausal questions across a wide range of applications, including agriculturalexperiments, clinical trials, industrial experiments, social experiments, anddigital experiments. Although valuable in such applications, the costs ofexperiments often drive experimenters to seek more efficient designs. Recently,experimenters have started to examine such efficiency questions from anoptimization perspective, as experimental design problems are fundamentallydecision-making problems. This perspective offers a lot of flexibility inleveraging various existing optimization tools to study experimental designproblems. This manuscript thus aims to examine the foundations of experimentaldesign problems in the context of causal inference as viewed through anoptimization lens.
对实验设计的研究为回答农业实验、临床试验、工业实验、社会实验和数字实验等广泛应用中的因果问题提供了巨大的益处。尽管在这些应用中很有价值,但实验成本往往促使实验人员寻求更高效的设计。最近,实验人员开始从优化的角度研究此类效率问题,因为实验设计问题从根本上说是决策问题。这一视角为利用各种现有优化工具研究实验设计问题提供了很大的灵活性。因此,本手稿旨在通过优化视角,研究因果推断背景下的实验设计问题的基础。
{"title":"Experimental Design For Causal Inference Through An Optimization Lens","authors":"Jinglong Zhao","doi":"arxiv-2408.09607","DOIUrl":"https://doi.org/arxiv-2408.09607","url":null,"abstract":"The study of experimental design offers tremendous benefits for answering\u0000causal questions across a wide range of applications, including agricultural\u0000experiments, clinical trials, industrial experiments, social experiments, and\u0000digital experiments. Although valuable in such applications, the costs of\u0000experiments often drive experimenters to seek more efficient designs. Recently,\u0000experimenters have started to examine such efficiency questions from an\u0000optimization perspective, as experimental design problems are fundamentally\u0000decision-making problems. This perspective offers a lot of flexibility in\u0000leveraging various existing optimization tools to study experimental design\u0000problems. This manuscript thus aims to examine the foundations of experimental\u0000design problems in the context of causal inference as viewed through an\u0000optimization lens.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning for the Estimation of Heterogeneous Parameters in Discrete Choice Models 离散选择模型中异质参数估计的深度学习
Pub Date : 2024-08-18 DOI: arxiv-2408.09560
Stephan Hetzenecker, Maximilian Osterhaus
This paper studies the finite sample performance of the flexible estimationapproach of Farrell, Liang, and Misra (2021a), who propose to use deep learningfor the estimation of heterogeneous parameters in economic models, in thecontext of discrete choice models. The approach combines the structure imposedby economic models with the flexibility of deep learning, which assures theinterpretebility of results on the one hand, and allows estimating flexiblefunctional forms of observed heterogeneity on the other hand. For inferenceafter the estimation with deep learning, Farrell et al. (2021a) derive aninfluence function that can be applied to many quantities of interest. Weconduct a series of Monte Carlo experiments that investigate the impact ofregularization on the proposed estimation and inference procedure in thecontext of discrete choice models. The results show that the deep learningapproach generally leads to precise estimates of the true average parametersand that regular robust standard errors lead to invalid inference results,showing the need for the influence function approach for inference. Withoutregularization, the influence function approach can lead to substantial biasand large estimated standard errors caused by extreme outliers. Regularizationreduces this property and stabilizes the estimation procedure, but at theexpense of inducing an additional bias. The bias in combination with decreasingvariance associated with increasing regularization leads to the construction ofinvalid inferential statements in our experiments. Repeated sample splitting,unlike regularization, stabilizes the estimation approach without introducingan additional bias, thereby allowing for the construction of valid inferentialstatements.
本文研究了 Farrell、Liang 和 Misra(2021a)的灵活估计方法的有限样本性能,他们提出在离散选择模型的背景下,使用深度学习来估计经济模型中的异质性参数。该方法将经济模型的结构与深度学习的灵活性相结合,一方面保证了结果的可解释性,另一方面允许对观察到的异质性进行灵活的功能形式估计。为了在深度学习估计之后进行推理,Farrell 等人(2021a)推导出了一种可应用于许多相关量的影响函数。我们进行了一系列蒙特卡罗实验,研究了离散选择模型背景下规范化对所提出的估计和推理过程的影响。结果表明,深度学习方法通常能精确估计真实的平均参数,而常规稳健标准误差会导致无效的推断结果,这表明推断需要使用影响函数方法。如果不进行正则化,影响函数方法可能会因极端离群值而导致严重偏差和较大的估计标准误差。正则化可以减少这种特性并稳定估计过程,但代价是引起额外的偏差。在我们的实验中,偏差与随着正则化程度增加而递减的方差相结合,导致了无效推断语句的产生。重复样本拆分与正则化不同,它能稳定估计方法,而不会引入额外的偏差,从而允许构建有效的推断陈述。
{"title":"Deep Learning for the Estimation of Heterogeneous Parameters in Discrete Choice Models","authors":"Stephan Hetzenecker, Maximilian Osterhaus","doi":"arxiv-2408.09560","DOIUrl":"https://doi.org/arxiv-2408.09560","url":null,"abstract":"This paper studies the finite sample performance of the flexible estimation\u0000approach of Farrell, Liang, and Misra (2021a), who propose to use deep learning\u0000for the estimation of heterogeneous parameters in economic models, in the\u0000context of discrete choice models. The approach combines the structure imposed\u0000by economic models with the flexibility of deep learning, which assures the\u0000interpretebility of results on the one hand, and allows estimating flexible\u0000functional forms of observed heterogeneity on the other hand. For inference\u0000after the estimation with deep learning, Farrell et al. (2021a) derive an\u0000influence function that can be applied to many quantities of interest. We\u0000conduct a series of Monte Carlo experiments that investigate the impact of\u0000regularization on the proposed estimation and inference procedure in the\u0000context of discrete choice models. The results show that the deep learning\u0000approach generally leads to precise estimates of the true average parameters\u0000and that regular robust standard errors lead to invalid inference results,\u0000showing the need for the influence function approach for inference. Without\u0000regularization, the influence function approach can lead to substantial bias\u0000and large estimated standard errors caused by extreme outliers. Regularization\u0000reduces this property and stabilizes the estimation procedure, but at the\u0000expense of inducing an additional bias. The bias in combination with decreasing\u0000variance associated with increasing regularization leads to the construction of\u0000invalid inferential statements in our experiments. Repeated sample splitting,\u0000unlike regularization, stabilizes the estimation approach without introducing\u0000an additional bias, thereby allowing for the construction of valid inferential\u0000statements.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"26 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Externally Valid Selection of Experimental Sites via the k-Median Problem 通过 k-Median 问题选择外部有效的实验点
Pub Date : 2024-08-17 DOI: arxiv-2408.09187
José Luis Montiel Olea, Brenda Prallon, Chen Qiu, Jörg Stoye, Yiwei Sun
We present a decision-theoretic justification for viewing the question of howto best choose where to experiment in order to optimize external validity as ak-median (clustering) problem, a popular problem in computer science andoperations research. We present conditions under which minimizing theworst-case, welfare-based regret among all nonrandom schemes that select ksites to experiment is approximately equal - and sometimes exactly equal - tofinding the k most central vectors of baseline site-level covariates. Thek-median problem can be formulated as a linear integer program. Two empiricalapplications illustrate the theoretical and computational benefits of thesuggested procedure.
我们提出了一种决策理论依据,将如何最佳选择实验地点以优化外部有效性的问题视为一个中值(聚类)问题,这是计算机科学和运营研究中的一个流行问题。我们提出了一些条件,在这些条件下,在所有选择 ks 个地点进行实验的非随机方案中,最小化最坏情况下基于福利的遗憾,近似等于(有时甚至完全等于)找到基线地点级协变量的 k 个最中心向量。k-中值问题可以表述为一个线性整数程序。两个经验应用说明了所建议程序的理论和计算优势。
{"title":"Externally Valid Selection of Experimental Sites via the k-Median Problem","authors":"José Luis Montiel Olea, Brenda Prallon, Chen Qiu, Jörg Stoye, Yiwei Sun","doi":"arxiv-2408.09187","DOIUrl":"https://doi.org/arxiv-2408.09187","url":null,"abstract":"We present a decision-theoretic justification for viewing the question of how\u0000to best choose where to experiment in order to optimize external validity as a\u0000k-median (clustering) problem, a popular problem in computer science and\u0000operations research. We present conditions under which minimizing the\u0000worst-case, welfare-based regret among all nonrandom schemes that select k\u0000sites to experiment is approximately equal - and sometimes exactly equal - to\u0000finding the k most central vectors of baseline site-level covariates. The\u0000k-median problem can be formulated as a linear integer program. Two empirical\u0000applications illustrate the theoretical and computational benefits of the\u0000suggested procedure.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142223851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Method of Moments Estimation for Affine Stochastic Volatility Models 仿随机波动率模型的矩估计法
Pub Date : 2024-08-17 DOI: arxiv-2408.09185
Yan-Feng Wu, Xiangyu Yang, Jian-Qiang Hu
We develop moment estimators for the parameters of affine stochasticvolatility models. We first address the challenge of calculating moments forthe models by introducing a recursive equation for deriving closed-formexpressions for moments of any order. Consequently, we propose our momentestimators. We then establish a central limit theorem for our estimators andderive the explicit formulas for the asymptotic covariance matrix. Finally, weprovide numerical results to validate our method.
我们开发了仿射随机波动模型参数的矩估计器。我们首先引入了一个递归方程,用于推导任意阶矩的闭式公式,从而解决了计算模型矩的难题。因此,我们提出了矩估计器。然后,我们建立了估计器的中心极限定理,并推导出渐近协方差矩阵的显式。最后,我们提供数值结果来验证我们的方法。
{"title":"Method of Moments Estimation for Affine Stochastic Volatility Models","authors":"Yan-Feng Wu, Xiangyu Yang, Jian-Qiang Hu","doi":"arxiv-2408.09185","DOIUrl":"https://doi.org/arxiv-2408.09185","url":null,"abstract":"We develop moment estimators for the parameters of affine stochastic\u0000volatility models. We first address the challenge of calculating moments for\u0000the models by introducing a recursive equation for deriving closed-form\u0000expressions for moments of any order. Consequently, we propose our moment\u0000estimators. We then establish a central limit theorem for our estimators and\u0000derive the explicit formulas for the asymptotic covariance matrix. Finally, we\u0000provide numerical results to validate our method.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"11 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Counterfactual and Synthetic Control Method: Causal Inference with Instrumented Principal Component Analysis 反事实和合成控制法:利用工具主成分分析进行因果推断
Pub Date : 2024-08-17 DOI: arxiv-2408.09271
Cong Wang
The fundamental problem of causal inference lies in the absence ofcounterfactuals. Traditional methodologies impute the missing counterfactualsimplicitly or explicitly based on untestable or overly stringent assumptions.Synthetic control method (SCM) utilizes a weighted average of control units toimpute the missing counterfactual for the treated unit. Although SCM relaxessome strict assumptions, it still requires the treated unit to be inside theconvex hull formed by the controls, avoiding extrapolation. In recent advances,researchers have modeled the entire data generating process (DGP) to explicitlyimpute the missing counterfactual. This paper expands the interactive fixedeffect (IFE) model by instrumenting covariates into factor loadings, addingadditional robustness. This methodology offers multiple benefits: firstly, itincorporates the strengths of previous SCM approaches, such as the relaxationof the untestable parallel trends assumption (PTA). Secondly, it does notrequire the targeted outcomes to be inside the convex hull formed by thecontrols. Thirdly, it eliminates the need for correct model specificationrequired by the IFE model. Finally, it inherits the ability of principalcomponent analysis (PCA) to effectively handle high-dimensional data andenhances the value extracted from numerous covariates.
因果推断的根本问题在于缺乏反事实。传统方法基于无法检验或过于严格的假设,或隐式或显式地计算缺失的反事实。虽然合成控制法放宽了一些严格的假设,但它仍然要求被处理单位位于控制单元形成的凸壳内,从而避免了外推。最近,研究人员对整个数据生成过程(DGP)进行了建模,以明确计算缺失的反事实。本文通过将协变量工具化为因子载荷,扩展了交互固定效应(IFE)模型,增加了额外的稳健性。这种方法有多个优点:首先,它吸收了以往单因素模型方法的优点,如放宽了无法检验的平行趋势假设(PTA)。其次,它不要求目标结果位于控制所形成的凸壳内。第三,它不需要 IFE 模型所要求的正确模型规范。最后,它继承了主成分分析(PCA)有效处理高维数据的能力,并提高了从众多协变量中提取的价值。
{"title":"Counterfactual and Synthetic Control Method: Causal Inference with Instrumented Principal Component Analysis","authors":"Cong Wang","doi":"arxiv-2408.09271","DOIUrl":"https://doi.org/arxiv-2408.09271","url":null,"abstract":"The fundamental problem of causal inference lies in the absence of\u0000counterfactuals. Traditional methodologies impute the missing counterfactuals\u0000implicitly or explicitly based on untestable or overly stringent assumptions.\u0000Synthetic control method (SCM) utilizes a weighted average of control units to\u0000impute the missing counterfactual for the treated unit. Although SCM relaxes\u0000some strict assumptions, it still requires the treated unit to be inside the\u0000convex hull formed by the controls, avoiding extrapolation. In recent advances,\u0000researchers have modeled the entire data generating process (DGP) to explicitly\u0000impute the missing counterfactual. This paper expands the interactive fixed\u0000effect (IFE) model by instrumenting covariates into factor loadings, adding\u0000additional robustness. This methodology offers multiple benefits: firstly, it\u0000incorporates the strengths of previous SCM approaches, such as the relaxation\u0000of the untestable parallel trends assumption (PTA). Secondly, it does not\u0000require the targeted outcomes to be inside the convex hull formed by the\u0000controls. Thirdly, it eliminates the need for correct model specification\u0000required by the IFE model. Finally, it inherits the ability of principal\u0000component analysis (PCA) to effectively handle high-dimensional data and\u0000enhances the value extracted from numerous covariates.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"141 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Revisiting the Many Instruments Problem using Random Matrix Theory 利用随机矩阵理论重新审视众多工具问题
Pub Date : 2024-08-16 DOI: arxiv-2408.08580
Helmut Farbmacher, Rebecca Groh, Michael Mühlegger, Gabriel Vollert
We use recent results from the theory of random matrices to improveinstrumental variables estimation with many instruments. In settings where thefirst-stage parameters are dense, we show that Ridge lowers the implicit priceof a bias adjustment. This comes along with improved (finite-sample) propertiesin the second stage regression. Our theoretical results nest existing resultson bias approximation and bias adjustment. Moreover, it extends them tosettings with more instruments than observations.
我们利用随机矩阵理论的最新成果来改进多工具的工具变量估计。在第一阶段参数密集的情况下,我们发现 Ridge 降低了偏差调整的隐含代价。同时,这也改善了第二阶段回归的(有限样本)特性。我们的理论结果与现有的偏差逼近和偏差调整结果相吻合。此外,它还将这些结果扩展到了工具多于观测值的情况下。
{"title":"Revisiting the Many Instruments Problem using Random Matrix Theory","authors":"Helmut Farbmacher, Rebecca Groh, Michael Mühlegger, Gabriel Vollert","doi":"arxiv-2408.08580","DOIUrl":"https://doi.org/arxiv-2408.08580","url":null,"abstract":"We use recent results from the theory of random matrices to improve\u0000instrumental variables estimation with many instruments. In settings where the\u0000first-stage parameters are dense, we show that Ridge lowers the implicit price\u0000of a bias adjustment. This comes along with improved (finite-sample) properties\u0000in the second stage regression. Our theoretical results nest existing results\u0000on bias approximation and bias adjustment. Moreover, it extends them to\u0000settings with more instruments than observations.","PeriodicalId":501293,"journal":{"name":"arXiv - ECON - Econometrics","volume":"23 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142184219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - ECON - Econometrics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1