首页 > 最新文献

Journal of the Japanese Society of Computational Statistics最新文献

英文 中文
Asymptotic cumulants of some information criteria 一些信息准则的渐近累积量
Pub Date : 2016-12-20 DOI: 10.5183/JJSCS.1512001_225
H. Ogasawara
Asymptotic cumulants of the Akaike and Takeuchi information criteria are given under possible model misspecification up to the fourth order with the higher-order asymptotic variances, where two versions of the latter information criterion are defined using observed and estimated expected information matrices. The asymptotic cumulants are provided before and after studentization using the parameter estimators by the weighted-score method, which include the maximum likelihood and Bayes modal estimators as special cases. Higher-order bias corrections of the criteria are derived using log-likelihood derivatives, which yields simple results for cases under canonical parametrization in the exponential family. It is shown that in these cases the Jeffreys prior gives the vanishing higher-order bias of the Akaike information criterion. The results are illustrated by three examples. Simulations for model selection in regression and interval estimation are also given.
Akaike和Takeuchi信息准则的渐近累积量在可能的模型错配下达到四阶,具有高阶渐近方差,其中后者信息准则的两个版本使用观察和估计的期望信息矩阵来定义。用加权分数法给出了参数估计量在学生化前后的渐近累积量,其中最大似然估计量和贝叶斯模态估计量是特例。使用对数似然导数推导了准则的高阶偏差修正,对于指数族中典型参数化的情况,得到了简单的结果。结果表明,在这些情况下,Jeffreys先验给出了Akaike信息准则的高阶偏差消失。通过三个算例说明了结果。给出了回归和区间估计中模型选择的仿真。
{"title":"Asymptotic cumulants of some information criteria","authors":"H. Ogasawara","doi":"10.5183/JJSCS.1512001_225","DOIUrl":"https://doi.org/10.5183/JJSCS.1512001_225","url":null,"abstract":"Asymptotic cumulants of the Akaike and Takeuchi information criteria are given under possible model misspecification up to the fourth order with the higher-order asymptotic variances, where two versions of the latter information criterion are defined using observed and estimated expected information matrices. The asymptotic cumulants are provided before and after studentization using the parameter estimators by the weighted-score method, which include the maximum likelihood and Bayes modal estimators as special cases. Higher-order bias corrections of the criteria are derived using log-likelihood derivatives, which yields simple results for cases under canonical parametrization in the exponential family. It is shown that in these cases the Jeffreys prior gives the vanishing higher-order bias of the Akaike information criterion. The results are illustrated by three examples. Simulations for model selection in regression and interval estimation are also given.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123169438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
ADHERENTLY PENALIZED LINEAR DISCRIMINANT ANALYSIS 坚持惩罚线性判别分析
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1412001_219
H. Hino, Jun Fujiki
A problem of supervised learning in which the data consist of p features and n observations is considered. Each observation is assumed to belong to either one of the two classes. Linear discriminant analysis (LDA) has been widely used for both classification and dimensionality reduction in this setting. However, when the dimensionality p is high and the observations are scarce, LDA does not offer a satisfactory result for classification. Witten & Tibshirani (2011) proposed the penalized LDA based on the Fisher’s discriminant problem with sparsity penalization. In this paper, an elastic-net type penalization is considered for LDA, and the corresponding optimization problem is efficiently solved.
考虑了一个由p个特征和n个观测值组成的数据的监督学习问题。假设每个观测值属于这两类中的任意一类。线性判别分析(LDA)已被广泛用于这种情况下的分类和降维。然而,当维数p较高且观测值较少时,LDA的分类效果并不理想。Witten & Tibshirani(2011)在Fisher判别问题的基础上提出了带有稀疏性惩罚的惩罚LDA。本文考虑了LDA的弹性网型惩罚,有效地解决了相应的优化问题。
{"title":"ADHERENTLY PENALIZED LINEAR DISCRIMINANT ANALYSIS","authors":"H. Hino, Jun Fujiki","doi":"10.5183/JJSCS.1412001_219","DOIUrl":"https://doi.org/10.5183/JJSCS.1412001_219","url":null,"abstract":"A problem of supervised learning in which the data consist of p features and n observations is considered. Each observation is assumed to belong to either one of the two classes. Linear discriminant analysis (LDA) has been widely used for both classification and dimensionality reduction in this setting. However, when the dimensionality p is high and the observations are scarce, LDA does not offer a satisfactory result for classification. Witten & Tibshirani (2011) proposed the penalized LDA based on the Fisher’s discriminant problem with sparsity penalization. In this paper, an elastic-net type penalization is considered for LDA, and the corresponding optimization problem is efficiently solved.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"128 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132256875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
POWER CALCULATIONS IN CLINICAL TRIALS WITH COMPLEX CLINICAL OBJECTIVES 具有复杂临床目的的临床试验中的功率计算
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1411001_213
A. Dmitrienko, G. Paux, T. Brechenmacher
Over the past decade, a variety of powerful multiple testing procedures have been developed for the analysis of clinical trials with multiple clinical objectives based, for example, on several endpoints, dose-placebo comparisons and patient subgroups. Sample size and power calculations in these complex settings are not straightforward and, in general, simulation-based methods are used. In this paper, we provide an overview of power evaluation approaches in the context of clinical trials with multiple objectives and illustrate the key principles using case studies commonly seen in the development of new therapies.
在过去的十年中,已经开发了各种强大的多重测试程序,用于分析基于多个临床目标的临床试验,例如,基于多个终点,剂量-安慰剂比较和患者亚组。在这些复杂的设置中,样本大小和功率计算并不简单,通常使用基于模拟的方法。在本文中,我们概述了在多目标临床试验背景下的功效评估方法,并通过新疗法开发中常见的案例研究说明了关键原则。
{"title":"POWER CALCULATIONS IN CLINICAL TRIALS WITH COMPLEX CLINICAL OBJECTIVES","authors":"A. Dmitrienko, G. Paux, T. Brechenmacher","doi":"10.5183/JJSCS.1411001_213","DOIUrl":"https://doi.org/10.5183/JJSCS.1411001_213","url":null,"abstract":"Over the past decade, a variety of powerful multiple testing procedures have been developed for the analysis of clinical trials with multiple clinical objectives based, for example, on several endpoints, dose-placebo comparisons and patient subgroups. Sample size and power calculations in these complex settings are not straightforward and, in general, simulation-based methods are used. In this paper, we provide an overview of power evaluation approaches in the context of clinical trials with multiple objectives and illustrate the key principles using case studies commonly seen in the development of new therapies.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131260606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
SPARSE PREDICTIVE MODELING FOR BANK TELEMARKETING SUCCESS USING SMOOTH-THRESHOLD ESTIMATING EQUATIONS 基于平滑阈值估计方程的银行电话营销成功稀疏预测模型
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1502003_217
Y. Kawasaki, Masao Ueki
In this paper, we attempt to build and evaluate several predictive models to predict success of telemarketing calls for selling bank long-term deposits using a publicly available set of data from a Portuguese retail bank collected from 2008 to 2013 (Moro et al., 2014, Decision Support Systems). The data include multiple predictor variables, either numeric or categorical, related with bank client, product and social-economic attributes. Dealing with a categorical predictor variable as multiple dummy variables increases model dimensionality, and redundancy in model parameterization must be of practical concern. This motivates us to assess prediction performance with more parsimonious modeling. We apply contemporary variable selection methods with penalization including lasso, elastic net, smoothly-clipped absolute deviation, minimum concave penalty as well as the smooth-threshold estimating equation. In addition to variable selection, the smooth-threshold estimating equation can achieve automatic grouping of predictor variables, which is an alternative sparse modeling to perform variable selection and could be suited to a certain problem, e.g., dummy variables created from categorical predictor variables. Predictive power of each modeling approach is assessed by repeating cross-validation experiments or sample splitting, one for training and another for testing.
在本文中,我们试图建立和评估几个预测模型,以预测电话营销电话销售银行长期存款的成功,使用2008年至2013年收集的葡萄牙零售银行公开可用的数据集(Moro等人,2014年,决策支持系统)。这些数据包括与银行客户、产品和社会经济属性相关的多个预测变量,有的是数字变量,有的是分类变量。将一个分类预测变量作为多个虚拟变量来处理,增加了模型的维数,模型参数化中的冗余必须得到实际的关注。这促使我们用更简洁的建模来评估预测性能。我们采用现代的变量选择方法,包括套索法、弹性网法、平滑夹持绝对偏差法、最小凹惩罚法以及平滑阈值估计方程。除了变量选择之外,平滑阈值估计方程还可以实现预测变量的自动分组,这是进行变量选择的另一种稀疏建模方法,可以适用于某些问题,例如由分类预测变量创建的虚拟变量。每种建模方法的预测能力通过重复交叉验证实验或样本分割来评估,一个用于训练,另一个用于测试。
{"title":"SPARSE PREDICTIVE MODELING FOR BANK TELEMARKETING SUCCESS USING SMOOTH-THRESHOLD ESTIMATING EQUATIONS","authors":"Y. Kawasaki, Masao Ueki","doi":"10.5183/JJSCS.1502003_217","DOIUrl":"https://doi.org/10.5183/JJSCS.1502003_217","url":null,"abstract":"In this paper, we attempt to build and evaluate several predictive models to predict success of telemarketing calls for selling bank long-term deposits using a publicly available set of data from a Portuguese retail bank collected from 2008 to 2013 (Moro et al., 2014, Decision Support Systems). The data include multiple predictor variables, either numeric or categorical, related with bank client, product and social-economic attributes. Dealing with a categorical predictor variable as multiple dummy variables increases model dimensionality, and redundancy in model parameterization must be of practical concern. This motivates us to assess prediction performance with more parsimonious modeling. We apply contemporary variable selection methods with penalization including lasso, elastic net, smoothly-clipped absolute deviation, minimum concave penalty as well as the smooth-threshold estimating equation. In addition to variable selection, the smooth-threshold estimating equation can achieve automatic grouping of predictor variables, which is an alternative sparse modeling to perform variable selection and could be suited to a certain problem, e.g., dummy variables created from categorical predictor variables. Predictive power of each modeling approach is assessed by repeating cross-validation experiments or sample splitting, one for training and another for testing.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122086869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
ESTIMATING SCALE-FREE NETWORKS VIA THE EXPONENTIATION OF MINIMAX CONCAVE PENALTY 利用极大极小凹惩罚指数估计无标度网络
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1503001_215
K. Hirose, Y. Ogura, Hidetoshi Shimodaira
We consider the problem of sparse estimation of undirected graphical models via the L1 regularization. The ordinary lasso encourages the sparsity on all edges equally likely, so that all nodes tend to have small degrees. On the other hand, many real-world networks are often scale-free, where some nodes have a large number of edges. In such cases, a penalty that induces structured sparsity, such as a log penalty, performs better than the ordinary lasso. In practical situations, however, it is difficult to determine an optimal penalty among the ordinary lasso, log penalty, or somewhere in between. In this paper, we introduce a new class of penalty that is based on the exponentiation of the minimax concave penalty. The proposed penalty includes both the lasso and the log penalty, and the gap between these two penalties is bridged by a tuning parameter. We apply cross-validation to select an appropriate value of the tuning parameter. Monte Carlo simulations are conducted to investigate the performance of our proposed procedure. The numerical result shows that the proposed method can perform better than the existing log penalty and the ordinary lasso.
利用L1正则化方法研究无向图模型的稀疏估计问题。普通套索鼓励所有边的稀疏性等可能,因此所有节点都倾向于具有较小的度。另一方面,许多现实世界的网络通常是无标度的,其中一些节点具有大量的边。在这种情况下,诱导结构化稀疏性的惩罚(例如log惩罚)比普通套索执行得更好。然而,在实际情况中,很难在普通套索惩罚、对数惩罚或两者之间确定最优惩罚。本文引入了一类新的基于极大极小凹惩罚的幂次惩罚。建议的惩罚包括套索惩罚和日志惩罚,这两种惩罚之间的差距通过调优参数来弥补。我们应用交叉验证来选择一个合适的调优参数值。通过蒙特卡罗模拟来研究我们提出的程序的性能。数值结果表明,该方法比现有的对数惩罚和普通套索具有更好的性能。
{"title":"ESTIMATING SCALE-FREE NETWORKS VIA THE EXPONENTIATION OF MINIMAX CONCAVE PENALTY","authors":"K. Hirose, Y. Ogura, Hidetoshi Shimodaira","doi":"10.5183/JJSCS.1503001_215","DOIUrl":"https://doi.org/10.5183/JJSCS.1503001_215","url":null,"abstract":"We consider the problem of sparse estimation of undirected graphical models via the L1 regularization. The ordinary lasso encourages the sparsity on all edges equally likely, so that all nodes tend to have small degrees. On the other hand, many real-world networks are often scale-free, where some nodes have a large number of edges. In such cases, a penalty that induces structured sparsity, such as a log penalty, performs better than the ordinary lasso. In practical situations, however, it is difficult to determine an optimal penalty among the ordinary lasso, log penalty, or somewhere in between. In this paper, we introduce a new class of penalty that is based on the exponentiation of the minimax concave penalty. The proposed penalty includes both the lasso and the log penalty, and the gap between these two penalties is bridged by a tuning parameter. We apply cross-validation to select an appropriate value of the tuning parameter. Monte Carlo simulations are conducted to investigate the performance of our proposed procedure. The numerical result shows that the proposed method can perform better than the existing log penalty and the ordinary lasso.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"161 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133107462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
STOCHASTIC ALTERNATING DIRECTION METHOD OF MULTIPLIERS FOR STRUCTURED REGULARIZATION 结构正则化乘法器的随机交替方向方法
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1502004_218
Taiji Suzuki
In this paper, we present stochastic optimization variants of the alternating direction method of multipliers (ADMM). ADMM is a useful method to solve a regularized risk minimization problem where the regularization term is complicated and not easily dealt with in an ordinary manner. For example, structured regularization is one of the typical applications of such regularization in which ADMM is effective. It includes group lasso regularization, low rank tensor regularization, and fused lasso regularization. Since ADMM is a general method and has wide applications, it is intensively studied and refined these days. However, ADMM is not suited to optimization problems with huge data. To resolve this problem, online stochastic optimization variants and a batch stochastic optimization variant of ADMM are presented. All the presented methods can be easily implemented and have wide applications. Moreover, the theoretical guarantees of the methods are given.
本文提出了乘法器交替方向法(ADMM)的随机优化变体。对于正则化项复杂且不易用常规方法处理的正则化风险最小化问题,ADMM是一种有用的方法。例如,结构化正则化是这种正则化的典型应用之一,其中ADMM是有效的。它包括群拉索正则化、低秩张量正则化和融合拉索正则化。由于ADMM是一种通用的方法,具有广泛的应用,因此近年来对其进行了深入的研究和完善。然而,ADMM并不适用于大数据的优化问题。为了解决这一问题,提出了在线随机优化变量和批量随机优化变量。所提出的方法易于实现,具有广泛的应用前景。并给出了方法的理论保证。
{"title":"STOCHASTIC ALTERNATING DIRECTION METHOD OF MULTIPLIERS FOR STRUCTURED REGULARIZATION","authors":"Taiji Suzuki","doi":"10.5183/JJSCS.1502004_218","DOIUrl":"https://doi.org/10.5183/JJSCS.1502004_218","url":null,"abstract":"In this paper, we present stochastic optimization variants of the alternating direction method of multipliers (ADMM). ADMM is a useful method to solve a regularized risk minimization problem where the regularization term is complicated and not easily dealt with in an ordinary manner. For example, structured regularization is one of the typical applications of such regularization in which ADMM is effective. It includes group lasso regularization, low rank tensor regularization, and fused lasso regularization. Since ADMM is a general method and has wide applications, it is intensively studied and refined these days. However, ADMM is not suited to optimization problems with huge data. To resolve this problem, online stochastic optimization variants and a batch stochastic optimization variant of ADMM are presented. All the presented methods can be easily implemented and have wide applications. Moreover, the theoretical guarantees of the methods are given.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"137 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124668964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PREDICTIVE MODEL SELECTION CRITERIA FOR BAYESIAN LASSO REGRESSION 贝叶斯套索回归的预测模型选择准则
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1501001_220
Shuichi Kawano, Ibuki Hoshina, Kaito Shimamura, S. Konishi
We consider the Bayesian lasso for regression, which can be interpreted as an L 1 norm regularization based on a Bayesian approach when the Laplace or double-exponential prior distribution is placed on the regression coefficients. A crucial issue is an appropriate choice of the values of hyperparameters included in the prior distributions, which essentially control the sparsity in the estimated model. To choose the values of tuning parameters, we introduce a model selection criterion for evaluating a Bayesian predictive distribution for the Bayesian lasso. Numerical results are presented to illustrate the properties of our sparse Bayesian modeling procedure.
我们考虑回归的贝叶斯套索,当回归系数上有拉普拉斯或双指数先验分布时,它可以被解释为基于贝叶斯方法的L 1范数正则化。一个关键问题是先验分布中包含的超参数值的适当选择,这从本质上控制了估计模型的稀疏性。为了选择调整参数的值,我们引入了一个模型选择准则来评估贝叶斯套索的贝叶斯预测分布。数值结果说明了稀疏贝叶斯建模方法的性质。
{"title":"PREDICTIVE MODEL SELECTION CRITERIA FOR BAYESIAN LASSO REGRESSION","authors":"Shuichi Kawano, Ibuki Hoshina, Kaito Shimamura, S. Konishi","doi":"10.5183/JJSCS.1501001_220","DOIUrl":"https://doi.org/10.5183/JJSCS.1501001_220","url":null,"abstract":"We consider the Bayesian lasso for regression, which can be interpreted as an L 1 norm regularization based on a Bayesian approach when the Laplace or double-exponential prior distribution is placed on the regression coefficients. A crucial issue is an appropriate choice of the values of hyperparameters included in the prior distributions, which essentially control the sparsity in the estimated model. To choose the values of tuning parameters, we introduce a model selection criterion for evaluating a Bayesian predictive distribution for the Bayesian lasso. Numerical results are presented to illustrate the properties of our sparse Bayesian modeling procedure.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114583956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
EDITORIAL: RECENT ADVANCES IN SPARSE STATISTICAL MODELING 编辑:稀疏统计建模的最新进展
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1510002_225
K. Hirose
The first term L(β) is a loss function and the second term λ ∑p j=1 |βj | is a penalty term. Here λ (λ > 0) is a tuning parameter which controls the sparsity and the model fitting. Because the penalty term consists of the sum of absolute values of the parameter, we can carry out the sparse estimation, that is, some of the elements of β are estimated by exactly zeros. It is well-known that we cannot often obtain the analytical solutions of the minimization problem (1), because the penalty term λ ∑p j=1 |βj | is indifferentiable when βj = 0 (j = 1, . . . , p). Therefore, it is important to develop efficient computational algorithms. This special issue includes six interesting papers related to sparse estimation. These papers cover a wide variety of topics, such as statistical modeling, computation, theoretical analysis, and applications. In particular, all of the papers deal with the issue of statistical computation. Kawasaki and Ueki (the first paper of this issue) apply smooth-threshold estimating equations (STEE, Ueki, 2009) to telemarketing success data collected from a Portuguese retail bank. In STEE, the penalty term consists of a quadratic form ∑p j=1 wjβ 2 j instead of ∑p j=1 |βj |, where wj (j = 1, . . . , p) are positive values allowed to be ∞, so that we do not need to implement a computational algorithm that is used in the L1 regularization. Kawano, Hoshina, Shimamura and Konishi (the second paper) propose a model selection criterion for choosing tuning parameters in the Bayesian lasso (Park and Casella, 2008). They use an efficient sparse estimation algorithm in the Bayesian lasso, referred to as the sparse algorithm. Matsui (the third paper) considers the problem of bi-level selection, which allows the selection of groups of variables and individuals simultaneously. The parameter estimation procedure is based on the coordinate descent algorithm, which is known as a remarkably fast algorithm (Friedman et al., 2010). Suzuki (the fourth paper) focuses attention on the alternating direction method of multipliers algorithm (ADMM algorithm, Boyd et al., 2011), which is applicable to various complex penalties such as the overlapping group lasso (Jacob et al., 2009). He reviews a stochastic version of the ADMM algorithm that allows the online learning. Hino and Fujiki (the fifth paper) propose a penalized linear discriminant analysis that adheres to the normal discriminant model. They apply the Majorize-Minimization algorithm (MM algorithm, Hunter and Lange 2004), which is often used to replace a non-convex optimization problem with a reweighted convex optimization
第一项L(β)是一个损失函数,第二项λ∑p j=1 |βj |是一个惩罚项。这里λ (λ > 0)是一个调节参数,它控制稀疏性和模型拟合。由于惩罚项由参数的绝对值和组成,我们可以进行稀疏估计,即β的一些元素被精确地估计为零。众所周知,我们不能经常得到最小化问题(1)的解析解,因为当βj = 0 (j =1,…)时,惩罚项λ∑p j=1 |βj |是不可微的。, p)。因此,开发高效的计算算法非常重要。本期特刊包括六篇与稀疏估计相关的有趣论文。这些论文涵盖了各种各样的主题,如统计建模、计算、理论分析和应用。特别地,所有的论文都涉及统计计算问题。Kawasaki和Ueki(本期的第一篇论文)将平滑阈值估计方程(STEE, Ueki, 2009)应用于从葡萄牙零售银行收集的电话营销成功数据。在STEE中,惩罚项由二次型∑p j=1 wjβ 2j代替∑p j=1 |βj |,其中wj (j =1,…, p)是允许为∞的正值,因此我们不需要实现L1正则化中使用的计算算法。Kawano, Hoshina, Shimamura和Konishi(第二篇论文)提出了在贝叶斯套索中选择调谐参数的模型选择标准(Park和Casella, 2008)。他们在贝叶斯套索中使用了一种高效的稀疏估计算法,称为稀疏算法。Matsui(第三篇论文)考虑了双水平选择问题,允许同时选择变量组和个体组。参数估计过程基于坐标下降算法,这是一种非常快速的算法(Friedman et al., 2010)。Suzuki(第四篇论文)重点研究了乘法器算法(ADMM算法,Boyd et al., 2011)的交替方向法,该算法适用于重叠组lasso (Jacob et al., 2009)等各种复杂处罚。他回顾了ADMM算法的随机版本,该算法允许在线学习。Hino和Fujiki(第五篇论文)提出了一种符合正常判别模型的惩罚线性判别分析。他们应用了最大化最小化算法(MM算法,Hunter and Lange 2004),该算法通常用于用重新加权的凸优化取代非凸优化问题
{"title":"EDITORIAL: RECENT ADVANCES IN SPARSE STATISTICAL MODELING","authors":"K. Hirose","doi":"10.5183/JJSCS.1510002_225","DOIUrl":"https://doi.org/10.5183/JJSCS.1510002_225","url":null,"abstract":"The first term L(β) is a loss function and the second term λ ∑p j=1 |βj | is a penalty term. Here λ (λ > 0) is a tuning parameter which controls the sparsity and the model fitting. Because the penalty term consists of the sum of absolute values of the parameter, we can carry out the sparse estimation, that is, some of the elements of β are estimated by exactly zeros. It is well-known that we cannot often obtain the analytical solutions of the minimization problem (1), because the penalty term λ ∑p j=1 |βj | is indifferentiable when βj = 0 (j = 1, . . . , p). Therefore, it is important to develop efficient computational algorithms. This special issue includes six interesting papers related to sparse estimation. These papers cover a wide variety of topics, such as statistical modeling, computation, theoretical analysis, and applications. In particular, all of the papers deal with the issue of statistical computation. Kawasaki and Ueki (the first paper of this issue) apply smooth-threshold estimating equations (STEE, Ueki, 2009) to telemarketing success data collected from a Portuguese retail bank. In STEE, the penalty term consists of a quadratic form ∑p j=1 wjβ 2 j instead of ∑p j=1 |βj |, where wj (j = 1, . . . , p) are positive values allowed to be ∞, so that we do not need to implement a computational algorithm that is used in the L1 regularization. Kawano, Hoshina, Shimamura and Konishi (the second paper) propose a model selection criterion for choosing tuning parameters in the Bayesian lasso (Park and Casella, 2008). They use an efficient sparse estimation algorithm in the Bayesian lasso, referred to as the sparse algorithm. Matsui (the third paper) considers the problem of bi-level selection, which allows the selection of groups of variables and individuals simultaneously. The parameter estimation procedure is based on the coordinate descent algorithm, which is known as a remarkably fast algorithm (Friedman et al., 2010). Suzuki (the fourth paper) focuses attention on the alternating direction method of multipliers algorithm (ADMM algorithm, Boyd et al., 2011), which is applicable to various complex penalties such as the overlapping group lasso (Jacob et al., 2009). He reviews a stochastic version of the ADMM algorithm that allows the online learning. Hino and Fujiki (the fifth paper) propose a penalized linear discriminant analysis that adheres to the normal discriminant model. They apply the Majorize-Minimization algorithm (MM algorithm, Hunter and Lange 2004), which is often used to replace a non-convex optimization problem with a reweighted convex optimization","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117292893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SPARSE REGULARIZATION FOR BI-LEVEL VARIABLE SELECTION 双水平变量选择的稀疏正则化
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1502001_216
H. Matsui
Sparse regularization provides solutions in which some parameters are exactly zero and therefore they can be used for selecting variables in regression models and so on. The lasso is proposed as a method for selecting individual variables for regression models. On the other hand, the group lasso selects groups of variables rather than individuals and therefore it has been used in various fields of applications. More recently, penalties that select variables at both the group and individual levels has been considered. They are so called bi-level selection. In this paper we focus on some penalties that aim for bi-level selection. We overview these penalties and estimation algorithms, and then compare the effectiveness of these penalties from the viewpoint of accuracy of prediction and selection of variables and groups through simulation studies.
稀疏正则化提供了一些参数完全为零的解决方案,因此它们可以用于选择回归模型中的变量等。套索是一种选择回归模型中单个变量的方法。另一方面,组套索选择一组变量而不是单个变量,因此它已被用于各种应用领域。最近,在群体和个人层面上选择变量的惩罚已经被考虑。它们被称为双级选择。在本文中,我们重点讨论了一些针对双级选择的惩罚。我们概述了这些惩罚和估计算法,然后通过仿真研究从预测的准确性和变量和组的选择的角度比较了这些惩罚的有效性。
{"title":"SPARSE REGULARIZATION FOR BI-LEVEL VARIABLE SELECTION","authors":"H. Matsui","doi":"10.5183/JJSCS.1502001_216","DOIUrl":"https://doi.org/10.5183/JJSCS.1502001_216","url":null,"abstract":"Sparse regularization provides solutions in which some parameters are exactly zero and therefore they can be used for selecting variables in regression models and so on. The lasso is proposed as a method for selecting individual variables for regression models. On the other hand, the group lasso selects groups of variables rather than individuals and therefore it has been used in various fields of applications. More recently, penalties that select variables at both the group and individual levels has been considered. They are so called bi-level selection. In this paper we focus on some penalties that aim for bi-level selection. We overview these penalties and estimation algorithms, and then compare the effectiveness of these penalties from the viewpoint of accuracy of prediction and selection of variables and groups through simulation studies.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127828020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A SEQUENTIAL MULTIPLE COMPARISON PROCEDURE FOR DETECTING A LOWEST DOSE HAVING INTERACTION IN A DOSE-RESPONSE TEST 用于检测剂量-反应试验中具有相互作用的最低剂量的顺序多重比较程序
Pub Date : 2015-12-20 DOI: 10.5183/JJSCS.1406001_212
Tomohiro Nakamura, H. Douke
In this study, we propose a multiple comparison procedure for detecting sequentially a lowest dose level having interaction based on two dose sample means on two treatments with increasing dose levels in a dose-response test. We apply a group sequential procedure in order to realize our method that tests sequentially the null hypotheses of no interaction based on tetrad differences. If we can first detect a dose level having interaction at an early stage in the sequential test, since we can terminate the procedure with just the few observations up to that stage, the procedure is useful from an economical point of view. In the procedure, we present an integral formula to determine the repeated confidence boundaries for satisfying a predefined type I familywise error rate. Furthermore, we show how to decide a required sample size in each cell so as to guarantee the power of the test. In the simulation studies, we evaluate the superiority among the procedures based on three α spending functions in terms of the power of the test and the required sample size for various configurations of population means.
在这项研究中,我们提出了一种多重比较程序,用于在剂量-反应试验中,根据两种剂量水平增加的治疗方法的两个剂量样本均值,依次检测具有相互作用的最低剂量水平。为了实现对基于四分体差异的无相互作用的零假设进行序贯检验的方法,我们采用了一组序贯过程。如果我们能够在顺序试验的早期阶段首先检测到具有相互作用的剂量水平,因为我们可以在该阶段之前仅进行少量观察就结束该程序,因此从经济的角度来看,该程序是有用的。在此过程中,我们提出了一个积分公式来确定满足预定义的I型家族错误率的重复置信边界。此外,我们展示了如何在每个单元中确定所需的样本量,以保证测试的功率。在仿真研究中,我们根据检验的威力和各种总体均值配置所需的样本量来评估基于三个α花费函数的程序之间的优越性。
{"title":"A SEQUENTIAL MULTIPLE COMPARISON PROCEDURE FOR DETECTING A LOWEST DOSE HAVING INTERACTION IN A DOSE-RESPONSE TEST","authors":"Tomohiro Nakamura, H. Douke","doi":"10.5183/JJSCS.1406001_212","DOIUrl":"https://doi.org/10.5183/JJSCS.1406001_212","url":null,"abstract":"In this study, we propose a multiple comparison procedure for detecting sequentially a lowest dose level having interaction based on two dose sample means on two treatments with increasing dose levels in a dose-response test. We apply a group sequential procedure in order to realize our method that tests sequentially the null hypotheses of no interaction based on tetrad differences. If we can first detect a dose level having interaction at an early stage in the sequential test, since we can terminate the procedure with just the few observations up to that stage, the procedure is useful from an economical point of view. In the procedure, we present an integral formula to determine the repeated confidence boundaries for satisfying a predefined type I familywise error rate. Furthermore, we show how to decide a required sample size in each cell so as to guarantee the power of the test. In the simulation studies, we evaluate the superiority among the procedures based on three α spending functions in terms of the power of the test and the required sample size for various configurations of population means.","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134152891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of the Japanese Society of Computational Statistics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1