EDITORIAL: RECENT ADVANCES IN SPARSE STATISTICAL MODELING

K. Hirose
{"title":"EDITORIAL: RECENT ADVANCES IN SPARSE STATISTICAL MODELING","authors":"K. Hirose","doi":"10.5183/JJSCS.1510002_225","DOIUrl":null,"url":null,"abstract":"The first term L(β) is a loss function and the second term λ ∑p j=1 |βj | is a penalty term. Here λ (λ > 0) is a tuning parameter which controls the sparsity and the model fitting. Because the penalty term consists of the sum of absolute values of the parameter, we can carry out the sparse estimation, that is, some of the elements of β are estimated by exactly zeros. It is well-known that we cannot often obtain the analytical solutions of the minimization problem (1), because the penalty term λ ∑p j=1 |βj | is indifferentiable when βj = 0 (j = 1, . . . , p). Therefore, it is important to develop efficient computational algorithms. This special issue includes six interesting papers related to sparse estimation. These papers cover a wide variety of topics, such as statistical modeling, computation, theoretical analysis, and applications. In particular, all of the papers deal with the issue of statistical computation. Kawasaki and Ueki (the first paper of this issue) apply smooth-threshold estimating equations (STEE, Ueki, 2009) to telemarketing success data collected from a Portuguese retail bank. In STEE, the penalty term consists of a quadratic form ∑p j=1 wjβ 2 j instead of ∑p j=1 |βj |, where wj (j = 1, . . . , p) are positive values allowed to be ∞, so that we do not need to implement a computational algorithm that is used in the L1 regularization. Kawano, Hoshina, Shimamura and Konishi (the second paper) propose a model selection criterion for choosing tuning parameters in the Bayesian lasso (Park and Casella, 2008). They use an efficient sparse estimation algorithm in the Bayesian lasso, referred to as the sparse algorithm. Matsui (the third paper) considers the problem of bi-level selection, which allows the selection of groups of variables and individuals simultaneously. The parameter estimation procedure is based on the coordinate descent algorithm, which is known as a remarkably fast algorithm (Friedman et al., 2010). Suzuki (the fourth paper) focuses attention on the alternating direction method of multipliers algorithm (ADMM algorithm, Boyd et al., 2011), which is applicable to various complex penalties such as the overlapping group lasso (Jacob et al., 2009). He reviews a stochastic version of the ADMM algorithm that allows the online learning. Hino and Fujiki (the fifth paper) propose a penalized linear discriminant analysis that adheres to the normal discriminant model. They apply the Majorize-Minimization algorithm (MM algorithm, Hunter and Lange 2004), which is often used to replace a non-convex optimization problem with a reweighted convex optimization","PeriodicalId":338719,"journal":{"name":"Journal of the Japanese Society of Computational Statistics","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the Japanese Society of Computational Statistics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5183/JJSCS.1510002_225","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The first term L(β) is a loss function and the second term λ ∑p j=1 |βj | is a penalty term. Here λ (λ > 0) is a tuning parameter which controls the sparsity and the model fitting. Because the penalty term consists of the sum of absolute values of the parameter, we can carry out the sparse estimation, that is, some of the elements of β are estimated by exactly zeros. It is well-known that we cannot often obtain the analytical solutions of the minimization problem (1), because the penalty term λ ∑p j=1 |βj | is indifferentiable when βj = 0 (j = 1, . . . , p). Therefore, it is important to develop efficient computational algorithms. This special issue includes six interesting papers related to sparse estimation. These papers cover a wide variety of topics, such as statistical modeling, computation, theoretical analysis, and applications. In particular, all of the papers deal with the issue of statistical computation. Kawasaki and Ueki (the first paper of this issue) apply smooth-threshold estimating equations (STEE, Ueki, 2009) to telemarketing success data collected from a Portuguese retail bank. In STEE, the penalty term consists of a quadratic form ∑p j=1 wjβ 2 j instead of ∑p j=1 |βj |, where wj (j = 1, . . . , p) are positive values allowed to be ∞, so that we do not need to implement a computational algorithm that is used in the L1 regularization. Kawano, Hoshina, Shimamura and Konishi (the second paper) propose a model selection criterion for choosing tuning parameters in the Bayesian lasso (Park and Casella, 2008). They use an efficient sparse estimation algorithm in the Bayesian lasso, referred to as the sparse algorithm. Matsui (the third paper) considers the problem of bi-level selection, which allows the selection of groups of variables and individuals simultaneously. The parameter estimation procedure is based on the coordinate descent algorithm, which is known as a remarkably fast algorithm (Friedman et al., 2010). Suzuki (the fourth paper) focuses attention on the alternating direction method of multipliers algorithm (ADMM algorithm, Boyd et al., 2011), which is applicable to various complex penalties such as the overlapping group lasso (Jacob et al., 2009). He reviews a stochastic version of the ADMM algorithm that allows the online learning. Hino and Fujiki (the fifth paper) propose a penalized linear discriminant analysis that adheres to the normal discriminant model. They apply the Majorize-Minimization algorithm (MM algorithm, Hunter and Lange 2004), which is often used to replace a non-convex optimization problem with a reweighted convex optimization
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
编辑:稀疏统计建模的最新进展
第一项L(β)是一个损失函数,第二项λ∑p j=1 |βj |是一个惩罚项。这里λ (λ > 0)是一个调节参数,它控制稀疏性和模型拟合。由于惩罚项由参数的绝对值和组成,我们可以进行稀疏估计,即β的一些元素被精确地估计为零。众所周知,我们不能经常得到最小化问题(1)的解析解,因为当βj = 0 (j =1,…)时,惩罚项λ∑p j=1 |βj |是不可微的。, p)。因此,开发高效的计算算法非常重要。本期特刊包括六篇与稀疏估计相关的有趣论文。这些论文涵盖了各种各样的主题,如统计建模、计算、理论分析和应用。特别地,所有的论文都涉及统计计算问题。Kawasaki和Ueki(本期的第一篇论文)将平滑阈值估计方程(STEE, Ueki, 2009)应用于从葡萄牙零售银行收集的电话营销成功数据。在STEE中,惩罚项由二次型∑p j=1 wjβ 2j代替∑p j=1 |βj |,其中wj (j =1,…, p)是允许为∞的正值,因此我们不需要实现L1正则化中使用的计算算法。Kawano, Hoshina, Shimamura和Konishi(第二篇论文)提出了在贝叶斯套索中选择调谐参数的模型选择标准(Park和Casella, 2008)。他们在贝叶斯套索中使用了一种高效的稀疏估计算法,称为稀疏算法。Matsui(第三篇论文)考虑了双水平选择问题,允许同时选择变量组和个体组。参数估计过程基于坐标下降算法,这是一种非常快速的算法(Friedman et al., 2010)。Suzuki(第四篇论文)重点研究了乘法器算法(ADMM算法,Boyd et al., 2011)的交替方向法,该算法适用于重叠组lasso (Jacob et al., 2009)等各种复杂处罚。他回顾了ADMM算法的随机版本,该算法允许在线学习。Hino和Fujiki(第五篇论文)提出了一种符合正常判别模型的惩罚线性判别分析。他们应用了最大化最小化算法(MM算法,Hunter and Lange 2004),该算法通常用于用重新加权的凸优化取代非凸优化问题
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
ENHANCING POWER OF SCORE TESTS FOR REGRESSION MODELS VIA FISHER TRANSFORMATION ANNOUNCEMENT: ON PUBLICATION OF THE JAPANESE JOURNAL OF STATISTICS AND DATA SCIENCE DISTRIBUTION OF THE LARGEST EIGENVALUE OF AN ELLIPTICAL WISHART MATRIX AND ITS SIMULATION COMMENT: ON CLOSING OF ENGLISH JOURNAL OF JSCS AND THE BIRTH OF NEW JOURNAL JJSD INFERENCE FOR THE EXTENT PARAMETER OF DAMAGE BY TSUNAMI WITH POINCARE CONES
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1