{"title":"针对结构化非凸和非光滑优化的带有一般松弛因子的部分布雷格曼 ADMM","authors":"Jianghua Yin, Chunming Tang, Jinbao Jian, Qiongxuan Huang","doi":"10.1007/s10898-024-01384-2","DOIUrl":null,"url":null,"abstract":"<p>In this paper, a partial Bregman alternating direction method of multipliers (ADMM) with a general relaxation factor <span>\\(\\alpha \\in (0,\\frac{1+\\sqrt{5}}{2})\\)</span> is proposed for structured nonconvex and nonsmooth optimization, where the objective function is the sum of a nonsmooth convex function and a smooth nonconvex function without coupled variables. We add a Bregman distance to alleviate the difficulty of solving the nonsmooth subproblem. For the smooth subproblem, we directly perform a gradient descent step of the augmented Lagrangian function, which makes the computational cost of each iteration of our method very cheap. To our knowledge, the nonconvex ADMM with a relaxation factor <span>\\(\\alpha \\ne 1\\)</span> in the literature has never been studied for the problem under consideration. Under some mild conditions, the boundedness of the generated sequence, the global convergence and the iteration complexity are established. The numerical results verify the efficiency and robustness of the proposed method.</p>","PeriodicalId":15961,"journal":{"name":"Journal of Global Optimization","volume":"11 1","pages":""},"PeriodicalIF":1.8000,"publicationDate":"2024-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A partial Bregman ADMM with a general relaxation factor for structured nonconvex and nonsmooth optimization\",\"authors\":\"Jianghua Yin, Chunming Tang, Jinbao Jian, Qiongxuan Huang\",\"doi\":\"10.1007/s10898-024-01384-2\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>In this paper, a partial Bregman alternating direction method of multipliers (ADMM) with a general relaxation factor <span>\\\\(\\\\alpha \\\\in (0,\\\\frac{1+\\\\sqrt{5}}{2})\\\\)</span> is proposed for structured nonconvex and nonsmooth optimization, where the objective function is the sum of a nonsmooth convex function and a smooth nonconvex function without coupled variables. We add a Bregman distance to alleviate the difficulty of solving the nonsmooth subproblem. For the smooth subproblem, we directly perform a gradient descent step of the augmented Lagrangian function, which makes the computational cost of each iteration of our method very cheap. To our knowledge, the nonconvex ADMM with a relaxation factor <span>\\\\(\\\\alpha \\\\ne 1\\\\)</span> in the literature has never been studied for the problem under consideration. Under some mild conditions, the boundedness of the generated sequence, the global convergence and the iteration complexity are established. The numerical results verify the efficiency and robustness of the proposed method.</p>\",\"PeriodicalId\":15961,\"journal\":{\"name\":\"Journal of Global Optimization\",\"volume\":\"11 1\",\"pages\":\"\"},\"PeriodicalIF\":1.8000,\"publicationDate\":\"2024-03-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Global Optimization\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1007/s10898-024-01384-2\",\"RegionNum\":3,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"Mathematics\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Global Optimization","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s10898-024-01384-2","RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Mathematics","Score":null,"Total":0}
A partial Bregman ADMM with a general relaxation factor for structured nonconvex and nonsmooth optimization
In this paper, a partial Bregman alternating direction method of multipliers (ADMM) with a general relaxation factor \(\alpha \in (0,\frac{1+\sqrt{5}}{2})\) is proposed for structured nonconvex and nonsmooth optimization, where the objective function is the sum of a nonsmooth convex function and a smooth nonconvex function without coupled variables. We add a Bregman distance to alleviate the difficulty of solving the nonsmooth subproblem. For the smooth subproblem, we directly perform a gradient descent step of the augmented Lagrangian function, which makes the computational cost of each iteration of our method very cheap. To our knowledge, the nonconvex ADMM with a relaxation factor \(\alpha \ne 1\) in the literature has never been studied for the problem under consideration. Under some mild conditions, the boundedness of the generated sequence, the global convergence and the iteration complexity are established. The numerical results verify the efficiency and robustness of the proposed method.
期刊介绍:
The Journal of Global Optimization publishes carefully refereed papers that encompass theoretical, computational, and applied aspects of global optimization. While the focus is on original research contributions dealing with the search for global optima of non-convex, multi-extremal problems, the journal’s scope covers optimization in the widest sense, including nonlinear, mixed integer, combinatorial, stochastic, robust, multi-objective optimization, computational geometry, and equilibrium problems. Relevant works on data-driven methods and optimization-based data mining are of special interest.
In addition to papers covering theory and algorithms of global optimization, the journal publishes significant papers on numerical experiments, new testbeds, and applications in engineering, management, and the sciences. Applications of particular interest include healthcare, computational biochemistry, energy systems, telecommunications, and finance. Apart from full-length articles, the journal features short communications on both open and solved global optimization problems. It also offers reviews of relevant books and publishes special issues.