{"title":"单边 $\\$sigma $$ 平滑函数下的在线学习","authors":"Hongxiang Zhang, Dachuan Xu, Ling Gai, Zhenning Zhang","doi":"10.1007/s10878-024-01174-2","DOIUrl":null,"url":null,"abstract":"<p>The online optimization model was first introduced in the research of machine learning problems (Zinkevich, Proceedings of ICML, 928–936, 2003). It is a powerful framework that combines the principles of optimization with the challenges of online decision-making. The present research mainly consider the case that the reveal objective functions are convex or submodular. In this paper, we focus on the online maximization problem under a special objective function <span>\\(\\varPhi (x):[0,1]^n\\rightarrow \\mathbb {R}_{+}\\)</span> which satisfies the inequality <span>\\(\\frac{1}{2}\\langle u^{T}\\nabla ^{2}\\varPhi (x),u\\rangle \\le \\sigma \\cdot \\frac{\\Vert u\\Vert _{1}}{\\Vert x\\Vert _{1}}\\langle u,\\nabla \\varPhi (x)\\rangle \\)</span> for any <span>\\(x,u\\in [0,1]^n, x\\ne 0\\)</span>. This objective function is named as one sided <span>\\(\\sigma \\)</span>-smooth (OSS) function. We achieve two conclusions here. Firstly, under the assumption that the gradient function of OSS function is L-smooth, we propose an <span>\\((1-\\exp ((\\theta -1)(\\theta /(1+\\theta ))^{2\\sigma }))\\)</span>- approximation algorithm with <span>\\(O(\\sqrt{T})\\)</span> regret upper bound, where <i>T</i> is the number of rounds in the online algorithm and <span>\\(\\theta , \\sigma \\in \\mathbb {R}_{+}\\)</span> are parameters. Secondly, if the gradient function of OSS function has no L-smoothness, we provide an <span>\\(\\left( 1+((\\theta +1)/\\theta )^{4\\sigma }\\right) ^{-1}\\)</span>-approximation projected gradient algorithm, and prove that the regret upper bound of the algorithm is <span>\\(O(\\sqrt{T})\\)</span>. We think that this research can provide different ideas for online non-convex and non-submodular learning.</p>","PeriodicalId":50231,"journal":{"name":"Journal of Combinatorial Optimization","volume":"48 1","pages":""},"PeriodicalIF":0.9000,"publicationDate":"2024-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Online learning under one sided $$\\\\sigma $$ -smooth function\",\"authors\":\"Hongxiang Zhang, Dachuan Xu, Ling Gai, Zhenning Zhang\",\"doi\":\"10.1007/s10878-024-01174-2\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>The online optimization model was first introduced in the research of machine learning problems (Zinkevich, Proceedings of ICML, 928–936, 2003). It is a powerful framework that combines the principles of optimization with the challenges of online decision-making. The present research mainly consider the case that the reveal objective functions are convex or submodular. In this paper, we focus on the online maximization problem under a special objective function <span>\\\\(\\\\varPhi (x):[0,1]^n\\\\rightarrow \\\\mathbb {R}_{+}\\\\)</span> which satisfies the inequality <span>\\\\(\\\\frac{1}{2}\\\\langle u^{T}\\\\nabla ^{2}\\\\varPhi (x),u\\\\rangle \\\\le \\\\sigma \\\\cdot \\\\frac{\\\\Vert u\\\\Vert _{1}}{\\\\Vert x\\\\Vert _{1}}\\\\langle u,\\\\nabla \\\\varPhi (x)\\\\rangle \\\\)</span> for any <span>\\\\(x,u\\\\in [0,1]^n, x\\\\ne 0\\\\)</span>. This objective function is named as one sided <span>\\\\(\\\\sigma \\\\)</span>-smooth (OSS) function. We achieve two conclusions here. Firstly, under the assumption that the gradient function of OSS function is L-smooth, we propose an <span>\\\\((1-\\\\exp ((\\\\theta -1)(\\\\theta /(1+\\\\theta ))^{2\\\\sigma }))\\\\)</span>- approximation algorithm with <span>\\\\(O(\\\\sqrt{T})\\\\)</span> regret upper bound, where <i>T</i> is the number of rounds in the online algorithm and <span>\\\\(\\\\theta , \\\\sigma \\\\in \\\\mathbb {R}_{+}\\\\)</span> are parameters. Secondly, if the gradient function of OSS function has no L-smoothness, we provide an <span>\\\\(\\\\left( 1+((\\\\theta +1)/\\\\theta )^{4\\\\sigma }\\\\right) ^{-1}\\\\)</span>-approximation projected gradient algorithm, and prove that the regret upper bound of the algorithm is <span>\\\\(O(\\\\sqrt{T})\\\\)</span>. We think that this research can provide different ideas for online non-convex and non-submodular learning.</p>\",\"PeriodicalId\":50231,\"journal\":{\"name\":\"Journal of Combinatorial Optimization\",\"volume\":\"48 1\",\"pages\":\"\"},\"PeriodicalIF\":0.9000,\"publicationDate\":\"2024-05-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Combinatorial Optimization\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1007/s10878-024-01174-2\",\"RegionNum\":4,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Combinatorial Optimization","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s10878-024-01174-2","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
Online learning under one sided $$\sigma $$ -smooth function
The online optimization model was first introduced in the research of machine learning problems (Zinkevich, Proceedings of ICML, 928–936, 2003). It is a powerful framework that combines the principles of optimization with the challenges of online decision-making. The present research mainly consider the case that the reveal objective functions are convex or submodular. In this paper, we focus on the online maximization problem under a special objective function \(\varPhi (x):[0,1]^n\rightarrow \mathbb {R}_{+}\) which satisfies the inequality \(\frac{1}{2}\langle u^{T}\nabla ^{2}\varPhi (x),u\rangle \le \sigma \cdot \frac{\Vert u\Vert _{1}}{\Vert x\Vert _{1}}\langle u,\nabla \varPhi (x)\rangle \) for any \(x,u\in [0,1]^n, x\ne 0\). This objective function is named as one sided \(\sigma \)-smooth (OSS) function. We achieve two conclusions here. Firstly, under the assumption that the gradient function of OSS function is L-smooth, we propose an \((1-\exp ((\theta -1)(\theta /(1+\theta ))^{2\sigma }))\)- approximation algorithm with \(O(\sqrt{T})\) regret upper bound, where T is the number of rounds in the online algorithm and \(\theta , \sigma \in \mathbb {R}_{+}\) are parameters. Secondly, if the gradient function of OSS function has no L-smoothness, we provide an \(\left( 1+((\theta +1)/\theta )^{4\sigma }\right) ^{-1}\)-approximation projected gradient algorithm, and prove that the regret upper bound of the algorithm is \(O(\sqrt{T})\). We think that this research can provide different ideas for online non-convex and non-submodular learning.
期刊介绍:
The objective of Journal of Combinatorial Optimization is to advance and promote the theory and applications of combinatorial optimization, which is an area of research at the intersection of applied mathematics, computer science, and operations research and which overlaps with many other areas such as computation complexity, computational biology, VLSI design, communication networks, and management science. It includes complexity analysis and algorithm design for combinatorial optimization problems, numerical experiments and problem discovery with applications in science and engineering.
The Journal of Combinatorial Optimization publishes refereed papers dealing with all theoretical, computational and applied aspects of combinatorial optimization. It also publishes reviews of appropriate books and special issues of journals.