{"title":"On the Hardness of Meaningful Local Guarantees in Nonsmooth Nonconvex Optimization","authors":"Guy Kornowski, Swati Padmanabhan, Ohad Shamir","doi":"arxiv-2409.10323","DOIUrl":null,"url":null,"abstract":"We study the oracle complexity of nonsmooth nonconvex optimization, with the\nalgorithm assumed to have access only to local function information. It has\nbeen shown by Davis, Drusvyatskiy, and Jiang (2023) that for nonsmooth\nLipschitz functions satisfying certain regularity and strictness conditions,\nperturbed gradient descent converges to local minimizers asymptotically.\nMotivated by this result and by other recent algorithmic advances in nonconvex\nnonsmooth optimization concerning Goldstein stationarity, we consider the\nquestion of obtaining a non-asymptotic rate of convergence to local minima for\nthis problem class. We provide the following negative answer to this question: Local algorithms\nacting on regular Lipschitz functions cannot, in the worst case, provide\nmeaningful local guarantees in terms of function value in sub-exponential time,\neven when all near-stationary points are global minima. This sharply contrasts\nwith the smooth setting, for which it is well-known that standard gradient\nmethods can do so in a dimension-independent rate. Our result complements the\nrich body of work in the theoretical computer science literature that provide\nhardness results conditional on conjectures such as $\\mathsf{P}\\neq\\mathsf{NP}$\nor cryptographic assumptions, in that ours holds unconditional of any such\nassumptions.","PeriodicalId":501286,"journal":{"name":"arXiv - MATH - Optimization and Control","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Optimization and Control","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10323","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We study the oracle complexity of nonsmooth nonconvex optimization, with the
algorithm assumed to have access only to local function information. It has
been shown by Davis, Drusvyatskiy, and Jiang (2023) that for nonsmooth
Lipschitz functions satisfying certain regularity and strictness conditions,
perturbed gradient descent converges to local minimizers asymptotically.
Motivated by this result and by other recent algorithmic advances in nonconvex
nonsmooth optimization concerning Goldstein stationarity, we consider the
question of obtaining a non-asymptotic rate of convergence to local minima for
this problem class. We provide the following negative answer to this question: Local algorithms
acting on regular Lipschitz functions cannot, in the worst case, provide
meaningful local guarantees in terms of function value in sub-exponential time,
even when all near-stationary points are global minima. This sharply contrasts
with the smooth setting, for which it is well-known that standard gradient
methods can do so in a dimension-independent rate. Our result complements the
rich body of work in the theoretical computer science literature that provide
hardness results conditional on conjectures such as $\mathsf{P}\neq\mathsf{NP}$
or cryptographic assumptions, in that ours holds unconditional of any such
assumptions.