Gautam Chandrasekaran, Adam Klivans, Vasilis Kontonis, Raghu Meka, Konstantinos Stavropoulos
{"title":"Smoothed Analysis for Learning Concepts with Low Intrinsic Dimension","authors":"Gautam Chandrasekaran, Adam Klivans, Vasilis Kontonis, Raghu Meka, Konstantinos Stavropoulos","doi":"arxiv-2407.00966","DOIUrl":null,"url":null,"abstract":"In traditional models of supervised learning, the goal of a learner -- given\nexamples from an arbitrary joint distribution on $\\mathbb{R}^d \\times \\{\\pm\n1\\}$ -- is to output a hypothesis that is competitive (to within $\\epsilon$) of\nthe best fitting concept from some class. In order to escape strong hardness\nresults for learning even simple concept classes, we introduce a\nsmoothed-analysis framework that requires a learner to compete only with the\nbest classifier that is robust to small random Gaussian perturbation. This subtle change allows us to give a wide array of learning results for any\nconcept that (1) depends on a low-dimensional subspace (aka multi-index model)\nand (2) has a bounded Gaussian surface area. This class includes functions of\nhalfspaces and (low-dimensional) convex sets, cases that are only known to be\nlearnable in non-smoothed settings with respect to highly structured\ndistributions such as Gaussians. Surprisingly, our analysis also yields new results for traditional\nnon-smoothed frameworks such as learning with margin. In particular, we obtain\nthe first algorithm for agnostically learning intersections of $k$-halfspaces\nin time $k^{poly(\\frac{\\log k}{\\epsilon \\gamma}) }$ where $\\gamma$ is the\nmargin parameter. Before our work, the best-known runtime was exponential in\n$k$ (Arriaga and Vempala, 1999).","PeriodicalId":501024,"journal":{"name":"arXiv - CS - Computational Complexity","volume":"18 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computational Complexity","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.00966","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In traditional models of supervised learning, the goal of a learner -- given
examples from an arbitrary joint distribution on $\mathbb{R}^d \times \{\pm
1\}$ -- is to output a hypothesis that is competitive (to within $\epsilon$) of
the best fitting concept from some class. In order to escape strong hardness
results for learning even simple concept classes, we introduce a
smoothed-analysis framework that requires a learner to compete only with the
best classifier that is robust to small random Gaussian perturbation. This subtle change allows us to give a wide array of learning results for any
concept that (1) depends on a low-dimensional subspace (aka multi-index model)
and (2) has a bounded Gaussian surface area. This class includes functions of
halfspaces and (low-dimensional) convex sets, cases that are only known to be
learnable in non-smoothed settings with respect to highly structured
distributions such as Gaussians. Surprisingly, our analysis also yields new results for traditional
non-smoothed frameworks such as learning with margin. In particular, we obtain
the first algorithm for agnostically learning intersections of $k$-halfspaces
in time $k^{poly(\frac{\log k}{\epsilon \gamma}) }$ where $\gamma$ is the
margin parameter. Before our work, the best-known runtime was exponential in
$k$ (Arriaga and Vempala, 1999).