Rayna Andreeva, James Ward, Primoz Skraba, Jie Gao, Rik Sarkar
{"title":"Approximating Metric Magnitude of Point Sets","authors":"Rayna Andreeva, James Ward, Primoz Skraba, Jie Gao, Rik Sarkar","doi":"arxiv-2409.04411","DOIUrl":null,"url":null,"abstract":"Metric magnitude is a measure of the \"size\" of point clouds with many\ndesirable geometric properties. It has been adapted to various mathematical\ncontexts and recent work suggests that it can enhance machine learning and\noptimization algorithms. But its usability is limited due to the computational\ncost when the dataset is large or when the computation must be carried out\nrepeatedly (e.g. in model training). In this paper, we study the magnitude\ncomputation problem, and show efficient ways of approximating it. We show that\nit can be cast as a convex optimization problem, but not as a submodular\noptimization. The paper describes two new algorithms - an iterative\napproximation algorithm that converges fast and is accurate, and a subset\nselection method that makes the computation even faster. It has been previously\nproposed that magnitude of model sequences generated during stochastic gradient\ndescent is correlated to generalization gap. Extension of this result using our\nmore scalable algorithms shows that longer sequences in fact bear higher\ncorrelations. We also describe new applications of magnitude in machine\nlearning - as an effective regularizer for neural network training, and as a\nnovel clustering criterion.","PeriodicalId":501444,"journal":{"name":"arXiv - MATH - Metric Geometry","volume":"33 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - MATH - Metric Geometry","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.04411","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Metric magnitude is a measure of the "size" of point clouds with many
desirable geometric properties. It has been adapted to various mathematical
contexts and recent work suggests that it can enhance machine learning and
optimization algorithms. But its usability is limited due to the computational
cost when the dataset is large or when the computation must be carried out
repeatedly (e.g. in model training). In this paper, we study the magnitude
computation problem, and show efficient ways of approximating it. We show that
it can be cast as a convex optimization problem, but not as a submodular
optimization. The paper describes two new algorithms - an iterative
approximation algorithm that converges fast and is accurate, and a subset
selection method that makes the computation even faster. It has been previously
proposed that magnitude of model sequences generated during stochastic gradient
descent is correlated to generalization gap. Extension of this result using our
more scalable algorithms shows that longer sequences in fact bear higher
correlations. We also describe new applications of magnitude in machine
learning - as an effective regularizer for neural network training, and as a
novel clustering criterion.