Low-Dimensional Invariant Embeddings for Universal Geometric Learning

IF 2.5 1区 数学 Q2 COMPUTER SCIENCE, THEORY & METHODS Foundations of Computational Mathematics Pub Date : 2024-02-08 DOI:10.1007/s10208-024-09641-2
Nadav Dym, Steven J. Gortler
{"title":"Low-Dimensional Invariant Embeddings for Universal Geometric Learning","authors":"Nadav Dym, Steven J. Gortler","doi":"10.1007/s10208-024-09641-2","DOIUrl":null,"url":null,"abstract":"<p>This paper studies separating invariants: mappings on <i>D</i>-dimensional domains which are invariant to an appropriate group action and which separate orbits. The motivation for this study comes from the usefulness of separating invariants in proving universality of equivariant neural network architectures. We observe that in several cases the cardinality of separating invariants proposed in the machine learning literature is much larger than the dimension <i>D</i>. As a result, the theoretical universal constructions based on these separating invariants are unrealistically large. Our goal in this paper is to resolve this issue. We show that when a continuous family of semi-algebraic separating invariants is available, separation can be obtained by randomly selecting <span>\\(2D+1 \\)</span> of these invariants. We apply this methodology to obtain an efficient scheme for computing separating invariants for several classical group actions which have been studied in the invariant learning literature. Examples include matrix multiplication actions on point clouds by permutations, rotations, and various other linear groups. Often the requirement of invariant separation is relaxed and only generic separation is required. In this case, we show that only <span>\\(D+1\\)</span> invariants are required. More importantly, generic invariants are often significantly easier to compute, as we illustrate by discussing generic and full separation for weighted graphs. Finally we outline an approach for proving that separating invariants can be constructed also when the random parameters have finite precision.</p>","PeriodicalId":55151,"journal":{"name":"Foundations of Computational Mathematics","volume":"25 1","pages":""},"PeriodicalIF":2.5000,"publicationDate":"2024-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Foundations of Computational Mathematics","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s10208-024-09641-2","RegionNum":1,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0

Abstract

This paper studies separating invariants: mappings on D-dimensional domains which are invariant to an appropriate group action and which separate orbits. The motivation for this study comes from the usefulness of separating invariants in proving universality of equivariant neural network architectures. We observe that in several cases the cardinality of separating invariants proposed in the machine learning literature is much larger than the dimension D. As a result, the theoretical universal constructions based on these separating invariants are unrealistically large. Our goal in this paper is to resolve this issue. We show that when a continuous family of semi-algebraic separating invariants is available, separation can be obtained by randomly selecting \(2D+1 \) of these invariants. We apply this methodology to obtain an efficient scheme for computing separating invariants for several classical group actions which have been studied in the invariant learning literature. Examples include matrix multiplication actions on point clouds by permutations, rotations, and various other linear groups. Often the requirement of invariant separation is relaxed and only generic separation is required. In this case, we show that only \(D+1\) invariants are required. More importantly, generic invariants are often significantly easier to compute, as we illustrate by discussing generic and full separation for weighted graphs. Finally we outline an approach for proving that separating invariants can be constructed also when the random parameters have finite precision.

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
通用几何学习的低维不变嵌入
本文研究分离不变式:D 维域上的映射,这些映射对适当的群作用是不变的,并且分离了轨道。这项研究的动机来自于分离不变式在证明等变神经网络架构普遍性方面的有用性。我们注意到,在一些情况下,机器学习文献中提出的分离不变式的万有性远远大于维数 D。我们在本文中的目标就是解决这个问题。我们证明,当半代数分离不变式的连续族可用时,可以通过随机选择这些不变式中的\(2D+1 \)来获得分离。我们应用这种方法获得了一种高效的方案,用于计算不变式学习文献中已经研究过的几种经典群作用的分离不变式。例如,通过排列、旋转和其他各种线性群对点云进行矩阵乘法运算。通常情况下,不变量分离的要求会被放宽,只要求通用分离。在这种情况下,我们证明只需要(D+1)个不变式。更重要的是,泛函不变式通常更容易计算,我们通过讨论加权图的泛函分离和完全分离来说明这一点。最后,我们概述了一种方法,用于证明当随机参数具有有限精度时,也可以构造分离不变式。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Foundations of Computational Mathematics
Foundations of Computational Mathematics 数学-计算机:理论方法
CiteScore
6.90
自引率
3.30%
发文量
46
审稿时长
>12 weeks
期刊介绍: Foundations of Computational Mathematics (FoCM) will publish research and survey papers of the highest quality which further the understanding of the connections between mathematics and computation. The journal aims to promote the exploration of all fundamental issues underlying the creative tension among mathematics, computer science and application areas unencumbered by any external criteria such as the pressure for applications. The journal will thus serve an increasingly important and applicable area of mathematics. The journal hopes to further the understanding of the deep relationships between mathematical theory: analysis, topology, geometry and algebra, and the computational processes as they are evolving in tandem with the modern computer. With its distinguished editorial board selecting papers of the highest quality and interest from the international community, FoCM hopes to influence both mathematics and computation. Relevance to applications will not constitute a requirement for the publication of articles. The journal does not accept code for review however authors who have code/data related to the submission should include a weblink to the repository where the data/code is stored.
期刊最新文献
Safely Learning Dynamical Systems Stabilizing Decomposition of Multiparameter Persistence Modules Optimal Regularization for a Data Source Sharp Bounds for Max-sliced Wasserstein Distances Optimal Convergence Rates for the Spectrum of the Graph Laplacian on Poisson Point Clouds
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1