首页 > 最新文献

Applied and Computational Harmonic Analysis最新文献

英文 中文
Unified stochastic framework for neural network quantization and pruning 神经网络量化与剪枝的统一随机框架
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-06-02 DOI: 10.1016/j.acha.2025.101778
Haoyu Zhang , Rayan Saab
Quantization and pruning are two essential techniques for compressing neural networks, yet they are often treated independently, with limited theoretical analysis connecting them. This paper introduces a unified framework for post-training quantization and pruning using stochastic path-following algorithms. Our approach builds on the Stochastic Path Following Quantization (SPFQ) method, extending its applicability to pruning and low-bit quantization, including challenging 1-bit regimes. By incorporating a scaling parameter and generalizing the stochastic operator, the proposed method achieves robust error correction and yields rigorous theoretical error bounds for both quantization and pruning as well as their combination.
量化和剪枝是压缩神经网络的两种基本技术,但它们往往被独立对待,很少有理论分析将它们联系起来。本文介绍了一个使用随机路径跟踪算法进行训练后量化和剪枝的统一框架。我们的方法建立在随机路径跟随量化(SPFQ)方法的基础上,扩展了其对剪枝和低比特量化的适用性,包括具有挑战性的1比特制度。通过引入尺度参数和推广随机算子,该方法实现了鲁棒误差校正,并为量化和剪枝及其组合提供了严格的理论误差界。
{"title":"Unified stochastic framework for neural network quantization and pruning","authors":"Haoyu Zhang ,&nbsp;Rayan Saab","doi":"10.1016/j.acha.2025.101778","DOIUrl":"10.1016/j.acha.2025.101778","url":null,"abstract":"<div><div>Quantization and pruning are two essential techniques for compressing neural networks, yet they are often treated independently, with limited theoretical analysis connecting them. This paper introduces a unified framework for post-training quantization and pruning using stochastic path-following algorithms. Our approach builds on the Stochastic Path Following Quantization (SPFQ) method, extending its applicability to pruning and low-bit quantization, including challenging 1-bit regimes. By incorporating a scaling parameter and generalizing the stochastic operator, the proposed method achieves robust error correction and yields rigorous theoretical error bounds for both quantization and pruning as well as their combination.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"79 ","pages":"Article 101778"},"PeriodicalIF":2.6,"publicationDate":"2025-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144230135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A tighter generalization error bound for wide GCN based on loss landscape 基于损失分布的广义GCN更严格的泛化误差界
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-05-21 DOI: 10.1016/j.acha.2025.101777
Xianchen Zhou , Kun Hu , Hongxia Wang
The generalization capability of Graph Convolutional Networks (GCNs) has been researched recently. The generalization error bound based on algorithmic stability is obtained for various structures of GCN. However, the generalization error bound computed by this method increases rapidly during the iteration since the algorithmic stability exponential depends on the number of iterations, which is not consistent with the performance of GCNs in practice. Based on the fact that the property of loss landscape, such as convex, exp-concave, or Polyak-Lojasiewicz* (PL*) leads to tighter stability and better generalization error bound, this paper focuses on the semi-supervised loss landscape of wide GCN. It shows that a wide GCN has a Hessian matrix with a small norm, which can lead to a positive definite training tangent kernel. Then GCN's loss can satisfy the PL* condition and lead to a tighter uniform stability independent of the iteration compared with previous work. Therefore, the generalization error bound in this paper depends on the graph filter's norm and layers, which is consistent with the experiments' results.
图卷积网络(GCNs)的泛化能力是近年来研究的热点。针对不同的GCN结构,给出了基于算法稳定性的泛化误差界。然而,由于算法稳定性指数依赖于迭代次数,该方法计算的泛化误差界在迭代过程中迅速增大,这与实际GCNs的性能不一致。基于损失格局如凸、expo -凹或Polyak-Lojasiewicz* (PL*)的性质导致更强的稳定性和更好的泛化误差界,本文重点研究了宽GCN的半监督损失格局。结果表明,宽GCN具有一个小范数的Hessian矩阵,可以得到正定的训练切核。那么GCN的损失可以满足PL*条件,并且与之前的工作相比具有更严格的不受迭代影响的均匀稳定性。因此,本文的泛化误差界取决于图滤波器的范数和层数,与实验结果一致。
{"title":"A tighter generalization error bound for wide GCN based on loss landscape","authors":"Xianchen Zhou ,&nbsp;Kun Hu ,&nbsp;Hongxia Wang","doi":"10.1016/j.acha.2025.101777","DOIUrl":"10.1016/j.acha.2025.101777","url":null,"abstract":"<div><div>The generalization capability of Graph Convolutional Networks (GCNs) has been researched recently. The generalization error bound based on algorithmic stability is obtained for various structures of GCN. However, the generalization error bound computed by this method increases rapidly during the iteration since the algorithmic stability exponential depends on the number of iterations, which is not consistent with the performance of GCNs in practice. Based on the fact that the property of loss landscape, such as convex, exp-concave, or Polyak-Lojasiewicz* (PL*) leads to tighter stability and better generalization error bound, this paper focuses on the semi-supervised loss landscape of wide GCN. It shows that a wide GCN has a Hessian matrix with a small norm, which can lead to a positive definite training tangent kernel. Then GCN's loss can satisfy the PL* condition and lead to a tighter uniform stability independent of the iteration compared with previous work. Therefore, the generalization error bound in this paper depends on the graph filter's norm and layers, which is consistent with the experiments' results.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"78 ","pages":"Article 101777"},"PeriodicalIF":2.6,"publicationDate":"2025-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144108039","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An eigenfunction approach to conversion of the Laplace transform of point masses on the real line to the Fourier domain 实线上质点的拉普拉斯变换到傅里叶域的特征函数转换方法
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-05-21 DOI: 10.1016/j.acha.2025.101776
Michael E. Mckenna , Hrushikesh N. Mhaskar , Richard G. Spencer
Motivated by applications in magnetic resonance relaxometry, we consider the following problem: given samples of a function tk=1KAkexp(tλk), where K2 is an integer, AkR, λk>0 for k=1,,K, determine K, Ak's and λk's. Unlike the case in which the λk's are purely imaginary, this problem is notoriously ill-posed. Our goal is to show that this problem can be transformed into an equivalent one in which the λk's are replaced by iλk. We show that this may be accomplished by approximation in terms of Hermite functions, and using the fact that these functions are eigenfunctions of the Fourier transform. We present a preliminary numerical exploration of parameter extraction from this formalism, including the effect of noise. The inherent ill-posedness of the original problem persists in the new domain, as reflected in the numerical results.
由磁共振弛豫测量中的应用驱动,我们考虑以下问题:给定函数t∈∑k=1KAkexp (- tλk)的样本,其中k≥2是整数,Ak∈R, λk>;0对于k=1,⋯k,确定k, Ak和λk。不像λk是纯虚的情况,这个问题是出了名的不适定的。我们的目标是证明这个问题可以转化成一个等价的问题其中λk被λk取代。我们证明这可以通过埃尔米特函数的近似来实现,并且利用这些函数是傅里叶变换的特征函数这一事实。我们提出了从这种形式中提取参数的初步数值探索,包括噪声的影响。正如数值结果所反映的那样,原问题固有的不适定性在新域中仍然存在。
{"title":"An eigenfunction approach to conversion of the Laplace transform of point masses on the real line to the Fourier domain","authors":"Michael E. Mckenna ,&nbsp;Hrushikesh N. Mhaskar ,&nbsp;Richard G. Spencer","doi":"10.1016/j.acha.2025.101776","DOIUrl":"10.1016/j.acha.2025.101776","url":null,"abstract":"<div><div>Motivated by applications in magnetic resonance relaxometry, we consider the following problem: given samples of a function <span><math><mi>t</mi><mo>↦</mo><msubsup><mrow><mo>∑</mo></mrow><mrow><mi>k</mi><mo>=</mo><mn>1</mn></mrow><mrow><mi>K</mi></mrow></msubsup><msub><mrow><mi>A</mi></mrow><mrow><mi>k</mi></mrow></msub><mi>exp</mi><mo>⁡</mo><mo>(</mo><mo>−</mo><mi>t</mi><msub><mrow><mi>λ</mi></mrow><mrow><mi>k</mi></mrow></msub><mo>)</mo></math></span>, where <span><math><mi>K</mi><mo>≥</mo><mn>2</mn></math></span> is an integer, <span><math><msub><mrow><mi>A</mi></mrow><mrow><mi>k</mi></mrow></msub><mo>∈</mo><mi>R</mi></math></span>, <span><math><msub><mrow><mi>λ</mi></mrow><mrow><mi>k</mi></mrow></msub><mo>&gt;</mo><mn>0</mn></math></span> for <span><math><mi>k</mi><mo>=</mo><mn>1</mn><mo>,</mo><mo>⋯</mo><mo>,</mo><mi>K</mi></math></span>, determine <em>K</em>, <span><math><msub><mrow><mi>A</mi></mrow><mrow><mi>k</mi></mrow></msub></math></span>'s and <span><math><msub><mrow><mi>λ</mi></mrow><mrow><mi>k</mi></mrow></msub></math></span>'s. Unlike the case in which the <span><math><msub><mrow><mi>λ</mi></mrow><mrow><mi>k</mi></mrow></msub></math></span>'s are purely imaginary, this problem is notoriously ill-posed. Our goal is to show that this problem can be transformed into an equivalent one in which the <span><math><msub><mrow><mi>λ</mi></mrow><mrow><mi>k</mi></mrow></msub></math></span>'s are replaced by <span><math><mi>i</mi><msub><mrow><mi>λ</mi></mrow><mrow><mi>k</mi></mrow></msub></math></span>. We show that this may be accomplished by approximation in terms of Hermite functions, and using the fact that these functions are eigenfunctions of the Fourier transform. We present a preliminary numerical exploration of parameter extraction from this formalism, including the effect of noise. The inherent ill-posedness of the original problem persists in the new domain, as reflected in the numerical results.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"79 ","pages":"Article 101776"},"PeriodicalIF":2.6,"publicationDate":"2025-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144222723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Framelet message passing 小框架消息传递
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-05-12 DOI: 10.1016/j.acha.2025.101773
Xinliang Liu , Bingxin Zhou , Chutian Zhang , Yu Guang Wang
Graph neural networks have achieved champions in wide applications. Neural message passing is a typical key module for feature propagation by aggregating neighboring features. In this work, we propose a new message passing based on multiscale framelet transforms, called Framelet Message Passing. Different from traditional spatial methods, it integrates framelet representation of neighbor nodes from multiple hops away in node message update. We also propose a continuous message passing using neural ODE solvers. Both discrete and continuous cases can provably mitigate oversmoothing and achieve superior performance. Numerical experiments on real graph datasets show that the continuous version of the framelet message passing significantly outperforms existing methods when learning heterogeneous graphs and achieves state-of-the-art performance on classic node classification tasks with low computational costs.
图神经网络已经取得了广泛的应用。神经信息传递是特征传播的典型关键模块,它通过对相邻特征的聚合实现特征传播。在这项工作中,我们提出了一种新的基于多尺度框架变换的消息传递,称为框架消息传递。与传统的空间方法不同,该方法在节点消息更新中集成了多跳相邻节点的框架表示。我们还提出了一个使用神经ODE求解器的连续消息传递。离散和连续两种情况都能有效地缓解过平滑,并获得较好的性能。在实际图数据集上的数值实验表明,连续版本的框架消息传递在学习异构图时明显优于现有的方法,并且在经典节点分类任务中具有较低的计算成本,达到了最先进的性能。
{"title":"Framelet message passing","authors":"Xinliang Liu ,&nbsp;Bingxin Zhou ,&nbsp;Chutian Zhang ,&nbsp;Yu Guang Wang","doi":"10.1016/j.acha.2025.101773","DOIUrl":"10.1016/j.acha.2025.101773","url":null,"abstract":"<div><div>Graph neural networks have achieved champions in wide applications. Neural message passing is a typical key module for feature propagation by aggregating neighboring features. In this work, we propose a new message passing based on multiscale framelet transforms, called Framelet Message Passing. Different from traditional spatial methods, it integrates framelet representation of neighbor nodes from multiple hops away in node message update. We also propose a continuous message passing using neural ODE solvers. Both discrete and continuous cases can provably mitigate oversmoothing and achieve superior performance. Numerical experiments on real graph datasets show that the continuous version of the framelet message passing significantly outperforms existing methods when learning heterogeneous graphs and achieves state-of-the-art performance on classic node classification tasks with low computational costs.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"78 ","pages":"Article 101773"},"PeriodicalIF":2.6,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144125322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An oracle gradient regularized Newton method for quadratic measurements regression 二次测量回归的oracle梯度正则牛顿法
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-05-08 DOI: 10.1016/j.acha.2025.101775
Jun Fan , Jie Sun , Ailing Yan , Shenglong Zhou
Recovering an unknown signal from quadratic measurements has gained popularity due to its wide range of applications, including phase retrieval, fusion frame phase retrieval, and positive operator-valued measures. In this paper, we employ a least squares approach to reconstruct the signal and establish its non-asymptotic statistical properties. Our analysis shows that the estimator perfectly recovers the true signal in the noiseless case, while the error between the estimator and the true signal is bounded by O(plog(1+2n)/n) in the noisy case, where n is the number of measurements and p is the dimension of the signal. We then develop a two-phase algorithm, gradient regularized Newton method (GRNM), to solve the least squares problem. It is proven that the first phase terminates within finitely many steps, and the sequence generated in the second phase converges to a unique local minimum at a superlinear rate under certain mild conditions. Beyond these deterministic results, GRNM is capable of exactly reconstructing the true signal in the noiseless case and achieving the stated error rate with a high probability in the noisy case. Numerical experiments demonstrate that GRNM offers a high level of recovery capability and accuracy as well as fast computational speed.
从二次测量中恢复未知信号由于其广泛的应用而受到欢迎,包括相位恢复,融合帧相位恢复和正算子值测量。本文采用最小二乘方法对信号进行重构,建立了信号的非渐近统计性质。我们的分析表明,在无噪声情况下,估计器完美地恢复了真实信号,而在有噪声情况下,估计器与真实信号之间的误差以O(plog (1+2n)/n)为界,其中n是测量次数,p是信号的维数。然后,我们开发了一种两阶段算法,梯度正则化牛顿法(GRNM),以解决最小二乘问题。证明了在一定温和条件下,第一阶段终止于有限多步内,第二阶段生成的序列以超线性速度收敛到唯一的局部极小值。除了这些确定性结果之外,GRNM能够在无噪声情况下准确地重建真实信号,并在有噪声情况下以高概率达到规定的错误率。数值实验表明,该算法具有较高的恢复能力和精度,计算速度快。
{"title":"An oracle gradient regularized Newton method for quadratic measurements regression","authors":"Jun Fan ,&nbsp;Jie Sun ,&nbsp;Ailing Yan ,&nbsp;Shenglong Zhou","doi":"10.1016/j.acha.2025.101775","DOIUrl":"10.1016/j.acha.2025.101775","url":null,"abstract":"<div><div>Recovering an unknown signal from quadratic measurements has gained popularity due to its wide range of applications, including phase retrieval, fusion frame phase retrieval, and positive operator-valued measures. In this paper, we employ a least squares approach to reconstruct the signal and establish its non-asymptotic statistical properties. Our analysis shows that the estimator perfectly recovers the true signal in the noiseless case, while the error between the estimator and the true signal is bounded by <span><math><mi>O</mi><mo>(</mo><msqrt><mrow><mi>p</mi><mi>log</mi><mo>⁡</mo><mo>(</mo><mn>1</mn><mo>+</mo><mn>2</mn><mi>n</mi><mo>)</mo><mo>/</mo><mi>n</mi></mrow></msqrt><mo>)</mo></math></span> in the noisy case, where <em>n</em> is the number of measurements and <em>p</em> is the dimension of the signal. We then develop a two-phase algorithm, gradient regularized Newton method (GRNM), to solve the least squares problem. It is proven that the first phase terminates within finitely many steps, and the sequence generated in the second phase converges to a unique local minimum at a superlinear rate under certain mild conditions. Beyond these deterministic results, GRNM is capable of exactly reconstructing the true signal in the noiseless case and achieving the stated error rate with a high probability in the noisy case. Numerical experiments demonstrate that GRNM offers a high level of recovery capability and accuracy as well as fast computational speed.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"78 ","pages":"Article 101775"},"PeriodicalIF":2.6,"publicationDate":"2025-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143935916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A parameter-free two-bit covariance estimator with improved operator norm error rate 一种改进算子范数错误率的无参数二位协方差估计器
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-05-02 DOI: 10.1016/j.acha.2025.101774
Junren Chen , Michael K. Ng
A covariance matrix estimator using two bits per entry was recently developed by Dirksen et al. (2022) [11]. The estimator achieves near minimax operator norm rate for general sub-Gaussian distributions, but also suffers from two downsides: theoretically, there is an essential gap on operator norm error between their estimator and sample covariance when the diagonal of the covariance matrix is dominated by only a few entries; practically, its performance heavily relies on the dithering scale, which needs to be tuned according to some unknown parameters. In this work, we propose a new 2-bit covariance matrix estimator that simultaneously addresses both issues. Unlike the sign quantizer associated with uniform dither in Dirksen et al., we adopt a triangular dither prior to a 2-bit quantizer inspired by the multi-bit uniform quantizer. By employing dithering scales varying across entries, our estimator enjoys an improved operator norm error rate that depends on the effective rank of the underlying covariance matrix rather than the ambient dimension, which is optimal up to logarithmic factors. Moreover, our proposed method eliminates the need of any tuning parameter, as the dithering scales are entirely determined by the data. While our estimator requires a pass of all unquantized samples to determine the dithering scales, it can be adapted to the online setting where the samples arise sequentially. Experimental results are provided to demonstrate the advantages of our estimators over the existing ones.
Dirksen等人(2022)最近开发了一种协方差矩阵估计器,每个条目使用两个比特。对于一般的亚高斯分布,该估计器实现了接近极大极小算子范数率,但也存在两个缺点:理论上,当协方差矩阵的对角线仅由少数项占主导时,其估计器与样本协方差之间的算子范数误差存在本质差距;实际上,它的性能很大程度上依赖于抖动尺度,抖动尺度需要根据一些未知参数进行调整。在这项工作中,我们提出了一个新的2位协方差矩阵估计器,同时解决了这两个问题。与Dirksen等人中与均匀抖动相关的符号量化器不同,我们在受多位均匀量化器启发的2位量化器之前采用了三角形抖动。通过使用不同条目的抖动尺度,我们的估计器具有改进的算子范数错误率,该错误率取决于底层协方差矩阵的有效秩,而不是环境维度,这是最优的,直到对数因子。此外,我们提出的方法不需要任何调优参数,因为抖动尺度完全由数据决定。虽然我们的估计器需要通过所有未量化的样本来确定抖动尺度,但它可以适应样本顺序出现的在线设置。实验结果证明了我们的估计器相对于现有估计器的优越性。
{"title":"A parameter-free two-bit covariance estimator with improved operator norm error rate","authors":"Junren Chen ,&nbsp;Michael K. Ng","doi":"10.1016/j.acha.2025.101774","DOIUrl":"10.1016/j.acha.2025.101774","url":null,"abstract":"<div><div>A covariance matrix estimator using two bits per entry was recently developed by Dirksen et al. (2022) <span><span>[11]</span></span>. The estimator achieves near minimax operator norm rate for general sub-Gaussian distributions, but also suffers from two downsides: theoretically, there is an essential gap on operator norm error between their estimator and sample covariance when the diagonal of the covariance matrix is dominated by only a few entries; practically, its performance heavily relies on the dithering scale, which needs to be tuned according to some unknown parameters. In this work, we propose a new 2-bit covariance matrix estimator that simultaneously addresses both issues. Unlike the sign quantizer associated with uniform dither in Dirksen et al., we adopt a triangular dither prior to a 2-bit quantizer inspired by the multi-bit uniform quantizer. By employing dithering scales varying across entries, our estimator enjoys an improved operator norm error rate that depends on the effective rank of the underlying covariance matrix rather than the ambient dimension, which is optimal up to logarithmic factors. Moreover, our proposed method eliminates the need of <em>any</em> tuning parameter, as the dithering scales are entirely determined by the data. While our estimator requires a pass of all unquantized samples to determine the dithering scales, it can be adapted to the online setting where the samples arise sequentially. Experimental results are provided to demonstrate the advantages of our estimators over the existing ones.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"78 ","pages":"Article 101774"},"PeriodicalIF":2.6,"publicationDate":"2025-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143903641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sparsification of the regularized magnetic Laplacian with multi-type spanning forests 具有多类型跨林的正则磁拉普拉斯算子的稀疏化
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-04-28 DOI: 10.1016/j.acha.2025.101766
M. Fanuel, R. Bardenet
In this paper, we consider a U(1)-connection graph, that is, a graph where each oriented edge is endowed with a unit modulus complex number that is conjugated under orientation flip. A natural replacement for the combinatorial Laplacian is then the magnetic Laplacian, an Hermitian matrix that includes information about the graph's connection. Magnetic Laplacians appear, e.g., in the problem of angular synchronization. In the context of large and dense graphs, we study here sparsifiers of the magnetic Laplacian Δ, i.e., spectral approximations based on subgraphs with few edges. Our approach relies on sampling multi-type spanning forests (MTSFs) using a custom determinantal point process, a probability distribution over edges that favors diversity. In a word, an MTSF is a spanning subgraph whose connected components are either trees or cycle-rooted trees. The latter partially capture the angular inconsistencies of the connection graph, and thus provide a way to compress the information contained in the connection. Interestingly, when the connection graph has weakly inconsistent cycles, samples from the determinantal point process under consideration can be obtained à la Wilson, using a random walk with cycle popping. We provide statistical guarantees for a choice of natural estimators of the connection Laplacian, and investigate two practical applications of our sparsifiers: ranking with angular synchronization and graph-based semi-supervised learning. From a statistical perspective, a side result of this paper of independent interest is a matrix Chernoff bound with intrinsic dimension, which allows considering the influence of a regularization – of the form Δ+qI with q>0 – on sparsification guarantees.
本文考虑一个U(1)-连接图,即每个有向边都有一个单位模复数,该复数在有向翻转下共轭。组合拉普拉斯的自然替代品是磁拉普拉斯,一个包含图连接信息的厄米矩阵。例如,在角同步问题中出现了磁拉普拉斯算子。在大而密集图的背景下,我们研究了磁拉普拉斯Δ的稀疏化算子,即基于少边子图的谱近似。我们的方法依赖于使用自定义确定性点过程对多类型跨越森林(mtsf)进行采样,这是一种有利于多样性的边缘概率分布。简而言之,MTSF是一个生成子图,其连接的组件要么是树,要么是环根树。后者部分捕获连接图的角度不一致,从而提供一种压缩连接中包含的信息的方法。有趣的是,当连接图具有弱不一致的循环时,可以使用带有循环弹出的随机漫步,从所考虑的确定性点过程中获得样本。我们为连接拉普拉斯的自然估计量的选择提供了统计保证,并研究了我们的稀疏化器的两个实际应用:角同步排序和基于图的半监督学习。从统计的角度来看,本文的一个独立的结果是一个具有固有维数的矩阵Chernoff界,它允许考虑形式为Δ+qI with q>;0的正则化-对稀疏化保证的影响。
{"title":"Sparsification of the regularized magnetic Laplacian with multi-type spanning forests","authors":"M. Fanuel,&nbsp;R. Bardenet","doi":"10.1016/j.acha.2025.101766","DOIUrl":"10.1016/j.acha.2025.101766","url":null,"abstract":"<div><div>In this paper, we consider a <span><math><mi>U</mi><mo>(</mo><mn>1</mn><mo>)</mo></math></span>-connection graph, that is, a graph where each oriented edge is endowed with a unit modulus complex number that is conjugated under orientation flip. A natural replacement for the combinatorial Laplacian is then the <em>magnetic</em> Laplacian, an Hermitian matrix that includes information about the graph's connection. Magnetic Laplacians appear, e.g., in the problem of angular synchronization. In the context of large and dense graphs, we study here sparsifiers of the magnetic Laplacian Δ, i.e., spectral approximations based on subgraphs with few edges. Our approach relies on sampling multi-type spanning forests (MTSFs) using a custom determinantal point process, a probability distribution over edges that favors diversity. In a word, an MTSF is a spanning subgraph whose connected components are either trees or cycle-rooted trees. The latter partially capture the angular inconsistencies of the connection graph, and thus provide a way to compress the information contained in the connection. Interestingly, when the connection graph has weakly inconsistent cycles, samples from the determinantal point process under consideration can be obtained <em>à la Wilson</em>, using a random walk with cycle popping. We provide statistical guarantees for a choice of natural estimators of the connection Laplacian, and investigate two practical applications of our sparsifiers: ranking with angular synchronization and graph-based semi-supervised learning. From a statistical perspective, a side result of this paper of independent interest is a matrix Chernoff bound with intrinsic dimension, which allows considering the influence of a regularization – of the form <span><math><mi>Δ</mi><mo>+</mo><mi>q</mi><mi>I</mi></math></span> with <span><math><mi>q</mi><mo>&gt;</mo><mn>0</mn></math></span> – on sparsification guarantees.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"78 ","pages":"Article 101766"},"PeriodicalIF":2.6,"publicationDate":"2025-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143891370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Duality for neural networks through Reproducing Kernel Banach Spaces 利用核Banach空间再现神经网络的对偶性
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-03-27 DOI: 10.1016/j.acha.2025.101765
Len Spek , Tjeerd Jan Heeringa , Felix Schwenninger , Christoph Brune
Reproducing Kernel Hilbert spaces (RKHS) have been a very successful tool in various areas of machine learning. Recently, Barron spaces have been used to prove bounds on the generalisation error for neural networks. Unfortunately, Barron spaces cannot be understood in terms of RKHS due to the strong nonlinear coupling of the weights. This can be solved by using the more general Reproducing Kernel Banach spaces (RKBS). We show that these Barron spaces belong to a class of integral RKBS. This class can also be understood as an infinite union of RKHS spaces. Furthermore, we show that the dual space of such RKBSs, is again an RKBS where the roles of the data and parameters are interchanged, forming an adjoint pair of RKBSs including a reproducing kernel. This allows us to construct the saddle point problem for neural networks, which can be used in the whole field of primal-dual optimisation.
再现核希尔伯特空间(RKHS)在机器学习的各个领域都是一个非常成功的工具。近年来,巴伦空间被用来证明神经网络泛化误差的界。不幸的是,由于权重的强非线性耦合,不能用RKHS来理解巴伦空间。这可以通过使用更通用的rereproduction Kernel Banach spaces (RKBS)来解决。我们证明了这些Barron空间属于一类积分RKBS。该类也可以理解为RKHS空间的无限并。此外,我们证明了这样的RKBS的对偶空间再次是一个RKBS,其中数据和参数的角色是互换的,形成了一个包含再现核的RKBS的伴随对。这允许我们构造神经网络的鞍点问题,它可以用于整个原始对偶优化领域。
{"title":"Duality for neural networks through Reproducing Kernel Banach Spaces","authors":"Len Spek ,&nbsp;Tjeerd Jan Heeringa ,&nbsp;Felix Schwenninger ,&nbsp;Christoph Brune","doi":"10.1016/j.acha.2025.101765","DOIUrl":"10.1016/j.acha.2025.101765","url":null,"abstract":"<div><div>Reproducing Kernel Hilbert spaces (RKHS) have been a very successful tool in various areas of machine learning. Recently, Barron spaces have been used to prove bounds on the generalisation error for neural networks. Unfortunately, Barron spaces cannot be understood in terms of RKHS due to the strong nonlinear coupling of the weights. This can be solved by using the more general Reproducing Kernel Banach spaces (RKBS). We show that these Barron spaces belong to a class of integral RKBS. This class can also be understood as an infinite union of RKHS spaces. Furthermore, we show that the dual space of such RKBSs, is again an RKBS where the roles of the data and parameters are interchanged, forming an adjoint pair of RKBSs including a reproducing kernel. This allows us to construct the saddle point problem for neural networks, which can be used in the whole field of primal-dual optimisation.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"78 ","pages":"Article 101765"},"PeriodicalIF":2.6,"publicationDate":"2025-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143869805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Controlled learning of pointwise nonlinearities in neural-network-like architectures 类神经网络结构中点非线性的受控学习
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-03-25 DOI: 10.1016/j.acha.2025.101764
Michael Unser, Alexis Goujon, Stanislas Ducotterd
We present a general variational framework for the training of freeform nonlinearities in layered computational architectures subject to some slope constraints. The regularization that we add to the traditional training loss penalizes the second-order total variation of each trainable activation. The slope constraints allow us to impose properties such as 1-Lipschitz stability, firm non-expansiveness, and monotonicity/invertibility. These properties are crucial to ensure the proper functioning of certain classes of signal-processing algorithms (e.g., plug-and-play schemes, unrolled proximal gradient, invertible flows). We prove that the global optimum of the stated constrained-optimization problem is achieved with nonlinearities that are adaptive nonuniform linear splines. We then show how to solve the resulting function-optimization problem numerically by representing the nonlinearities in a suitable (nonuniform) B-spline basis. Finally, we illustrate the use of our framework with the data-driven design of (weakly) convex regularizers for the denoising of images and the resolution of inverse problems.
我们提出了一个通用的变分框架,用于在受某些斜率约束的分层计算架构中训练自由形式非线性。我们加入传统训练损失的正则化惩罚了每个可训练激活的二阶总变化。斜率约束允许我们施加诸如1-Lipschitz稳定性,坚固非扩张性和单调性/可逆性等性质。这些特性对于确保某些类型的信号处理算法(例如,即插即用方案、展开的近端梯度、可逆流)的正常运行至关重要。证明了所述约束优化问题的全局最优解是用自适应非均匀线性样条实现的。然后,我们展示了如何通过在合适的(非均匀的)b样条基中表示非线性来解决结果函数优化问题。最后,我们用数据驱动的(弱)凸正则化设计来说明我们的框架在图像去噪和反问题解决中的应用。
{"title":"Controlled learning of pointwise nonlinearities in neural-network-like architectures","authors":"Michael Unser,&nbsp;Alexis Goujon,&nbsp;Stanislas Ducotterd","doi":"10.1016/j.acha.2025.101764","DOIUrl":"10.1016/j.acha.2025.101764","url":null,"abstract":"<div><div>We present a general variational framework for the training of freeform nonlinearities in layered computational architectures subject to some slope constraints. The regularization that we add to the traditional training loss penalizes the second-order total variation of each trainable activation. The slope constraints allow us to impose properties such as 1-Lipschitz stability, firm non-expansiveness, and monotonicity/invertibility. These properties are crucial to ensure the proper functioning of certain classes of signal-processing algorithms (e.g., plug-and-play schemes, unrolled proximal gradient, invertible flows). We prove that the global optimum of the stated constrained-optimization problem is achieved with nonlinearities that are adaptive nonuniform linear splines. We then show how to solve the resulting function-optimization problem numerically by representing the nonlinearities in a suitable (nonuniform) B-spline basis. Finally, we illustrate the use of our framework with the data-driven design of (weakly) convex regularizers for the denoising of images and the resolution of inverse problems.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"77 ","pages":"Article 101764"},"PeriodicalIF":2.6,"publicationDate":"2025-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143738438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mathematical algorithm design for deep learning under societal and judicial constraints: The algorithmic transparency requirement 社会和司法约束下深度学习的数学算法设计:算法透明度要求
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-03-24 DOI: 10.1016/j.acha.2025.101763
Holger Boche , Adalbert Fono , Gitta Kutyniok
Deep learning still has drawbacks regarding trustworthiness, which describes a comprehensible, fair, safe, and reliable method. To mitigate the potential risk of AI, clear obligations associated with trustworthiness have been proposed via regulatory guidelines, e.g., in the European AI Act. Therefore, a central question is to what extent trustworthy deep learning can be realized. Establishing the described properties constituting trustworthiness requires that the factors influencing an algorithmic computation can be retraced, i.e., the algorithmic implementation is transparent. Motivated by the observation that the current evolution of deep learning models necessitates a change in computing technology, we derive a mathematical framework that enables us to analyze whether a transparent implementation in a computing model is feasible. The core idea is to formalize and subsequently relate the properties of a transparent algorithmic implementation to the mathematical model of the computing platform, thereby establishing verifiable criteria.
We exemplarily apply our trustworthiness framework to analyze deep learning approaches for inverse problems in digital and analog computing models represented by Turing and Blum-Shub-Smale machines, respectively. Based on previous results, we find that Blum-Shub-Smale machines have the potential to establish trustworthy solvers for inverse problems under fairly general conditions, whereas Turing machines cannot guarantee trustworthiness to the same degree.
深度学习在可信度方面仍然存在缺陷,它描述了一种可理解、公平、安全、可靠的方法。为了减轻人工智能的潜在风险,已经通过监管指南(例如欧洲人工智能法案)提出了与可信度相关的明确义务。因此,一个核心问题是在多大程度上可以实现可信赖的深度学习。建立所描述的构成可信度的属性要求影响算法计算的因素可以追溯,即算法实现是透明的。由于观察到当前深度学习模型的发展需要改变计算技术,我们推导了一个数学框架,使我们能够分析计算模型中的透明实现是否可行。核心思想是形式化并随后将透明算法实现的属性与计算平台的数学模型联系起来,从而建立可验证的标准。例如,我们应用我们的可信度框架来分析数字和模拟计算模型中逆问题的深度学习方法,分别以图灵和Blum-Shub-Smale机器为代表。基于之前的结果,我们发现Blum-Shub-Smale机器有潜力在相当一般的条件下为逆问题建立可信解,而图灵机不能保证相同程度的可信度。
{"title":"Mathematical algorithm design for deep learning under societal and judicial constraints: The algorithmic transparency requirement","authors":"Holger Boche ,&nbsp;Adalbert Fono ,&nbsp;Gitta Kutyniok","doi":"10.1016/j.acha.2025.101763","DOIUrl":"10.1016/j.acha.2025.101763","url":null,"abstract":"<div><div>Deep learning still has drawbacks regarding trustworthiness, which describes a comprehensible, fair, safe, and reliable method. To mitigate the potential risk of AI, clear obligations associated with trustworthiness have been proposed via regulatory guidelines, e.g., in the European AI Act. Therefore, a central question is to what extent trustworthy deep learning can be realized. Establishing the described properties constituting trustworthiness requires that the factors influencing an algorithmic computation can be retraced, i.e., the algorithmic implementation is transparent. Motivated by the observation that the current evolution of deep learning models necessitates a change in computing technology, we derive a mathematical framework that enables us to analyze whether a transparent implementation in a computing model is feasible. The core idea is to formalize and subsequently relate the properties of a transparent algorithmic implementation to the mathematical model of the computing platform, thereby establishing verifiable criteria.</div><div>We exemplarily apply our trustworthiness framework to analyze deep learning approaches for inverse problems in digital and analog computing models represented by Turing and Blum-Shub-Smale machines, respectively. Based on previous results, we find that Blum-Shub-Smale machines have the potential to establish trustworthy solvers for inverse problems under fairly general conditions, whereas Turing machines cannot guarantee trustworthiness to the same degree.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"77 ","pages":"Article 101763"},"PeriodicalIF":2.6,"publicationDate":"2025-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143696911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied and Computational Harmonic Analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1