首页 > 最新文献

Applied and Computational Harmonic Analysis最新文献

英文 中文
The Large Deviation Principle for W-random spectral measures
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-02-21 DOI: 10.1016/j.acha.2025.101756
Mahya Ghandehari , Georgi S. Medvedev
The W-random graphs provide a flexible framework for modeling large random networks. Using the Large Deviation Principle (LDP) for W-random graphs from [19], we prove the LDP for the corresponding class of random symmetric Hilbert-Schmidt integral operators. Our main result describes how the eigenvalues and the eigenspaces of the integral operator are affected by large deviations in the underlying random graphon. To prove the LDP, we demonstrate continuous dependence of the spectral measures associated with integral operators on the corresponding graphons and use the Contraction Principle. To illustrate our results, we obtain leading order asymptotics of the eigenvalues of small-world and bipartite random graphs conditioned on atypical edge counts. These examples suggest several representative scenarios of how the eigenvalues and the eigenspaces are affected by large deviations. We discuss the implications of these observations for bifurcation analysis of Dynamical Systems and Graph Signal Processing.
W 随机图为大型随机网络建模提供了一个灵活的框架。利用文献[19]中针对 W-随机图的大偏差原理(LDP),我们证明了相应类别的随机对称希尔伯特-施密特积分算子的大偏差原理。我们的主要结果描述了积分算子的特征值和特征空间如何受到底层随机图元大偏差的影响。为了证明 LDP,我们证明了与积分算子相关的谱度量对相应图元的连续依赖性,并使用了收缩原理。为了说明我们的结果,我们获得了以非典型边数为条件的小世界和双方形随机图特征值的前阶渐近性。这些例子说明了特征值和特征空间如何受到大偏差的影响。我们将讨论这些观察结果对动态系统分岔分析和图信号处理的影响。
{"title":"The Large Deviation Principle for W-random spectral measures","authors":"Mahya Ghandehari ,&nbsp;Georgi S. Medvedev","doi":"10.1016/j.acha.2025.101756","DOIUrl":"10.1016/j.acha.2025.101756","url":null,"abstract":"<div><div>The <em>W</em>-random graphs provide a flexible framework for modeling large random networks. Using the Large Deviation Principle (LDP) for <em>W</em>-random graphs from <span><span>[19]</span></span>, we prove the LDP for the corresponding class of random symmetric Hilbert-Schmidt integral operators. Our main result describes how the eigenvalues and the eigenspaces of the integral operator are affected by large deviations in the underlying random graphon. To prove the LDP, we demonstrate continuous dependence of the spectral measures associated with integral operators on the corresponding graphons and use the Contraction Principle. To illustrate our results, we obtain leading order asymptotics of the eigenvalues of small-world and bipartite random graphs conditioned on atypical edge counts. These examples suggest several representative scenarios of how the eigenvalues and the eigenspaces are affected by large deviations. We discuss the implications of these observations for bifurcation analysis of Dynamical Systems and Graph Signal Processing.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"77 ","pages":"Article 101756"},"PeriodicalIF":2.6,"publicationDate":"2025-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143479228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient identification of wide shallow neural networks with biases
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-02-17 DOI: 10.1016/j.acha.2025.101749
Massimo Fornasier , Timo Klock , Marco Mondelli , Michael Rauchensteiner
The identification of the parameters of a neural network from finite samples of input-output pairs is often referred to as the teacher-student model, and this model has represented a popular framework for understanding training and generalization. Even if the problem is NP-complete in the worst case, a rapidly growing literature – after adding suitable distributional assumptions – has established finite sample identification of two-layer networks with a number of neurons m=O(D), D being the input dimension. For the range D<m<D2 the problem becomes harder, and truly little is known for networks parametrized by biases as well. This paper fills the gap by providing efficient algorithms and rigorous theoretical guarantees of finite sample identification for such wider shallow networks with biases. Our approach is based on a two-step pipeline: first, we recover the direction of the weights, by exploiting second order information; next, we identify the signs by suitable algebraic evaluations, and we recover the biases by empirical risk minimization via gradient descent. Numerical results demonstrate the effectiveness of our approach.
{"title":"Efficient identification of wide shallow neural networks with biases","authors":"Massimo Fornasier ,&nbsp;Timo Klock ,&nbsp;Marco Mondelli ,&nbsp;Michael Rauchensteiner","doi":"10.1016/j.acha.2025.101749","DOIUrl":"10.1016/j.acha.2025.101749","url":null,"abstract":"<div><div>The identification of the parameters of a neural network from finite samples of input-output pairs is often referred to as the <em>teacher-student model</em>, and this model has represented a popular framework for understanding training and generalization. Even if the problem is NP-complete in the worst case, a rapidly growing literature – after adding suitable distributional assumptions – has established finite sample identification of two-layer networks with a number of neurons <span><math><mi>m</mi><mo>=</mo><mi>O</mi><mo>(</mo><mi>D</mi><mo>)</mo></math></span>, <em>D</em> being the input dimension. For the range <span><math><mi>D</mi><mo>&lt;</mo><mi>m</mi><mo>&lt;</mo><msup><mrow><mi>D</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> the problem becomes harder, and truly little is known for networks parametrized by biases as well. This paper fills the gap by providing efficient algorithms and rigorous theoretical guarantees of finite sample identification for such wider shallow networks with biases. Our approach is based on a two-step pipeline: first, we recover the direction of the weights, by exploiting second order information; next, we identify the signs by suitable algebraic evaluations, and we recover the biases by empirical risk minimization via gradient descent. Numerical results demonstrate the effectiveness of our approach.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"77 ","pages":"Article 101749"},"PeriodicalIF":2.6,"publicationDate":"2025-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143429952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Kadec-type theorems for sampled group orbits 抽样群轨道的kadec型定理
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-01-10 DOI: 10.1016/j.acha.2025.101748
Ilya Krishtal, Brendan Miller
We extend the classical Kadec 14 theorem for systems of exponential functions on an interval to frames and atomic decompositions formed by sampling an orbit of a vector under an isometric group representation.
我们将区间上指数函数系统的经典 Kadec 14 定理扩展到等距群表示下的向量轨道采样所形成的框架和原子分解。
{"title":"Kadec-type theorems for sampled group orbits","authors":"Ilya Krishtal,&nbsp;Brendan Miller","doi":"10.1016/j.acha.2025.101748","DOIUrl":"10.1016/j.acha.2025.101748","url":null,"abstract":"<div><div>We extend the classical Kadec <span><math><mfrac><mrow><mn>1</mn></mrow><mrow><mn>4</mn></mrow></mfrac></math></span> theorem for systems of exponential functions on an interval to frames and atomic decompositions formed by sampling an orbit of a vector under an isometric group representation.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"76 ","pages":"Article 101748"},"PeriodicalIF":2.6,"publicationDate":"2025-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142990415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the non-frame property of Gabor systems with Hermite generators and the frame set conjecture 带有Hermite发生器的Gabor系统的非框架性质及框架集猜想
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2025-01-10 DOI: 10.1016/j.acha.2025.101747
Andreas Horst , Jakob Lemvig , Allan Erlang Videbæk
The frame set conjecture for Hermite functions formulated in [13] states that the Gabor frame set for these generators is the largest possible, that is, the time-frequency shifts of the Hermite functions associated with sampling rates α and modulation rates β that avoid all known obstructions lead to Gabor frames for L2(R). By results in [24], [25] and [22], it is known that the conjecture is true for the Gaussian, the 0th order Hermite functions, and false for Hermite functions of order 2,3,6,7,10,11,, respectively. In this paper we disprove the remaining cases except for the 1st order Hermite function.
在[13]中表述的Hermite函数的帧集猜想表明,这些发生器的Gabor帧集是可能最大的,也就是说,与采样率α和调制率β相关的Hermite函数的时频移可以避免所有已知的障碍,从而导致L2(R)的Gabor帧。由[24,25]和[22]的结果可知,该猜想对高斯、0阶Hermite函数为真,对2、3、6、7、10、11、…阶Hermite函数为假。本文证明了除一阶Hermite函数外的其他情况。
{"title":"On the non-frame property of Gabor systems with Hermite generators and the frame set conjecture","authors":"Andreas Horst ,&nbsp;Jakob Lemvig ,&nbsp;Allan Erlang Videbæk","doi":"10.1016/j.acha.2025.101747","DOIUrl":"10.1016/j.acha.2025.101747","url":null,"abstract":"<div><div>The frame set conjecture for Hermite functions formulated in <span><span>[13]</span></span> states that the Gabor frame set for these generators is the largest possible, that is, the time-frequency shifts of the Hermite functions associated with sampling rates <em>α</em> and modulation rates <em>β</em> that avoid all known obstructions lead to Gabor frames for <span><math><msup><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msup><mo>(</mo><mi>R</mi><mo>)</mo></math></span>. By results in <span><span>[24]</span></span>, <span><span>[25]</span></span> and <span><span>[22]</span></span>, it is known that the conjecture is true for the Gaussian, the 0th order Hermite functions, and false for Hermite functions of order <span><math><mn>2</mn><mo>,</mo><mn>3</mn><mo>,</mo><mn>6</mn><mo>,</mo><mn>7</mn><mo>,</mo><mn>10</mn><mo>,</mo><mn>11</mn><mo>,</mo><mo>…</mo></math></span>, respectively. In this paper we disprove the remaining cases <em>except</em> for the 1st order Hermite function.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"76 ","pages":"Article 101747"},"PeriodicalIF":2.6,"publicationDate":"2025-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142990416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
How robust is randomized blind deconvolution via nuclear norm minimization against adversarial noise? 通过核范数最小化随机盲反卷积对对抗噪声的鲁棒性如何?
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-12-30 DOI: 10.1016/j.acha.2024.101746
Julia Kostin , Felix Krahmer , Dominik Stöger
In this paper, we study the problem of recovering two unknown signals from their convolution, which is commonly referred to as blind deconvolution. Reformulation of blind deconvolution as a low-rank recovery problem has led to multiple theoretical recovery guarantees in the past decade due to the success of the nuclear norm minimization heuristic. In particular, in the absence of noise, exact recovery has been established for sufficiently incoherent signals contained in lower-dimensional subspaces. However, if the convolution is corrupted by additive bounded noise, the stability of the recovery problem remains much less understood. In particular, existing reconstruction bounds involve large dimension factors and therefore fail to explain the empirical evidence for dimension-independent robustness of nuclear norm minimization. Recently, theoretical evidence has emerged for ill-posed behaviour of low-rank matrix recovery for sufficiently small noise levels. In this work, we develop improved recovery guarantees for blind deconvolution with adversarial noise which exhibit square-root scaling in the noise level. Hence, our results are consistent with existing counterexamples which speak against linear scaling in the noise level as demonstrated for related low-rank matrix recovery problems.
本文研究了从卷积中恢复两个未知信号的问题,这通常被称为盲反卷积。在过去的十年中,由于核范数最小化启发式的成功,盲反卷积作为一个低秩恢复问题的重新表述已经导致了多个理论上的恢复保证。特别是,在没有噪声的情况下,对于包含在低维子空间中的充分不相干的信号,已经建立了精确的恢复。然而,如果卷积被加性有界噪声破坏,恢复问题的稳定性仍然很少被理解。特别是,现有的重建边界涉及大维度因素,因此无法解释核范数最小化的维无关鲁棒性的经验证据。最近,理论证据已经出现了低秩矩阵恢复的病态行为足够小的噪声水平。在这项工作中,我们开发了具有对抗性噪声的盲反卷积的改进恢复保证,该噪声在噪声水平上表现为平方根缩放。因此,我们的结果与现有的反例一致,这些反例反对噪声水平的线性缩放,如相关的低秩矩阵恢复问题所示。
{"title":"How robust is randomized blind deconvolution via nuclear norm minimization against adversarial noise?","authors":"Julia Kostin ,&nbsp;Felix Krahmer ,&nbsp;Dominik Stöger","doi":"10.1016/j.acha.2024.101746","DOIUrl":"10.1016/j.acha.2024.101746","url":null,"abstract":"<div><div>In this paper, we study the problem of recovering two unknown signals from their convolution, which is commonly referred to as blind deconvolution. Reformulation of blind deconvolution as a low-rank recovery problem has led to multiple theoretical recovery guarantees in the past decade due to the success of the nuclear norm minimization heuristic. In particular, in the absence of noise, exact recovery has been established for sufficiently incoherent signals contained in lower-dimensional subspaces. However, if the convolution is corrupted by additive bounded noise, the stability of the recovery problem remains much less understood. In particular, existing reconstruction bounds involve large dimension factors and therefore fail to explain the empirical evidence for dimension-independent robustness of nuclear norm minimization. Recently, theoretical evidence has emerged for ill-posed behaviour of low-rank matrix recovery for sufficiently small noise levels. In this work, we develop improved recovery guarantees for blind deconvolution with adversarial noise which exhibit square-root scaling in the noise level. Hence, our results are consistent with existing counterexamples which speak against linear scaling in the noise level as demonstrated for related low-rank matrix recovery problems.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"76 ","pages":"Article 101746"},"PeriodicalIF":2.6,"publicationDate":"2024-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142968152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal rates for functional linear regression with general regularization 一般正则化函数线性回归的最优率
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-12-17 DOI: 10.1016/j.acha.2024.101745
Naveen Gupta , S. Sivananthan , Bharath K. Sriperumbudur
Functional linear regression is one of the fundamental and well-studied methods in functional data analysis. In this work, we investigate the functional linear regression model within the context of reproducing kernel Hilbert space by employing general spectral regularization to approximate the slope function with certain smoothness assumptions. We establish optimal convergence rates for estimation and prediction errors associated with the proposed method under Hölder type source condition, which generalizes and sharpens all the known results in the literature.
函数线性回归是函数数据分析中最基本、研究最充分的方法之一。在此工作中,我们研究了在核希尔伯特空间再现背景下的函数线性回归模型,采用一般谱正则化方法在一定的平滑假设下近似斜率函数。我们建立了在Hölder类型源条件下与所提出方法相关的估计和预测误差的最优收敛率,它推广和锐化了文献中所有已知的结果。
{"title":"Optimal rates for functional linear regression with general regularization","authors":"Naveen Gupta ,&nbsp;S. Sivananthan ,&nbsp;Bharath K. Sriperumbudur","doi":"10.1016/j.acha.2024.101745","DOIUrl":"10.1016/j.acha.2024.101745","url":null,"abstract":"<div><div>Functional linear regression is one of the fundamental and well-studied methods in functional data analysis. In this work, we investigate the functional linear regression model within the context of reproducing kernel Hilbert space by employing general spectral regularization to approximate the slope function with certain smoothness assumptions. We establish optimal convergence rates for estimation and prediction errors associated with the proposed method under Hölder type source condition, which generalizes and sharpens all the known results in the literature.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"76 ","pages":"Article 101745"},"PeriodicalIF":2.6,"publicationDate":"2024-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142874073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty principles, restriction, Bourgain's Λq theorem, and signal recovery 不确定性原理,限制,布尔甘Λq定理,和信号恢复
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-12-09 DOI: 10.1016/j.acha.2024.101734
A. Iosevich , A. Mayeli
Let G be a finite abelian group. Let f:GC be a signal (i.e. function). The classical uncertainty principle asserts that the product of the size of the support of f and its Fourier transform fˆ, supp(f) and supp(fˆ) respectively, must satisfy the condition:|supp(f)||supp(fˆ)||G|.
In the first part of this paper, we improve the uncertainty principle for signals with Fourier transform supported on generic sets. This improvement is achieved by employing the restriction theory, including Bourgain celebrate result on Λq-sets, and the Salem set mechanism from harmonic analysis. Then we investigate some applications of uncertainty principles that were developed in the first part of this paper, to the problem of unique recovery of finite sparse signals in the absence of some frequencies.
Donoho and Stark ([14]), and, independently, Matolcsi and Szucs ([33]) showed that a signal of length N can be recovered exactly, even if some of the frequencies are unobserved, provided that the product of the size of the number of non-zero entries of the signal and the number of missing frequencies is not too large, leveraging the classical uncertainty principle for vectors. Our results broaden the scope for a natural class of signals in higher-dimensional spaces. In the case when the signal is binary, we provide a very simple exact recovery mechanism through the DRA algorithm.
设G是一个有限阿贝尔群。设f:G→C为信号(即函数)。经典的不确定性原理断言,f的支持大小与其傅里叶变换f的乘积,分别为supp(f)和supp(f),必须满足条件:|supp(f)|·|supp(f)|≥|G|。
{"title":"Uncertainty principles, restriction, Bourgain's Λq theorem, and signal recovery","authors":"A. Iosevich ,&nbsp;A. Mayeli","doi":"10.1016/j.acha.2024.101734","DOIUrl":"10.1016/j.acha.2024.101734","url":null,"abstract":"<div><div>Let <em>G</em> be a finite abelian group. Let <span><math><mi>f</mi><mo>:</mo><mi>G</mi><mo>→</mo><mi>C</mi></math></span> be a signal (i.e. function). The classical uncertainty principle asserts that the product of the size of the support of <em>f</em> and its Fourier transform <span><math><mover><mrow><mi>f</mi></mrow><mrow><mo>ˆ</mo></mrow></mover></math></span>, <span><math><mtext>supp</mtext><mo>(</mo><mi>f</mi><mo>)</mo></math></span> and <span><math><mtext>supp</mtext><mo>(</mo><mover><mrow><mi>f</mi></mrow><mrow><mo>ˆ</mo></mrow></mover><mo>)</mo></math></span> respectively, must satisfy the condition:<span><span><span><math><mo>|</mo><mtext>supp</mtext><mo>(</mo><mi>f</mi><mo>)</mo><mo>|</mo><mo>⋅</mo><mo>|</mo><mtext>supp</mtext><mo>(</mo><mover><mrow><mi>f</mi></mrow><mrow><mo>ˆ</mo></mrow></mover><mo>)</mo><mo>|</mo><mo>≥</mo><mo>|</mo><mi>G</mi><mo>|</mo><mo>.</mo></math></span></span></span></div><div>In the first part of this paper, we improve the uncertainty principle for signals with Fourier transform supported on generic sets. This improvement is achieved by employing <em>the restriction theory</em>, including Bourgain celebrate result on <span><math><msub><mrow><mi>Λ</mi></mrow><mrow><mi>q</mi></mrow></msub></math></span>-sets, and <em>the Salem set</em> mechanism from harmonic analysis. Then we investigate some applications of uncertainty principles that were developed in the first part of this paper, to the problem of unique recovery of finite sparse signals in the absence of some frequencies.</div><div>Donoho and Stark (<span><span>[14]</span></span>), and, independently, Matolcsi and Szucs (<span><span>[33]</span></span>) showed that a signal of length <em>N</em> can be recovered exactly, even if some of the frequencies are unobserved, provided that the product of the size of the number of non-zero entries of the signal and the number of missing frequencies is not too large, leveraging the classical uncertainty principle for vectors. Our results broaden the scope for a natural class of signals in higher-dimensional spaces. In the case when the signal is binary, we provide a very simple exact recovery mechanism through the DRA algorithm.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"76 ","pages":"Article 101734"},"PeriodicalIF":2.6,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142874074","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tikhonov regularization for Gaussian empirical gain maximization in RKHS is consistent RKHS 中高斯经验增益最大化的 Tikhonov 正则化是一致的
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-12-09 DOI: 10.1016/j.acha.2024.101735
Yunlong Feng , Qiang Wu
Without imposing light-tailed noise assumptions, we prove that Tikhonov regularization for Gaussian Empirical Gain Maximization (EGM) in a reproducing kernel Hilbert space is consistent and further establish its fast exponential type convergence rates. In the literature, Gaussian EGM was proposed in various contexts to tackle robust estimation problems and has been applied extensively in a great variety of real-world applications. A reproducing kernel Hilbert space is frequently chosen as the hypothesis space, and Tikhonov regularization plays a crucial role in model selection. Although Gaussian EGM has been studied theoretically in a series of papers recently and has been well-understood, theoretical understanding of its Tikhonov regularized variants in RKHS is still limited. Several fundamental challenges remain, especially when light-tailed noise assumptions are absent. To fill the gap and address these challenges, we conduct the present study and make the following contributions. First, under weak moment conditions, we establish a new comparison theorem that enables the investigation of the asymptotic mean calibration properties of regularized Gaussian EGM. Second, under the same weak moment conditions, we show that regularized Gaussian EGM estimators are consistent and further establish their fast exponential-type convergence rates. Our study justifies its feasibility in tackling robust regression problems and explains its robustness from a theoretical viewpoint. Moreover, new technical tools including probabilistic initial upper bounds, confined effective hypothesis spaces, and novel comparison theorems are introduced and developed, which can faciliate the analysis of general regularized empirical gain maximization schemes that fall into the same vein as regularized Gaussian EGM.
在不施加轻尾噪声假设的情况下,证明了再现核Hilbert空间中高斯经验增益最大化(EGM)的Tikhonov正则化是一致的,并进一步建立了其快速指数型收敛速率。在文献中,高斯EGM在各种情况下被提出来解决鲁棒估计问题,并已广泛应用于各种实际应用中。假设空间通常选择再现核希尔伯特空间,吉洪诺夫正则化在模型选择中起着至关重要的作用。虽然最近有一系列论文从理论上对高斯EGM进行了研究,并得到了很好的理解,但对其在RKHS中的Tikhonov正则化变体的理论认识仍然有限。几个基本的挑战仍然存在,特别是在没有光尾噪声假设的情况下。为了填补空白和应对这些挑战,我们进行了本研究,并做出了以下贡献。首先,在弱矩条件下,我们建立了一个新的比较定理,用于研究正则化高斯方程的渐近均值校准性质。其次,在相同的弱矩条件下,我们证明了正则化高斯EGM估计是一致的,并进一步建立了它们的快速指数型收敛速率。我们的研究证明了它在解决鲁棒回归问题上的可行性,并从理论角度解释了它的鲁棒性。此外,引入和发展了新的技术工具,包括概率初始上界、有限有效假设空间和新的比较定理,这些工具可以促进与正则化高斯EGM相同的一般正则化经验增益最大化方案的分析。
{"title":"Tikhonov regularization for Gaussian empirical gain maximization in RKHS is consistent","authors":"Yunlong Feng ,&nbsp;Qiang Wu","doi":"10.1016/j.acha.2024.101735","DOIUrl":"10.1016/j.acha.2024.101735","url":null,"abstract":"<div><div>Without imposing light-tailed noise assumptions, we prove that Tikhonov regularization for Gaussian Empirical Gain Maximization (EGM) in a reproducing kernel Hilbert space is consistent and further establish its fast exponential type convergence rates. In the literature, Gaussian EGM was proposed in various contexts to tackle robust estimation problems and has been applied extensively in a great variety of real-world applications. A reproducing kernel Hilbert space is frequently chosen as the hypothesis space, and Tikhonov regularization plays a crucial role in model selection. Although Gaussian EGM has been studied theoretically in a series of papers recently and has been well-understood, theoretical understanding of its Tikhonov regularized variants in RKHS is still limited. Several fundamental challenges remain, especially when light-tailed noise assumptions are absent. To fill the gap and address these challenges, we conduct the present study and make the following contributions. First, under weak moment conditions, we establish a new comparison theorem that enables the investigation of the asymptotic mean calibration properties of regularized Gaussian EGM. Second, under the same weak moment conditions, we show that regularized Gaussian EGM estimators are consistent and further establish their fast exponential-type convergence rates. Our study justifies its feasibility in tackling robust regression problems and explains its robustness from a theoretical viewpoint. Moreover, new technical tools including probabilistic initial upper bounds, confined effective hypothesis spaces, and novel comparison theorems are introduced and developed, which can faciliate the analysis of general regularized empirical gain maximization schemes that fall into the same vein as regularized Gaussian EGM.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"76 ","pages":"Article 101735"},"PeriodicalIF":2.6,"publicationDate":"2024-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142823231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Injectivity of ReLU networks: Perspectives from statistical physics ReLU网络的注入性:来自统计物理学的观点
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-12-03 DOI: 10.1016/j.acha.2024.101736
Antoine Maillard , Afonso S. Bandeira , David Belius , Ivan Dokmanić , Shuta Nakajima
When can the input of a ReLU neural network be inferred from its output? In other words, when is the network injective? We consider a single layer, xReLU(Wx), with a random Gaussian m×n matrix W, in a high-dimensional setting where n,m. Recent work connects this problem to spherical integral geometry giving rise to a conjectured sharp injectivity threshold for α=m/n by studying the expected Euler characteristic of a certain random set. We adopt a different perspective and show that injectivity is equivalent to a property of the ground state of the spherical perceptron, an important spin glass model in statistical physics. By leveraging the (non-rigorous) replica symmetry-breaking theory, we derive analytical equations for the threshold whose solution is at odds with that from the Euler characteristic. Furthermore, we use Gordon's min–max theorem to prove that a replica-symmetric upper bound refutes the Euler characteristic prediction. Along the way we aim to give a tutorial-style introduction to key ideas from statistical physics in an effort to make the exposition accessible to a broad audience. Our analysis establishes a connection between spin glasses and integral geometry but leaves open the problem of explaining the discrepancies.
何时可以从ReLU神经网络的输出推断其输入?换句话说,什么时候网络是注入的?我们考虑在n,m→∞的高维环境下,具有随机高斯m×n矩阵W的单层x, ReLU(Wx)。最近的工作将这一问题与球面积分几何联系起来,通过研究某随机集的期望欧拉特性,提出了α=m/n的猜想尖锐注入阈值。我们采用了不同的视角,并证明了注入性相当于球形感知器基态的一个性质,球面感知器是统计物理中一个重要的自旋玻璃模型。通过利用(非严格的)副本对称破缺理论,我们推导出阈值的解析方程,其解与欧拉特性的解不一致。此外,我们用Gordon的最小-最大定理证明了一个复制对称上界驳斥了欧拉特征预测。在此过程中,我们的目标是对统计物理学的关键思想进行教程式的介绍,以使广泛的受众能够访问该博览会。我们的分析建立了自旋玻璃和积分几何之间的联系,但没有解释这些差异的问题。
{"title":"Injectivity of ReLU networks: Perspectives from statistical physics","authors":"Antoine Maillard ,&nbsp;Afonso S. Bandeira ,&nbsp;David Belius ,&nbsp;Ivan Dokmanić ,&nbsp;Shuta Nakajima","doi":"10.1016/j.acha.2024.101736","DOIUrl":"10.1016/j.acha.2024.101736","url":null,"abstract":"<div><div>When can the input of a ReLU neural network be inferred from its output? In other words, when is the network injective? We consider a single layer, <span><math><mi>x</mi><mo>↦</mo><mrow><mi>ReLU</mi></mrow><mo>(</mo><mi>W</mi><mi>x</mi><mo>)</mo></math></span>, with a random Gaussian <span><math><mi>m</mi><mo>×</mo><mi>n</mi></math></span> matrix <em>W</em>, in a high-dimensional setting where <span><math><mi>n</mi><mo>,</mo><mi>m</mi><mo>→</mo><mo>∞</mo></math></span>. Recent work connects this problem to spherical integral geometry giving rise to a conjectured sharp injectivity threshold for <span><math><mi>α</mi><mo>=</mo><mi>m</mi><mo>/</mo><mi>n</mi></math></span> by studying the expected Euler characteristic of a certain random set. We adopt a different perspective and show that injectivity is equivalent to a property of the ground state of the spherical perceptron, an important spin glass model in statistical physics. By leveraging the (non-rigorous) replica symmetry-breaking theory, we derive analytical equations for the threshold whose solution is at odds with that from the Euler characteristic. Furthermore, we use Gordon's min–max theorem to prove that a replica-symmetric upper bound refutes the Euler characteristic prediction. Along the way we aim to give a tutorial-style introduction to key ideas from statistical physics in an effort to make the exposition accessible to a broad audience. Our analysis establishes a connection between spin glasses and integral geometry but leaves open the problem of explaining the discrepancies.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"76 ","pages":"Article 101736"},"PeriodicalIF":2.6,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142790150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Group projected subspace pursuit for block sparse signal reconstruction: Convergence analysis and applications 1 块稀疏信号重构中的群投影子空间追踪:收敛性分析及应用
IF 2.6 2区 数学 Q1 MATHEMATICS, APPLIED Pub Date : 2024-11-28 DOI: 10.1016/j.acha.2024.101726
Roy Y. He , Haixia Liu , Hao Liu
In this paper, we present a convergence analysis of the Group Projected Subspace Pursuit (GPSP) algorithm proposed by He et al. [26] (Group Projected subspace pursuit for IDENTification of variable coefficient differential equations (GP-IDENT), Journal of Computational Physics, 494, 112526) and extend its application to general tasks of block sparse signal recovery. Given an observation y and sampling matrix A, we focus on minimizing the approximation error Acy22 with respect to the signal c with block sparsity constraints. We prove that when the sampling matrix A satisfies the Block Restricted Isometry Property (BRIP) with a sufficiently small Block Restricted Isometry Constant (BRIC), GPSP exactly recovers the true block sparse signals. When the observations are noisy, this convergence property of GPSP remains valid if the magnitude of the true signal is sufficiently large. GPSP selects the features by subspace projection criterion (SPC) for candidate inclusion and response magnitude criterion (RMC) for candidate exclusion. We compare these criteria with counterparts of other state-of-the-art greedy algorithms. Our theoretical analysis and numerical ablation studies reveal that SPC is critical to the superior performances of GPSP, and that RMC can enhance the robustness of feature identification when observations contain noises. We test and compare GPSP with other methods in diverse settings, including heterogeneous random block matrices, inexact observations, face recognition, and PDE identification. We find that GPSP outperforms the other algorithms in most cases for various levels of block sparsity and block sizes, justifying its effectiveness for general applications.
本文对He et al. [26] (Group Projected Subspace Pursuit for IDENTification of variable coefficient differential equations (GP-IDENT), Journal of Computational Physics, 494, 112526)提出的Group Projected Subspace Pursuit (GPSP)算法进行了收敛性分析,并将其应用于块稀疏信号恢复的一般任务。给定观测值y和采样矩阵A,我们专注于最小化相对于具有块稀疏性约束的信号c的近似误差‖Ac−y‖22。证明了当采样矩阵A满足块受限等距特性(BRIP)且块受限等距常数(BRIC)足够小时,GPSP能准确地恢复出真实的块稀疏信号。当观测值有噪声时,如果真信号的幅度足够大,GPSP的这种收敛性仍然有效。GPSP通过子空间投影准则(SPC)选择候选包含特征,通过响应大小准则(RMC)选择候选排除特征。我们将这些标准与其他最先进的贪婪算法进行比较。我们的理论分析和数值消融研究表明,SPC是GPSP优越性能的关键,RMC可以增强观测值包含噪声时特征识别的鲁棒性。我们在不同的环境下测试并比较了GPSP和其他方法,包括异构随机块矩阵、不精确观察、人脸识别和PDE识别。我们发现,在大多数情况下,对于不同级别的块稀疏性和块大小,GPSP优于其他算法,证明其在一般应用中的有效性。
{"title":"Group projected subspace pursuit for block sparse signal reconstruction: Convergence analysis and applications 1","authors":"Roy Y. He ,&nbsp;Haixia Liu ,&nbsp;Hao Liu","doi":"10.1016/j.acha.2024.101726","DOIUrl":"10.1016/j.acha.2024.101726","url":null,"abstract":"<div><div>In this paper, we present a convergence analysis of the Group Projected Subspace Pursuit (GPSP) algorithm proposed by He et al. <span><span>[26]</span></span> (Group Projected subspace pursuit for IDENTification of variable coefficient differential equations (GP-IDENT), <em>Journal of Computational Physics</em>, 494, 112526) and extend its application to general tasks of block sparse signal recovery. Given an observation <strong>y</strong> and sampling matrix <strong>A</strong>, we focus on minimizing the approximation error <span><math><msubsup><mrow><mo>‖</mo><mi>A</mi><mi>c</mi><mo>−</mo><mi>y</mi><mo>‖</mo></mrow><mrow><mn>2</mn></mrow><mrow><mn>2</mn></mrow></msubsup></math></span> with respect to the signal <strong>c</strong> with block sparsity constraints. We prove that when the sampling matrix <strong>A</strong> satisfies the Block Restricted Isometry Property (BRIP) with a sufficiently small Block Restricted Isometry Constant (BRIC), GPSP exactly recovers the true block sparse signals. When the observations are noisy, this convergence property of GPSP remains valid if the magnitude of the true signal is sufficiently large. GPSP selects the features by subspace projection criterion (SPC) for candidate inclusion and response magnitude criterion (RMC) for candidate exclusion. We compare these criteria with counterparts of other state-of-the-art greedy algorithms. Our theoretical analysis and numerical ablation studies reveal that SPC is critical to the superior performances of GPSP, and that RMC can enhance the robustness of feature identification when observations contain noises. We test and compare GPSP with other methods in diverse settings, including heterogeneous random block matrices, inexact observations, face recognition, and PDE identification. We find that GPSP outperforms the other algorithms in most cases for various levels of block sparsity and block sizes, justifying its effectiveness for general applications.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"75 ","pages":"Article 101726"},"PeriodicalIF":2.6,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142759319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Applied and Computational Harmonic Analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1