Pub Date : 2025-01-01Epub Date: 2024-11-13DOI: 10.1016/j.acha.2024.101721
Jeremy Hoskins , Manas Rachh , Bowei Wu
This paper describes a trapezoidal quadrature method for the discretization of weakly singular, and hypersingular boundary integral operators with complex symmetric quadratic forms. Such integral operators naturally arise when complex coordinate methods or complexified contour methods are used for the solution of time-harmonic acoustic and electromagnetic interface problems in three dimensions. The quadrature is an extension of a locally corrected punctured trapezoidal rule in parameter space wherein the correction weights are determined by fitting moments of error in the punctured trapezoidal rule, which is known analytically in terms of the Epstein zeta function. In this work, we analyze the analytic continuation of the Epstein zeta function and the generalized Wigner limits to complex quadratic forms; this analysis is essential to apply the fitting procedure for computing the correction weights. We illustrate the high-order convergence of this approach through several numerical examples.
{"title":"On quadrature for singular integral operators with complex symmetric quadratic forms","authors":"Jeremy Hoskins , Manas Rachh , Bowei Wu","doi":"10.1016/j.acha.2024.101721","DOIUrl":"10.1016/j.acha.2024.101721","url":null,"abstract":"<div><div>This paper describes a trapezoidal quadrature method for the discretization of weakly singular, and hypersingular boundary integral operators with complex symmetric quadratic forms. Such integral operators naturally arise when complex coordinate methods or complexified contour methods are used for the solution of time-harmonic acoustic and electromagnetic interface problems in three dimensions. The quadrature is an extension of a locally corrected punctured trapezoidal rule in parameter space wherein the correction weights are determined by fitting moments of error in the punctured trapezoidal rule, which is known analytically in terms of the Epstein zeta function. In this work, we analyze the analytic continuation of the Epstein zeta function and the generalized Wigner limits to complex quadratic forms; this analysis is essential to apply the fitting procedure for computing the correction weights. We illustrate the high-order convergence of this approach through several numerical examples.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101721"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-11-19DOI: 10.1016/j.acha.2024.101725
Lexing Ying
This note considers the multidimensional unstructured sparse recovery problems. Examples include Fourier inversion and sparse deconvolution. The eigenmatrix is a data-driven construction with desired approximate eigenvalues and eigenvectors proposed for the one-dimensional problems. This note extends the eigenmatrix approach to multidimensional problems, providing a rather unified treatment for general kernels and unstructured sampling grids in both real and complex settings. Numerical results are provided to demonstrate the performance of the proposed method.
{"title":"Multidimensional unstructured sparse recovery via eigenmatrix","authors":"Lexing Ying","doi":"10.1016/j.acha.2024.101725","DOIUrl":"10.1016/j.acha.2024.101725","url":null,"abstract":"<div><div>This note considers the multidimensional unstructured sparse recovery problems. Examples include Fourier inversion and sparse deconvolution. The eigenmatrix is a data-driven construction with desired approximate eigenvalues and eigenvectors proposed for the one-dimensional problems. This note extends the eigenmatrix approach to multidimensional problems, providing a rather unified treatment for general kernels and unstructured sampling grids in both real and complex settings. Numerical results are provided to demonstrate the performance of the proposed method.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101725"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142696457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-10-10DOI: 10.1016/j.acha.2024.101713
Ronald DeVore , Robert D. Nowak , Rahul Parhi , Jonathan W. Siegel
We investigate the approximation of functions f on a bounded domain by the outputs of single-hidden-layer ReLU neural networks of width n. This form of nonlinear n-term dictionary approximation has been intensely studied since it is the simplest case of neural network approximation (NNA). There are several celebrated approximation results for this form of NNA that introduce novel model classes of functions on Ω whose approximation rates do not grow unbounded with the input dimension. These novel classes include Barron classes, and classes based on sparsity or variation such as the Radon-domain BV classes. The present paper is concerned with the definition of these novel model classes on domains Ω. The current definition of these model classes does not depend on the domain Ω. A new and more proper definition of model classes on domains is given by introducing the concept of weighted variation spaces. These new model classes are intrinsic to the domain itself. The importance of these new model classes is that they are strictly larger than the classical (domain-independent) classes. Yet, it is shown that they maintain the same NNA rates.
我们研究了宽度为 n 的单隐层 ReLU 神经网络输出对有界域 Ω⊂Rd 上函数 f 的逼近。这种形式的 NNA 有几个著名的逼近结果,它们引入了 Ω 上函数的新模型类,其逼近率不会随着输入维度的增加而无限制地增长。这些新类包括巴伦类,以及基于稀疏性或变化的类,如拉顿域 BV 类。目前这些模型类的定义并不依赖于域 Ω。通过引入加权变异空间的概念,我们给出了关于域上模型类的更恰当的新定义。这些新的模型类是领域本身所固有的。这些新模型类的重要性在于,它们严格来说比经典(与域无关)类大。然而,研究表明它们保持了相同的 NNA 率。
{"title":"Weighted variation spaces and approximation by shallow ReLU networks","authors":"Ronald DeVore , Robert D. Nowak , Rahul Parhi , Jonathan W. Siegel","doi":"10.1016/j.acha.2024.101713","DOIUrl":"10.1016/j.acha.2024.101713","url":null,"abstract":"<div><div>We investigate the approximation of functions <em>f</em> on a bounded domain <span><math><mi>Ω</mi><mo>⊂</mo><msup><mrow><mi>R</mi></mrow><mrow><mi>d</mi></mrow></msup></math></span> by the outputs of single-hidden-layer ReLU neural networks of width <em>n</em>. This form of nonlinear <em>n</em>-term dictionary approximation has been intensely studied since it is the simplest case of neural network approximation (NNA). There are several celebrated approximation results for this form of NNA that introduce novel model classes of functions on Ω whose approximation rates do not grow unbounded with the input dimension. These novel classes include Barron classes, and classes based on sparsity or variation such as the Radon-domain BV classes. The present paper is concerned with the definition of these novel model classes on domains Ω. The current definition of these model classes does not depend on the domain Ω. A new and more proper definition of model classes on domains is given by introducing the concept of weighted variation spaces. These new model classes are intrinsic to the domain itself. The importance of these new model classes is that they are strictly larger than the classical (domain-independent) classes. Yet, it is shown that they maintain the same NNA rates.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101713"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142432242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-10-24DOI: 10.1016/j.acha.2024.101715
Eyar Azar , Satish Mulleti , Yonina C. Eldar
Analog-to-digital converters (ADCs) act as a bridge between the analog and digital domains. Two important attributes of any ADC are sampling rate and its dynamic range. For bandlimited signals, the sampling should be above the Nyquist rate. It is also desired that the signals' dynamic range should be within that of the ADC's; otherwise, the signal will be clipped. Nonlinear operators such as modulo or companding can be used prior to sampling to avoid clipping. To recover the true signal from the samples of the nonlinear operator, either high sampling rates are required, or strict constraints on the nonlinear operations are imposed, both of which are not desirable in practice. In this paper, we propose a generalized flexible nonlinear operator which is sampling efficient. Moreover, by carefully choosing its parameters, clipping, modulo, and companding can be seen as special cases of it. We show that bandlimited signals are uniquely identified from the nonlinear samples of the proposed operator when sampled above the Nyquist rate. Furthermore, we propose a robust algorithm to recover the true signal from the nonlinear samples. Compared to the existing methods, our approach has a lower mean-squared error for a given sampling rate, noise level, and dynamic range. Our results lead to less constrained hardware design to address the dynamic range issues while operating at the lowest rate possible.
{"title":"Unlimited sampling beyond modulo","authors":"Eyar Azar , Satish Mulleti , Yonina C. Eldar","doi":"10.1016/j.acha.2024.101715","DOIUrl":"10.1016/j.acha.2024.101715","url":null,"abstract":"<div><div>Analog-to-digital converters (ADCs) act as a bridge between the analog and digital domains. Two important attributes of any ADC are sampling rate and its dynamic range. For bandlimited signals, the sampling should be above the Nyquist rate. It is also desired that the signals' dynamic range should be within that of the ADC's; otherwise, the signal will be clipped. Nonlinear operators such as modulo or companding can be used prior to sampling to avoid clipping. To recover the true signal from the samples of the nonlinear operator, either high sampling rates are required, or strict constraints on the nonlinear operations are imposed, both of which are not desirable in practice. In this paper, we propose a generalized flexible nonlinear operator which is sampling efficient. Moreover, by carefully choosing its parameters, clipping, modulo, and companding can be seen as special cases of it. We show that bandlimited signals are uniquely identified from the nonlinear samples of the proposed operator when sampled above the Nyquist rate. Furthermore, we propose a robust algorithm to recover the true signal from the nonlinear samples. Compared to the existing methods, our approach has a lower mean-squared error for a given sampling rate, noise level, and dynamic range. Our results lead to less constrained hardware design to address the dynamic range issues while operating at the lowest rate possible.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101715"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142554887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-11-13DOI: 10.1016/j.acha.2024.101722
Gi-Ren Liu , Yuan-Chung Sheu , Hau-Tieng Wu
The moving average of the complex modulus of the analytic wavelet transform provides a robust time-scale representation for signals to small time shifts and deformation. In this work, we derive the Wiener chaos expansion of this representation for stationary Gaussian processes by the Malliavin calculus and combinatorial techniques. The expansion allows us to obtain a lower bound for the Wasserstein distance between the time-scale representations of two long-range dependent Gaussian processes in terms of Hurst indices. Moreover, we apply the expansion to establish an upper bound for the smooth Wasserstein distance and the Kolmogorov distance between the distributions of a random vector derived from the time-scale representation and its normal counterpart. It is worth mentioning that the expansion consists of infinite Wiener chaos, and the projection coefficients converge to zero slowly as the order of the Wiener chaos increases. We provide a rational-decay upper bound for these distribution distances, the rate of which depends on the nonlinear transformation of the amplitude of the complex wavelet coefficients.
{"title":"Gaussian approximation for the moving averaged modulus wavelet transform and its variants","authors":"Gi-Ren Liu , Yuan-Chung Sheu , Hau-Tieng Wu","doi":"10.1016/j.acha.2024.101722","DOIUrl":"10.1016/j.acha.2024.101722","url":null,"abstract":"<div><div>The moving average of the complex modulus of the analytic wavelet transform provides a robust time-scale representation for signals to small time shifts and deformation. In this work, we derive the Wiener chaos expansion of this representation for stationary Gaussian processes by the Malliavin calculus and combinatorial techniques. The expansion allows us to obtain a lower bound for the Wasserstein distance between the time-scale representations of two long-range dependent Gaussian processes in terms of Hurst indices. Moreover, we apply the expansion to establish an upper bound for the smooth Wasserstein distance and the Kolmogorov distance between the distributions of a random vector derived from the time-scale representation and its normal counterpart. It is worth mentioning that the expansion consists of infinite Wiener chaos, and the projection coefficients converge to zero slowly as the order of the Wiener chaos increases. We provide a rational-decay upper bound for these distribution distances, the rate of which depends on the nonlinear transformation of the amplitude of the complex wavelet coefficients.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101722"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-09-26DOI: 10.1016/j.acha.2024.101709
Luís Daniel Abreu , Michael Speckbacher
In this paper we formulate Donoho and Logan's large sieve principle for the wavelet transform on the Hardy space, adapting the concept of maximum Nyquist density to the hyperbolic geometry of the underlying space. The results provide deterministic guarantees for -minimization methods and hold for a class of mother wavelets that constitutes an orthonormal basis of the Hardy space and can be associated with higher hyperbolic Landau levels. Explicit calculations of the basis functions reveal a connection with the Zernike polynomials. We prove a novel local reproducing formula for the spaces in consideration and use it to derive concentration estimates of the large sieve type for the corresponding wavelet transforms. We conclude with a discussion of optimality of localization and Lieb inequalities in the analytic case by building on recent results of Kulikov, Ramos and Tilli based on the groundbreaking methods of Nicola and Tilli. This leads to a sharp uncertainty principle and a local Lieb inequality for the wavelet transform.
{"title":"Donoho-Logan large sieve principles for the wavelet transform","authors":"Luís Daniel Abreu , Michael Speckbacher","doi":"10.1016/j.acha.2024.101709","DOIUrl":"10.1016/j.acha.2024.101709","url":null,"abstract":"<div><div>In this paper we formulate Donoho and Logan's large sieve principle for the wavelet transform on the Hardy space, adapting the concept of maximum Nyquist density to the hyperbolic geometry of the underlying space. The results provide deterministic guarantees for <span><math><msub><mrow><mi>L</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span>-minimization methods and hold for a class of mother wavelets that constitutes an orthonormal basis of the Hardy space and can be associated with higher hyperbolic Landau levels. Explicit calculations of the basis functions reveal a connection with the Zernike polynomials. We prove a novel local reproducing formula for the spaces in consideration and use it to derive concentration estimates of the large sieve type for the corresponding wavelet transforms. We conclude with a discussion of optimality of localization and Lieb inequalities in the analytic case by building on recent results of Kulikov, Ramos and Tilli based on the groundbreaking methods of Nicola and Tilli. This leads to a sharp uncertainty principle and a local Lieb inequality for the wavelet transform.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101709"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142425138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We introduce LOT Wassmap, a computationally feasible algorithm to uncover low-dimensional structures in the Wasserstein space. The algorithm is motivated by the observation that many datasets are naturally interpreted as probability measures rather than points in , and that finding low-dimensional descriptions of such datasets requires manifold learning algorithms in the Wasserstein space. Most available algorithms are based on computing the pairwise Wasserstein distance matrix, which can be computationally challenging for large datasets in high dimensions. Our algorithm leverages approximation schemes such as Sinkhorn distances and linearized optimal transport to speed-up computations, and in particular, avoids computing a pairwise distance matrix. We provide guarantees on the embedding quality under such approximations, including when explicit descriptions of the probability measures are not available and one must deal with finite samples instead. Experiments demonstrate that LOT Wassmap attains correct embeddings and that the quality improves with increased sample size. We also show how LOT Wassmap significantly reduces the computational cost when compared to algorithms that depend on pairwise distance computations.
{"title":"Linearized Wasserstein dimensionality reduction with approximation guarantees","authors":"Alexander Cloninger , Keaton Hamm , Varun Khurana , Caroline Moosmüller","doi":"10.1016/j.acha.2024.101718","DOIUrl":"10.1016/j.acha.2024.101718","url":null,"abstract":"<div><div>We introduce LOT Wassmap, a computationally feasible algorithm to uncover low-dimensional structures in the Wasserstein space. The algorithm is motivated by the observation that many datasets are naturally interpreted as probability measures rather than points in <span><math><msup><mrow><mi>R</mi></mrow><mrow><mi>n</mi></mrow></msup></math></span>, and that finding low-dimensional descriptions of such datasets requires manifold learning algorithms in the Wasserstein space. Most available algorithms are based on computing the pairwise Wasserstein distance matrix, which can be computationally challenging for large datasets in high dimensions. Our algorithm leverages approximation schemes such as Sinkhorn distances and linearized optimal transport to speed-up computations, and in particular, avoids computing a pairwise distance matrix. We provide guarantees on the embedding quality under such approximations, including when explicit descriptions of the probability measures are not available and one must deal with finite samples instead. Experiments demonstrate that LOT Wassmap attains correct embeddings and that the quality improves with increased sample size. We also show how LOT Wassmap significantly reduces the computational cost when compared to algorithms that depend on pairwise distance computations.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101718"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142535998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-10-21DOI: 10.1016/j.acha.2024.101710
Gary Froyland, Christopher P. Rock
This paper investigates links between the eigenvalues and eigenfunctions of the Laplace-Beltrami operator, and the higher Cheeger constants of smooth Riemannian manifolds, possibly weighted and/or with boundary. The higher Cheeger constants give a loose description of the major geometric features of a manifold. We give a constructive upper bound on the higher Cheeger constants, in terms of the eigenvalue of any eigenfunction with the corresponding number of nodal domains. Specifically, we show that for each such eigenfunction, a positive-measure collection of its superlevel sets have their Cheeger ratios bounded above in terms of the corresponding eigenvalue.
Some manifolds have their major features entwined across several eigenfunctions, and no single eigenfunction contains all the major features. In this case, there may exist carefully chosen linear combinations of the eigenfunctions, each with large values on a single feature, and small values elsewhere. We can then apply a soft-thresholding operator to these linear combinations to obtain new functions, each supported on a single feature. We show that the Cheeger ratios of the level sets of these functions also give an upper bound on the Laplace-Beltrami eigenvalues. We extend these level set results to nonautonomous dynamical systems, and show that the dynamic Laplacian eigenfunctions reveal sets with small dynamic Cheeger ratios.
{"title":"Higher Cheeger ratios of features in Laplace-Beltrami eigenfunctions","authors":"Gary Froyland, Christopher P. Rock","doi":"10.1016/j.acha.2024.101710","DOIUrl":"10.1016/j.acha.2024.101710","url":null,"abstract":"<div><div>This paper investigates links between the eigenvalues and eigenfunctions of the Laplace-Beltrami operator, and the higher Cheeger constants of smooth Riemannian manifolds, possibly weighted and/or with boundary. The higher Cheeger constants give a loose description of the major geometric features of a manifold. We give a constructive upper bound on the higher Cheeger constants, in terms of the eigenvalue of any eigenfunction with the corresponding number of nodal domains. Specifically, we show that for each such eigenfunction, a positive-measure collection of its superlevel sets have their Cheeger ratios bounded above in terms of the corresponding eigenvalue.</div><div>Some manifolds have their major features entwined across several eigenfunctions, and no single eigenfunction contains all the major features. In this case, there may exist carefully chosen linear combinations of the eigenfunctions, each with large values on a single feature, and small values elsewhere. We can then apply a soft-thresholding operator to these linear combinations to obtain new functions, each supported on a single feature. We show that the Cheeger ratios of the level sets of these functions also give an upper bound on the Laplace-Beltrami eigenvalues. We extend these level set results to nonautonomous dynamical systems, and show that the dynamic Laplacian eigenfunctions reveal sets with small dynamic Cheeger ratios.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101710"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142535999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-10-24DOI: 10.1016/j.acha.2024.101719
Holger Boche , Adalbert Fono , Gitta Kutyniok
Despite the success of Deep Learning (DL) serious reliability issues such as non-robustness persist. An interesting aspect is, whether these problems arise due to insufficient tools or fundamental limitations of DL. We study this question from the computability perspective by characterizing the limits the applied hardware imposes. For this, we focus on the class of inverse problems, which, in particular, encompasses any task to reconstruct data from measurements. On digital hardware, a conceptual barrier on the capabilities of DL for solving finite-dimensional inverse problems has in fact already been derived. This paper investigates the general computation framework of Blum-Shub-Smale (BSS) machines, describing the processing and storage of arbitrary real values. Although a corresponding real-world computing device does not exist, research and development towards real number computing hardware, usually referred to by “neuromorphic computing”, has increased in recent years. In this work, we show that the framework of BSS machines does enable the algorithmic solvability of finite dimensional inverse problems. Our results emphasize the influence of the considered computing model in questions of accuracy and reliability.
{"title":"Inverse problems are solvable on real number signal processing hardware","authors":"Holger Boche , Adalbert Fono , Gitta Kutyniok","doi":"10.1016/j.acha.2024.101719","DOIUrl":"10.1016/j.acha.2024.101719","url":null,"abstract":"<div><div>Despite the success of Deep Learning (DL) serious reliability issues such as non-robustness persist. An interesting aspect is, whether these problems arise due to insufficient tools or fundamental limitations of DL. We study this question from the computability perspective by characterizing the limits the applied hardware imposes. For this, we focus on the class of inverse problems, which, in particular, encompasses any task to reconstruct data from measurements. On digital hardware, a conceptual barrier on the capabilities of DL for solving finite-dimensional inverse problems has in fact already been derived. This paper investigates the general computation framework of Blum-Shub-Smale (BSS) machines, describing the processing and storage of arbitrary real values. Although a corresponding real-world computing device does not exist, research and development towards real number computing hardware, usually referred to by “neuromorphic computing”, has increased in recent years. In this work, we show that the framework of BSS machines does enable the algorithmic solvability of finite dimensional inverse problems. Our results emphasize the influence of the considered computing model in questions of accuracy and reliability.</div></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101719"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142561034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2024-09-12DOI: 10.1016/j.acha.2024.101698
Pei-Chun Su , Hau-Tieng Wu
We develop a data-driven optimal shrinkage algorithm, named extended OptShrink (eOptShrink), for matrix denoising with high-dimensional noise and a separable covariance structure. This noise is colored and dependent across samples. The algorithm leverages the asymptotic behavior of singular values and vectors of the noisy data's random matrix. Our theory includes the sticking property of non-outlier singular values, delocalization of weak signal singular vectors, and the spectral behavior of outlier singular values and vectors. We introduce three estimators: a novel rank estimator, an estimator for the spectral distribution of the pure noise matrix, and the optimal shrinker eOptShrink. Notably, eOptShrink does not require estimating the noise's separable covariance structure. We provide a theoretical guarantee for these estimators with a convergence rate. Through numerical simulations and comparisons with state-of-the-art optimal shrinkage algorithms, we demonstrate eOptShrink's application in extracting maternal and fetal electrocardiograms from single-channel trans-abdominal maternal electrocardiograms.
{"title":"Data-driven optimal shrinkage of singular values under high-dimensional noise with separable covariance structure with application","authors":"Pei-Chun Su , Hau-Tieng Wu","doi":"10.1016/j.acha.2024.101698","DOIUrl":"10.1016/j.acha.2024.101698","url":null,"abstract":"<div><p>We develop a data-driven optimal shrinkage algorithm, named <em>extended OptShrink</em> (eOptShrink), for matrix denoising with high-dimensional noise and a separable covariance structure. This noise is colored and dependent across samples. The algorithm leverages the asymptotic behavior of singular values and vectors of the noisy data's random matrix. Our theory includes the sticking property of non-outlier singular values, delocalization of weak signal singular vectors, and the spectral behavior of outlier singular values and vectors. We introduce three estimators: a novel rank estimator, an estimator for the spectral distribution of the pure noise matrix, and the optimal shrinker eOptShrink. Notably, eOptShrink does not require estimating the noise's separable covariance structure. We provide a theoretical guarantee for these estimators with a convergence rate. Through numerical simulations and comparisons with state-of-the-art optimal shrinkage algorithms, we demonstrate eOptShrink's application in extracting maternal and fetal electrocardiograms from single-channel trans-abdominal maternal electrocardiograms.</p></div>","PeriodicalId":55504,"journal":{"name":"Applied and Computational Harmonic Analysis","volume":"74 ","pages":"Article 101698"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142241768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}