首页 > 最新文献

Information and Inference-A Journal of the Ima最新文献

英文 中文
Multivariate super-resolution without separation 无分离的多元超分辨率
4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-27 DOI: 10.1093/imaiai/iaad024
Bakytzhan Kurmanbek, Elina Robeva
Abstract In this paper, we study the high-dimensional super-resolution imaging problem. Here, we are given an image of a number of point sources of light whose locations and intensities are unknown. The image is pixelized and is blurred by a known point-spread function arising from the imaging device. We encode the unknown point sources and their intensities via a non-negative measure and we propose a convex optimization program to find it. Assuming the device’s point-spread function is componentwise decomposable, we show that the optimal solution is the true measure in the noiseless case, and it approximates the true measure well in the noisy case with respect to the generalized Wasserstein distance. Our main assumption is that the components of the point-spread function form a Tchebychev system ($T$-system) in the noiseless case and a $T^{*}$-system in the noisy case, mild conditions that are satisfied by Gaussian point-spread functions. Our work is a generalization to all dimensions of the work [14] where the same analysis is carried out in two dimensions. We also extend results in [27] to the high-dimensional case when the point-spread function decomposes.
摘要本文研究了高维超分辨率成像问题。在这里,我们得到了一些点光源的图像,它们的位置和强度都是未知的。图像被像素化,并由成像装置产生的已知点扩散函数模糊。我们通过非负测度对未知点源及其强度进行编码,并提出了一个凸优化程序来寻找未知点源及其强度。假设装置的点扩散函数是可分解的,我们证明了最优解是无噪声情况下的真测度,并且它很好地近似于有噪声情况下的广义Wasserstein距离的真测度。我们的主要假设是,点扩展函数的分量在无噪声情况下形成一个Tchebychev系统($T$-system),在有噪声情况下形成一个$T^{*}$-系统,高斯点扩展函数满足温和的条件。我们的工作是对工作的所有维度的推广[14],其中在二维中进行了相同的分析。我们还将[27]中的结果推广到点扩散函数分解时的高维情况。
{"title":"Multivariate super-resolution without separation","authors":"Bakytzhan Kurmanbek, Elina Robeva","doi":"10.1093/imaiai/iaad024","DOIUrl":"https://doi.org/10.1093/imaiai/iaad024","url":null,"abstract":"Abstract In this paper, we study the high-dimensional super-resolution imaging problem. Here, we are given an image of a number of point sources of light whose locations and intensities are unknown. The image is pixelized and is blurred by a known point-spread function arising from the imaging device. We encode the unknown point sources and their intensities via a non-negative measure and we propose a convex optimization program to find it. Assuming the device’s point-spread function is componentwise decomposable, we show that the optimal solution is the true measure in the noiseless case, and it approximates the true measure well in the noisy case with respect to the generalized Wasserstein distance. Our main assumption is that the components of the point-spread function form a Tchebychev system ($T$-system) in the noiseless case and a $T^{*}$-system in the noisy case, mild conditions that are satisfied by Gaussian point-spread functions. Our work is a generalization to all dimensions of the work [14] where the same analysis is carried out in two dimensions. We also extend results in [27] to the high-dimensional case when the point-spread function decomposes.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"138 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136266723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A manifold two-sample test study: integral probability metric with neural networks 流形双样本检验研究:基于神经网络的积分概率度量
4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-27 DOI: 10.1093/imaiai/iaad018
Jie Wang, Minshuo Chen, Tuo Zhao, Wenjing Liao, Yao Xie
Abstract Two-sample tests are important areas aiming to determine whether two collections of observations follow the same distribution or not. We propose two-sample tests based on integral probability metric (IPM) for high-dimensional samples supported on a low-dimensional manifold. We characterize the properties of proposed tests with respect to the number of samples $n$ and the structure of the manifold with intrinsic dimension $d$. When an atlas is given, we propose a two-step test to identify the difference between general distributions, which achieves the type-II risk in the order of $n^{-1/max {d,2}}$. When an atlas is not given, we propose Hölder IPM test that applies for data distributions with $(s,beta )$-Hölder densities, which achieves the type-II risk in the order of $n^{-(s+beta )/d}$. To mitigate the heavy computation burden of evaluating the Hölder IPM, we approximate the Hölder function class using neural networks. Based on the approximation theory of neural networks, we show that the neural network IPM test has the type-II risk in the order of $n^{-(s+beta )/d}$, which is in the same order of the type-II risk as the Hölder IPM test. Our proposed tests are adaptive to low-dimensional geometric structure because their performance crucially depends on the intrinsic dimension instead of the data dimension.
摘要双样本检验是确定两个观测值集合是否遵循相同分布的重要领域。我们提出了基于积分概率度量(IPM)的两样本测试,用于支持在低维流形上的高维样本。我们根据样本数$n$和具有固有维数$d$的流形的结构来描述所提出的测试的性质。当给定地图集时,我们提出了两步检验来识别一般分布之间的差异,从而实现了$n^{-1/max {d,2}}$顺序的ii型风险。在没有给出地图集的情况下,我们提出了Hölder IPM检验,适用于密度为$(s,beta )$ -Hölder的数据分布,达到了以$n^{-(s+beta )/d}$为顺序的ii型风险。为了减轻评估Hölder IPM的繁重计算负担,我们使用神经网络近似Hölder函数类。基于神经网络逼近理论,我们证明了神经网络IPM检验具有$n^{-(s+beta )/d}$数量级的ii型风险,与Hölder IPM检验具有相同的ii型风险数量级。我们提出的测试适合低维几何结构,因为它们的性能主要取决于内在维数而不是数据维数。
{"title":"A manifold two-sample test study: integral probability metric with neural networks","authors":"Jie Wang, Minshuo Chen, Tuo Zhao, Wenjing Liao, Yao Xie","doi":"10.1093/imaiai/iaad018","DOIUrl":"https://doi.org/10.1093/imaiai/iaad018","url":null,"abstract":"Abstract Two-sample tests are important areas aiming to determine whether two collections of observations follow the same distribution or not. We propose two-sample tests based on integral probability metric (IPM) for high-dimensional samples supported on a low-dimensional manifold. We characterize the properties of proposed tests with respect to the number of samples $n$ and the structure of the manifold with intrinsic dimension $d$. When an atlas is given, we propose a two-step test to identify the difference between general distributions, which achieves the type-II risk in the order of $n^{-1/max {d,2}}$. When an atlas is not given, we propose Hölder IPM test that applies for data distributions with $(s,beta )$-Hölder densities, which achieves the type-II risk in the order of $n^{-(s+beta )/d}$. To mitigate the heavy computation burden of evaluating the Hölder IPM, we approximate the Hölder function class using neural networks. Based on the approximation theory of neural networks, we show that the neural network IPM test has the type-II risk in the order of $n^{-(s+beta )/d}$, which is in the same order of the type-II risk as the Hölder IPM test. Our proposed tests are adaptive to low-dimensional geometric structure because their performance crucially depends on the intrinsic dimension instead of the data dimension.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136266917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast and provable tensor robust principal component analysis via scaled gradient descent 快速和可证明的张量鲁棒主成分分析通过缩放梯度下降
4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-27 DOI: 10.1093/imaiai/iaad019
Harry Dong, Tian Tong, Cong Ma, Yuejie Chi
Abstract An increasing number of data science and machine learning problems rely on computation with tensors, which better capture the multi-way relationships and interactions of data than matrices. When tapping into this critical advantage, a key challenge is to develop computationally efficient and provably correct algorithms for extracting useful information from tensor data that are simultaneously robust to corruptions and ill-conditioning. This paper tackles tensor robust principal component analysis (RPCA), which aims to recover a low-rank tensor from its observations contaminated by sparse corruptions, under the Tucker decomposition. To minimize the computation and memory footprints, we propose to directly recover the low-dimensional tensor factors—starting from a tailored spectral initialization—via scaled gradient descent (ScaledGD), coupled with an iteration-varying thresholding operation to adaptively remove the impact of corruptions. Theoretically, we establish that the proposed algorithm converges linearly to the true low-rank tensor at a constant rate that is independent with its condition number, as long as the level of corruptions is not too large. Empirically, we demonstrate that the proposed algorithm achieves better and more scalable performance than state-of-the-art tensor RPCA algorithms through synthetic experiments and real-world applications.
越来越多的数据科学和机器学习问题依赖于张量计算,它比矩阵更能捕捉数据的多方向关系和相互作用。当利用这一关键优势时,一个关键的挑战是开发计算效率高且可证明正确的算法,用于从张量数据中提取有用的信息,同时对损坏和病态具有鲁棒性。本文研究了张量鲁棒主成分分析(RPCA),其目的是在Tucker分解下从被稀疏腐蚀污染的观测中恢复低秩张量。为了最大限度地减少计算和内存占用,我们建议通过缩放梯度下降(ScaledGD)从定制谱初始化开始直接恢复低维张量因子,再加上迭代变化的阈值操作,以自适应地消除损坏的影响。从理论上讲,我们建立了该算法以与条件数无关的常数速率线性收敛到真正的低秩张量,只要破坏程度不太大。通过综合实验和实际应用,我们证明了所提出的算法比最先进的张量RPCA算法具有更好的可扩展性。
{"title":"Fast and provable tensor robust principal component analysis via scaled gradient descent","authors":"Harry Dong, Tian Tong, Cong Ma, Yuejie Chi","doi":"10.1093/imaiai/iaad019","DOIUrl":"https://doi.org/10.1093/imaiai/iaad019","url":null,"abstract":"Abstract An increasing number of data science and machine learning problems rely on computation with tensors, which better capture the multi-way relationships and interactions of data than matrices. When tapping into this critical advantage, a key challenge is to develop computationally efficient and provably correct algorithms for extracting useful information from tensor data that are simultaneously robust to corruptions and ill-conditioning. This paper tackles tensor robust principal component analysis (RPCA), which aims to recover a low-rank tensor from its observations contaminated by sparse corruptions, under the Tucker decomposition. To minimize the computation and memory footprints, we propose to directly recover the low-dimensional tensor factors—starting from a tailored spectral initialization—via scaled gradient descent (ScaledGD), coupled with an iteration-varying thresholding operation to adaptively remove the impact of corruptions. Theoretically, we establish that the proposed algorithm converges linearly to the true low-rank tensor at a constant rate that is independent with its condition number, as long as the level of corruptions is not too large. Empirically, we demonstrate that the proposed algorithm achieves better and more scalable performance than state-of-the-art tensor RPCA algorithms through synthetic experiments and real-world applications.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136267080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalization error bounds for iterative recovery algorithms unfolded as neural networks 迭代恢复算法的泛化误差边界以神经网络的形式展开
4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-27 DOI: 10.1093/imaiai/iaad023
Ekkehard Schnoor, Arash Behboodi, Holger Rauhut
Abstract Motivated by the learned iterative soft thresholding algorithm (LISTA), we introduce a general class of neural networks suitable for sparse reconstruction from few linear measurements. By allowing a wide range of degrees of weight-sharing between the flayers, we enable a unified analysis for very different neural network types, ranging from recurrent ones to networks more similar to standard feedforward neural networks. Based on training samples, via empirical risk minimization, we aim at learning the optimal network parameters and thereby the optimal network that reconstructs signals from their low-dimensional linear measurements. We derive generalization bounds by analyzing the Rademacher complexity of hypothesis classes consisting of such deep networks, that also take into account the thresholding parameters. We obtain estimates of the sample complexity that essentially depend only linearly on the number of parameters and on the depth. We apply our main result to obtain specific generalization bounds for several practical examples, including different algorithms for (implicit) dictionary learning, and convolutional neural networks.
摘要:在学习迭代软阈值算法(LISTA)的激励下,我们引入了一类适用于从少量线性测量稀疏重建的神经网络。通过允许在剥层器之间广泛程度的权重共享,我们能够对非常不同的神经网络类型进行统一分析,从循环网络到更类似于标准前馈神经网络的网络。基于训练样本,通过经验风险最小化,我们的目标是学习最优网络参数,从而获得从低维线性测量中重建信号的最优网络。我们通过分析由这种深度网络组成的假设类的Rademacher复杂度来推导泛化边界,并且考虑了阈值参数。我们得到的样本复杂度的估计基本上只线性地依赖于参数的数量和深度。我们将我们的主要结果应用于几个实际示例,包括(隐式)字典学习和卷积神经网络的不同算法,以获得特定的泛化界限。
{"title":"Generalization error bounds for iterative recovery algorithms unfolded as neural networks","authors":"Ekkehard Schnoor, Arash Behboodi, Holger Rauhut","doi":"10.1093/imaiai/iaad023","DOIUrl":"https://doi.org/10.1093/imaiai/iaad023","url":null,"abstract":"Abstract Motivated by the learned iterative soft thresholding algorithm (LISTA), we introduce a general class of neural networks suitable for sparse reconstruction from few linear measurements. By allowing a wide range of degrees of weight-sharing between the flayers, we enable a unified analysis for very different neural network types, ranging from recurrent ones to networks more similar to standard feedforward neural networks. Based on training samples, via empirical risk minimization, we aim at learning the optimal network parameters and thereby the optimal network that reconstructs signals from their low-dimensional linear measurements. We derive generalization bounds by analyzing the Rademacher complexity of hypothesis classes consisting of such deep networks, that also take into account the thresholding parameters. We obtain estimates of the sample complexity that essentially depend only linearly on the number of parameters and on the depth. We apply our main result to obtain specific generalization bounds for several practical examples, including different algorithms for (implicit) dictionary learning, and convolutional neural networks.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"263 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136266735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Separation-free super-resolution from compressed measurements is possible: an orthonormal atomic norm minimization approach 从压缩测量中实现无分离的超分辨率是可能的:一种标准正交原子范数最小化方法
4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-27 DOI: 10.1093/imaiai/iaad033
Jirong Yi, Soura Dasgupta, Jian-Feng Cai, Mathews Jacob, Jingchao Gao, Myung Cho, Weiyu Xu
Abstract We consider the problem of recovering the superposition of $R$ distinct complex exponential functions from compressed non-uniform time-domain samples. Total variation (TV) minimization or atomic norm minimization was proposed in the literature to recover the $R$ frequencies or the missing data. However, it is known that in order for TV minimization and atomic norm minimization to recover the missing data or the frequencies, the underlying $R$ frequencies are required to be well separated, even when the measurements are noiseless. This paper shows that the Hankel matrix recovery approach can super-resolve the $R$ complex exponentials and their frequencies from compressed non-uniform measurements, regardless of how close their frequencies are to each other. We propose a new concept of orthonormal atomic norm minimization (OANM), and demonstrate that the success of Hankel matrix recovery in separation-free super-resolution comes from the fact that the nuclear norm of a Hankel matrix is an orthonormal atomic norm. More specifically, we show that, in traditional atomic norm minimization, the underlying parameter values must be well separated to achieve successful signal recovery, if the atoms are changing continuously with respect to the continuously valued parameter. In contrast, for the OANM, it is possible the OANM is successful even though the original atoms can be arbitrarily close. As a byproduct of this research, we provide one matrix-theoretic inequality of nuclear norm, and give its proof using the theory of compressed sensing.
摘要研究了从压缩的非均匀时域样本中恢复R不同复指数函数叠加的问题。文献中提出了总变差最小化或原子范数最小化来恢复R频率或丢失的数据。然而,众所周知,为了使TV最小化和原子范数最小化来恢复丢失的数据或频率,即使在测量是无噪声的情况下,也需要很好地分离底层R频率。本文证明了汉克尔矩阵恢复方法可以从压缩的非均匀测量中超分辨出$R$复指数及其频率,而不管它们的频率彼此有多接近。我们提出了标准正交原子范数最小化(OANM)的新概念,并证明了Hankel矩阵在无分离超分辨中恢复的成功源于Hankel矩阵的核范数是一个标准正交原子范数。更具体地说,我们表明,在传统的原子范数最小化中,如果原子相对于连续值参数连续变化,则必须很好地分离底层参数值以实现成功的信号恢复。相比之下,对于OANM,即使原始原子可以任意接近,OANM也有可能成功。作为本研究的副产品,我们给出了核范数的一个矩阵理论不等式,并利用压缩感知理论给出了它的证明。
{"title":"Separation-free super-resolution from compressed measurements is possible: an orthonormal atomic norm minimization approach","authors":"Jirong Yi, Soura Dasgupta, Jian-Feng Cai, Mathews Jacob, Jingchao Gao, Myung Cho, Weiyu Xu","doi":"10.1093/imaiai/iaad033","DOIUrl":"https://doi.org/10.1093/imaiai/iaad033","url":null,"abstract":"Abstract We consider the problem of recovering the superposition of $R$ distinct complex exponential functions from compressed non-uniform time-domain samples. Total variation (TV) minimization or atomic norm minimization was proposed in the literature to recover the $R$ frequencies or the missing data. However, it is known that in order for TV minimization and atomic norm minimization to recover the missing data or the frequencies, the underlying $R$ frequencies are required to be well separated, even when the measurements are noiseless. This paper shows that the Hankel matrix recovery approach can super-resolve the $R$ complex exponentials and their frequencies from compressed non-uniform measurements, regardless of how close their frequencies are to each other. We propose a new concept of orthonormal atomic norm minimization (OANM), and demonstrate that the success of Hankel matrix recovery in separation-free super-resolution comes from the fact that the nuclear norm of a Hankel matrix is an orthonormal atomic norm. More specifically, we show that, in traditional atomic norm minimization, the underlying parameter values must be well separated to achieve successful signal recovery, if the atoms are changing continuously with respect to the continuously valued parameter. In contrast, for the OANM, it is possible the OANM is successful even though the original atoms can be arbitrarily close. As a byproduct of this research, we provide one matrix-theoretic inequality of nuclear norm, and give its proof using the theory of compressed sensing.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"203 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136266739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Minimax detection of localized signals in statistical inverse problems 统计逆问题中局域信号的极大极小检测
4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-27 DOI: 10.1093/imaiai/iaad026
Markus Pohlmann, Frank Werner, Axel Munk
Abstract We investigate minimax testing for detecting local signals or linear combinations of such signals when only indirect data are available. Naturally, in the presence of noise, signals that are too small cannot be reliably detected. In a Gaussian white noise model, we discuss upper and lower bounds for the minimal size of the signal such that testing with small error probabilities is possible. In certain situations we are able to characterize the asymptotic minimax detection boundary. Our results are applied to inverse problems such as numerical differentiation, deconvolution and the inversion of the Radon transform.
摘要:我们研究了在只有间接数据可用的情况下检测局部信号或这些信号的线性组合的极大极小检验。当然,在噪声存在的情况下,太小的信号不能被可靠地检测到。在高斯白噪声模型中,我们讨论了信号最小尺寸的上界和下界,使小误差概率的测试成为可能。在某些情况下,我们能够描述渐近极大极小检测边界。我们的结果应用于数值微分、反卷积和Radon变换反演等反问题。
{"title":"Minimax detection of localized signals in statistical inverse problems","authors":"Markus Pohlmann, Frank Werner, Axel Munk","doi":"10.1093/imaiai/iaad026","DOIUrl":"https://doi.org/10.1093/imaiai/iaad026","url":null,"abstract":"Abstract We investigate minimax testing for detecting local signals or linear combinations of such signals when only indirect data are available. Naturally, in the presence of noise, signals that are too small cannot be reliably detected. In a Gaussian white noise model, we discuss upper and lower bounds for the minimal size of the signal such that testing with small error probabilities is possible. In certain situations we are able to characterize the asymptotic minimax detection boundary. Our results are applied to inverse problems such as numerical differentiation, deconvolution and the inversion of the Radon transform.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"136266731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sharp, strong and unique minimizers for low complexity robust recovery 锐利,强大和独特的最小化低复杂性稳健恢复
4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-23 DOI: 10.1093/imaiai/iaad005
Jalal Fadili, Tran T. A. Nghia, Trinh T. T. Tran
Abstract In this paper, we show the important roles of sharp minima and strong minima for robust recovery. We also obtain several characterizations of sharp minima for convex regularized optimization problems. Our characterizations are quantitative and verifiable especially for the case of decomposable norm regularized problems including sparsity, group-sparsity and low-rank convex problems. For group-sparsity optimization problems, we show that a unique solution is a strong solution and obtains quantitative characterizations for solution uniqueness.
摘要本文证明了锐极小值和强极小值在鲁棒恢复中的重要作用。对于凸正则化优化问题,我们也得到了尖锐极小值的几个特征。特别是对于可分解范数正则化问题,包括稀疏性、群稀疏性和低秩凸性问题,我们的刻画是定量的和可验证的。对于群稀疏优化问题,我们证明了唯一解是强解,并得到了解唯一性的定量表征。
{"title":"Sharp, strong and unique minimizers for low complexity robust recovery","authors":"Jalal Fadili, Tran T. A. Nghia, Trinh T. T. Tran","doi":"10.1093/imaiai/iaad005","DOIUrl":"https://doi.org/10.1093/imaiai/iaad005","url":null,"abstract":"Abstract In this paper, we show the important roles of sharp minima and strong minima for robust recovery. We also obtain several characterizations of sharp minima for convex regularized optimization problems. Our characterizations are quantitative and verifiable especially for the case of decomposable norm regularized problems including sparsity, group-sparsity and low-rank convex problems. For group-sparsity optimization problems, we show that a unique solution is a strong solution and obtains quantitative characterizations for solution uniqueness.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134956496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Non-adaptive algorithms for threshold group testing with consecutive positives 连续阳性阈值组检测的非自适应算法
IF 1.6 4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-04-04 DOI: 10.1093/imaiai/iaad009
Given up to $d$ positive items in a large population of $n$ items ($d ll n$), the goal of threshold group testing is to efficiently identify the positives via tests, where a test on a subset of items is positive if the subset contains at least $u$ positive items, negative if it contains up to $ell $ positive items and arbitrary (either positive or negative) otherwise. The parameter $g = u - ell - 1$ is called the gap. In non-adaptive strategies, all tests are fixed in advance and can be represented as a measurement matrix, in which each row and column represent a test and an item, respectively. In this paper, we consider non-adaptive threshold group testing with consecutive positives in which the items are linearly ordered and the positives are consecutive in that order. We show that by designing deterministic and strongly explicit measurement matrices, $lceil log _{2}{lceil frac {n}{d} rceil } rceil + 2d + 3$ (respectively, $lceil log _{2}{lceil frac {n}{d} rceil } rceil + 3d$) tests suffice to identify the positives in $O left ( log _{2}{frac {n}{d}} + d right )$ time when $g = 0$ (respectively, $g> 0$). The results significantly improve the state-of-the-art scheme that needs $15 lceil log _{2}{lceil frac {n}{d} rceil } rceil + 4d + 71$ tests to identify the positives in $O left ( frac {n}{d} log _{2}{frac {n}{d}} + ud^{2} right )$ time, and whose associated measurement matrices are random and (non-strongly) explicit.
在大量的$n$项($d ll n$)中给出$d$阳性项目,阈值组测试的目标是通过测试有效地识别阳性项目,其中,如果对项目子集的测试至少包含$u$个阳性项目,则为阳性,如果包含多达$ell $个阳性项目,则为阴性,否则为任意(阳性或阴性)。参数$g = u - ell - 1$称为间隙。在非自适应策略中,所有的测试都是预先固定的,可以用测量矩阵表示,其中每一行和每一列分别代表一个测试和一个项目。在本文中,我们考虑具有连续阳性的非自适应阈值群检验,其中项目是线性有序的,阳性是连续的。我们表明,通过设计确定性和强显式测量矩阵,$lceil log _{2}{lceil frac {n}{d} rceil } rceil + 2d + 3$(分别,$lceil log _{2}{lceil frac {n}{d} rceil } rceil + 3d$)测试足以识别$g = 0$(分别,$g> 0$)时$O left ( log _{2}{frac {n}{d}} + d right )$时间的阳性。结果显著改进了最先进的方案,该方案需要$15 lceil log _{2}{lceil frac {n}{d} rceil } rceil + 4d + 71$测试来识别$O left ( frac {n}{d} log _{2}{frac {n}{d}} + ud^{2} right )$时间内的阳性,并且其相关的测量矩阵是随机和(非强)显式的。
{"title":"Non-adaptive algorithms for threshold group testing with consecutive positives","authors":"","doi":"10.1093/imaiai/iaad009","DOIUrl":"https://doi.org/10.1093/imaiai/iaad009","url":null,"abstract":"\u0000 Given up to $d$ positive items in a large population of $n$ items ($d ll n$), the goal of threshold group testing is to efficiently identify the positives via tests, where a test on a subset of items is positive if the subset contains at least $u$ positive items, negative if it contains up to $ell $ positive items and arbitrary (either positive or negative) otherwise. The parameter $g = u - ell - 1$ is called the gap. In non-adaptive strategies, all tests are fixed in advance and can be represented as a measurement matrix, in which each row and column represent a test and an item, respectively. In this paper, we consider non-adaptive threshold group testing with consecutive positives in which the items are linearly ordered and the positives are consecutive in that order. We show that by designing deterministic and strongly explicit measurement matrices, $lceil log _{2}{lceil frac {n}{d} rceil } rceil + 2d + 3$ (respectively, $lceil log _{2}{lceil frac {n}{d} rceil } rceil + 3d$) tests suffice to identify the positives in $O left ( log _{2}{frac {n}{d}} + d right )$ time when $g = 0$ (respectively, $g> 0$). The results significantly improve the state-of-the-art scheme that needs $15 lceil log _{2}{lceil frac {n}{d} rceil } rceil + 4d + 71$ tests to identify the positives in $O left ( frac {n}{d} log _{2}{frac {n}{d}} + ud^{2} right )$ time, and whose associated measurement matrices are random and (non-strongly) explicit.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"25 1","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74278014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Theoretical analysis and computation of the sample Fréchet mean of sets of large graphs for various metrics 各种指标的大图集样本均值的理论分析与计算
IF 1.6 4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-03-28 DOI: 10.1093/imaiai/iaad002
Daniel Ferguson, F. G. Meyer
To characterize the location (mean, median) of a set of graphs, one needs a notion of centrality that has been adapted to metric spaces. A standard approach is to consider the Fréchet mean. In practice, computing the Fréchet mean for sets of large graphs presents many computational issues. In this work, we suggest a method that may be used to compute the Fréchet mean for sets of graphs which is metric independent. We show that the technique proposed can be used to determine the Fréchet mean when considering the Hamming distance or a distance defined by the difference between the spectra of the adjacency matrices of the graphs.
为了描述一组图的位置(平均值,中位数),我们需要一个适用于度量空间的中心性概念。一种标准的方法是考虑fr切特平均值。在实践中,计算大型图集的fr平均值会出现许多计算问题。在这项工作中,我们提出了一种方法,可用于计算与度量无关的图集的fr平均值。我们证明,当考虑汉明距离或由图的邻接矩阵的谱之间的差定义的距离时,所提出的技术可以用来确定fr平均。
{"title":"Theoretical analysis and computation of the sample Fréchet mean of sets of large graphs for various metrics","authors":"Daniel Ferguson, F. G. Meyer","doi":"10.1093/imaiai/iaad002","DOIUrl":"https://doi.org/10.1093/imaiai/iaad002","url":null,"abstract":"\u0000 To characterize the location (mean, median) of a set of graphs, one needs a notion of centrality that has been adapted to metric spaces. A standard approach is to consider the Fréchet mean. In practice, computing the Fréchet mean for sets of large graphs presents many computational issues. In this work, we suggest a method that may be used to compute the Fréchet mean for sets of graphs which is metric independent. We show that the technique proposed can be used to determine the Fréchet mean when considering the Hamming distance or a distance defined by the difference between the spectra of the adjacency matrices of the graphs.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"105 1","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89007827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Local Viterbi property in decoding 译码中的局部维特比特性
IF 1.6 4区 数学 Q2 MATHEMATICS, APPLIED Pub Date : 2023-03-20 DOI: 10.1093/imaiai/iaad004
J. Lember
The article studies the decoding problem (also known as the classification or the segmentation problem) with pairwise Markov models (PMMs). A PMM is a process where the observation process and the underlying state sequence form a two-dimensional Markov chain, a natural generalization of hidden Markov model. The standard solutions to the decoding problem are the so-called Viterbi path—a sequence with maximum state path probability given the observations—or the pointwise maximum a posteriori (PMAP) path that maximizes the expected number of correctly classified entries. When the goal is to simultaneously maximize both criterions—conditional probability (corresponding to Viterbi path) and pointwise conditional probability (corresponding to PMAP path)—then they are combined into one single criterion via the regularization parameter $C$. The main objective of the article is to study the behaviour of the solution—called the hybrid path—as $C$ grows. Increasing $C$ increases the conditional probability of the hybrid path and when $C$ is big enough then every hybrid path is a Viterbi path. We show that hybrid paths also approach the Viterbi path locally: we define $m$-locally Viterbi paths and show that the hybrid path is $m$-locally Viterbi whenever $C$ is big enough. This all might lead to an impression that when $C$ is relatively big then any hybrid path that is not yet Viterbi differs from the Viterbi path by a few single entries only. We argue that this intuition is wrong, because when unique and $m$-locally Viterbi, then different hybrid paths differ by at least $m$ entries. Thus, when $C$ increases then the different hybrid paths tend to differ from each other by larger and larger intervals. Hence the hybrid paths might offer a variety of rather different solutions to the decoding problem.
本文研究了基于成对马尔可夫模型的解码问题(也称为分类或分割问题)。PMM是观测过程和底层状态序列形成二维马尔可夫链的过程,是隐马尔可夫模型的自然推广。解码问题的标准解决方案是所谓的Viterbi路径——在给定观测值的情况下具有最大状态路径概率的序列——或者是使正确分类条目的期望数量最大化的点向最大后验路径(PMAP)。当目标是同时最大化条件概率(对应于Viterbi路径)和点向条件概率(对应于PMAP路径)这两个标准时,它们通过正则化参数$C$组合成一个标准。本文的主要目的是研究解(称为混合路径)随着C的增长的行为。增加C会增加混合路径的条件概率当C足够大时每个混合路径都是维特比路径。我们证明了混合路径也接近局部Viterbi路径:我们定义了$m$-局部Viterbi路径,并证明当$C$足够大时混合路径是$m$-局部Viterbi。这可能会给人一种印象,当$C$比较大时,任何还不是Viterbi的混合路径与Viterbi路径只相差几个单条目。我们认为这种直觉是错误的,因为当唯一且$m$-局部维特比时,不同的混合路径至少相差$m$项。因此,当$C$增加时,不同的混合路径之间的差异会越来越大。因此,混合路径可能为解码问题提供各种不同的解决方案。
{"title":"Local Viterbi property in decoding","authors":"J. Lember","doi":"10.1093/imaiai/iaad004","DOIUrl":"https://doi.org/10.1093/imaiai/iaad004","url":null,"abstract":"\u0000 The article studies the decoding problem (also known as the classification or the segmentation problem) with pairwise Markov models (PMMs). A PMM is a process where the observation process and the underlying state sequence form a two-dimensional Markov chain, a natural generalization of hidden Markov model. The standard solutions to the decoding problem are the so-called Viterbi path—a sequence with maximum state path probability given the observations—or the pointwise maximum a posteriori (PMAP) path that maximizes the expected number of correctly classified entries. When the goal is to simultaneously maximize both criterions—conditional probability (corresponding to Viterbi path) and pointwise conditional probability (corresponding to PMAP path)—then they are combined into one single criterion via the regularization parameter $C$. The main objective of the article is to study the behaviour of the solution—called the hybrid path—as $C$ grows. Increasing $C$ increases the conditional probability of the hybrid path and when $C$ is big enough then every hybrid path is a Viterbi path. We show that hybrid paths also approach the Viterbi path locally: we define $m$-locally Viterbi paths and show that the hybrid path is $m$-locally Viterbi whenever $C$ is big enough. This all might lead to an impression that when $C$ is relatively big then any hybrid path that is not yet Viterbi differs from the Viterbi path by a few single entries only. We argue that this intuition is wrong, because when unique and $m$-locally Viterbi, then different hybrid paths differ by at least $m$ entries. Thus, when $C$ increases then the different hybrid paths tend to differ from each other by larger and larger intervals. Hence the hybrid paths might offer a variety of rather different solutions to the decoding problem.","PeriodicalId":45437,"journal":{"name":"Information and Inference-A Journal of the Ima","volume":"95 1","pages":""},"PeriodicalIF":1.6,"publicationDate":"2023-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82746142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Information and Inference-A Journal of the Ima
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1